Exllamav3 quantization of Qwen/Qwen3-Coder-480B-A35B-Instruct
2.00 bpw h6 114.396 GiB
The 2.00bpw quant will fit in six 24 GB cards with 40k of fp16 context.
Model tree for MikeRoz/Qwen3-Coder-480B-A35B-Instruct-exl3
Base model
Qwen/Qwen3-Coder-480B-A35B-Instruct