Qwen3-Coder-480B-A35B-Instruct gguf format

Recommend to use q8_0 model. I will add Q2_K, Q3_K_S later.

tq2_0 and tq1_0 gave weird results. I removed them from repository. I will add Q2_K, Q3_K_S, Q4_K_S (in progress but upload speed is 1.5 MB/s now)

Please Note: Only q8_0 is working now

Other models, such as q4_k_s is still loading. The upload speed of hugging face today is at 1.5MB/s, so it may take some time. I will update the loading status continuously.

Downloads last month
749
GGUF
Hardware compatibility
Log In to view the estimation

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for John1604/Qwen3-Coder-480B-A35B-Instruct-gguf

Quantized
(31)
this model