works on the latest llama.cpp version. Since this is a quant the original license of the model still applies!

Downloads last month
475
GGUF
Model size
20.9B params
Architecture
gpt-oss
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for QuantStack/InternVL3_5-GPT-OSS-20B-A4B-Preview-gguf

Quantized
(2)
this model

Collection including QuantStack/InternVL3_5-GPT-OSS-20B-A4B-Preview-gguf