Model Details

This is Qwen/QwQ-32B quantized with AutoRound (symmetric quantization) and serialized with the GPTQ format in 2-bit (group size of 32). The model has been created, tested, and evaluated by The Kaitchup. The model is compatible with vLLM and Transformers.

image/png

Details on the quantization process and how to use the model here: The Kaitchup

  • Developed by: The Kaitchup
  • Language(s) (NLP): English
  • License: Apache 2.0 license

How to Support My Work

Subscribe to The Kaitchup. This helps me a lot to continue quantizing and evaluating models for free.

Downloads last month
11
Safetensors
Model size
4.55B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for kaitchup/QwQ-32B-AutoRoundGPTQ-2bit

Base model

Qwen/Qwen2.5-32B
Finetuned
Qwen/QwQ-32B
Quantized
(93)
this model

Collection including kaitchup/QwQ-32B-AutoRoundGPTQ-2bit