Model details

This is Qwen/Qwen2.5-14B-Instruct quantized with AutoRound (symmetric quantization) and serialized with the GPTQ format in 4-bit.

Downloads last month
7
Safetensors
Model size
3.33B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including Siddharth63/Qwen2.5-14B-Instruct-AutoRound-GPTQ-4bit