Text Generation
Transformers
Safetensors
mistral
4-bit precision
AWQ
Inference Endpoints
chatml
text-generation-inference
awq
Wukong-0.1-Mistral-7B-v0.2-AWQ / quant_config.json
Shaun Prince
add quant config
d268c49
raw
history blame
82 Bytes
{
"zero_point": true,
"q_group_size": 128,
"w_bit": 4,
"version": "GEMM"
}