license: apache-2.0 | |
datasets: | |
- lars1234/story_writing_benchmark | |
base_model: | |
- lars1234/Mistral-Small-24B-Instruct-2501-writer | |
# Mistral-Small-24B-Instruct-2501-writer-AWQ | |
This model is the 4-bit AWQ-quantized version of [Mistral-Small-24B-Instruct-2501-writer](lars1234/Mistral-Small-24B-Instruct-2501-writer). | |
- **Quantization Method**: AWQ (Activation-aware Weight Quantization) | |
- **Quantization Configuration**: | |
- Bit Width: 4-bit | |
- Group Size: 128 | |
- Zero Point: Enabled | |
- Version: GEMM |