QwQ-32B-AWQ-4bit / README.md
bnjmnmarie's picture
Update README.md
2db6953 verified
|
raw
history blame
933 Bytes
metadata
language:
  - en
library_name: transformers
tags:
  - autoround
license: apache-2.0
base_model:
  - Qwen/QwQ-32B

Model Details

This is Qwen/QwQ-32B quantized with AutoAWQ in 4-bit. The model has been created, tested, and evaluated by The Kaitchup. The model is compatible with vLLM and Transformers.

image/png

Details on the quantization process and how to use the model here: The Kaitchup

  • Developed by: The Kaitchup
  • Language(s) (NLP): English
  • License: Apache 2.0 license

How to Support My Work

Subscribe to The Kaitchup. This helps me a lot to continue quantizing and evaluating models for free.