Qwen
/

Text Generation
GGUF
English
chat
Inference Endpoints
conversational
feihu.hf commited on
Commit
da58b03
·
1 Parent(s): c38cd4a

update readme

Browse files
Files changed (2) hide show
  1. README.md +7 -1
  2. figures/benchmark.jpg +0 -0
README.md CHANGED
@@ -19,6 +19,10 @@ tags:
19
 
20
  QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
21
 
 
 
 
 
22
  **This repo contains the QwQ 32B model in the GGUF Format**, which has the following features:
23
  - Type: Causal Language Models
24
  - Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
@@ -32,6 +36,8 @@ QwQ is the reasoning model of the Qwen series. Compared with conventional instru
32
 
33
  **Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
34
 
 
 
35
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
36
 
37
  ## Requirements
@@ -103,7 +109,7 @@ If you find our work helpful, feel free to give us a cite.
103
 
104
  ```
105
  @misc{qwq32b,
106
- title = {Qwen2.5: A Party of Foundation Models},
107
  url = {https://qwenlm.github.io/blog/qwq-32b/},
108
  author = {Qwen Team},
109
  month = {March},
 
19
 
20
  QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
21
 
22
+ <p align="center">
23
+ <img width="100%" src="figures/benchmark.jpg">
24
+ </p>
25
+
26
  **This repo contains the QwQ 32B model in the GGUF Format**, which has the following features:
27
  - Type: Causal Language Models
28
  - Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
 
36
 
37
  **Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
38
 
39
+ You can try our [demo](https://huggingface.co/spaces/Qwen/QwQ-32B-Demo) or access QwQ models via [QwenChat](https://chat.qwen.ai).
40
+
41
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
42
 
43
  ## Requirements
 
109
 
110
  ```
111
  @misc{qwq32b,
112
+ title = {QwQ-32B: The Power of Scaling RL},
113
  url = {https://qwenlm.github.io/blog/qwq-32b/},
114
  author = {Qwen Team},
115
  month = {March},
figures/benchmark.jpg ADDED