Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,27 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
pipeline_tag: text-generation
|
6 |
+
tags:
|
7 |
+
- chat
|
8 |
+
---
|
9 |
+
|
10 |
+
# Qwen2-Wukong-7B
|
11 |
+
|
12 |
+

|
13 |
+
|
14 |
+
|
15 |
+
Qwen2-Wukong-7B is a dealigned chat finetune of the original fantastic Qwen2-7B model by the Qwen team.
|
16 |
+
|
17 |
+
This model was trained on the teknium OpenHeremes-2.5 dataset and some supplementary datasets from Cognitive Computations
|
18 |
+
|
19 |
+
This model was trained for 3 epochs with a custom FA2 implementation for AMD cards.
|
20 |
+
|
21 |
+
Special thanks to [Tensorwave](https://tensorwave.com/) for providing the compute for this training with their fantastic Mi300x Instict nodes.
|
22 |
+
|
23 |
+
[<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
|
24 |
+
|
25 |
+
# Example Outputs
|
26 |
+
|
27 |
+
TBD
|