Update README.md
Browse files
README.md
CHANGED
@@ -12,15 +12,44 @@ base_model:
|
|
12 |
- DRXD1000/Phoenix
|
13 |
- VAGOsolutions/SauerkrautLM-7b-v1-mistral
|
14 |
- malteos/hermeo-7b
|
|
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
# Wiedervereinigung-7b-dpo
|
18 |
|
19 |

|
20 |
|
21 |
-
|
|
|
22 |
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
|
23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
|
25 |
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
|
26 |
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
|
@@ -59,27 +88,6 @@ parameters:
|
|
59 |
dtype: bfloat16
|
60 |
```
|
61 |
|
62 |
-
## mt-bench-de
|
63 |
-
|
64 |
-
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
|
65 |
-
|
66 |
-
```json
|
67 |
-
{
|
68 |
-
"first_turn": 7.3,
|
69 |
-
"second_turn": 6.925,
|
70 |
-
"categories": {
|
71 |
-
"writing": 8.425,
|
72 |
-
"roleplay": 8.6,
|
73 |
-
"reasoning": 5.4,
|
74 |
-
"math": 4.35,
|
75 |
-
"coding": 4.3,
|
76 |
-
"extraction": 7.975,
|
77 |
-
"stem": 8.5,
|
78 |
-
"humanities": 9.35
|
79 |
-
},
|
80 |
-
"average": 7.1125
|
81 |
-
}
|
82 |
-
```
|
83 |
|
84 |
## 💻 Usage
|
85 |
|
|
|
12 |
- DRXD1000/Phoenix
|
13 |
- VAGOsolutions/SauerkrautLM-7b-v1-mistral
|
14 |
- malteos/hermeo-7b
|
15 |
+
license: apache-2.0
|
16 |
+
language:
|
17 |
+
- de
|
18 |
+
- en
|
19 |
---
|
20 |
|
21 |
# Wiedervereinigung-7b-dpo
|
22 |
|
23 |

|
24 |
|
25 |
+
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
|
26 |
+
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
|
27 |
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
|
28 |
+
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
|
29 |
+
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
|
30 |
+
|
31 |
+
## mt-bench-de
|
32 |
+
|
33 |
+
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
|
34 |
+
|
35 |
+
```json
|
36 |
+
{
|
37 |
+
"first_turn": 7.3,
|
38 |
+
"second_turn": 6.925,
|
39 |
+
"categories": {
|
40 |
+
"writing": 8.425,
|
41 |
+
"roleplay": 8.6,
|
42 |
+
"reasoning": 5.4,
|
43 |
+
"math": 4.35,
|
44 |
+
"coding": 4.3,
|
45 |
+
"extraction": 7.975,
|
46 |
+
"stem": 8.5,
|
47 |
+
"humanities": 9.35
|
48 |
+
},
|
49 |
+
"average": 7.1125
|
50 |
+
}
|
51 |
+
```
|
52 |
+
|
53 |
|
54 |
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
|
55 |
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
|
|
|
88 |
dtype: bfloat16
|
89 |
```
|
90 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
|
92 |
## 💻 Usage
|
93 |
|