johannhartmann commited on
Commit
3179d3c
·
verified ·
1 Parent(s): 1955c8f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -23
README.md CHANGED
@@ -12,15 +12,44 @@ base_model:
12
  - DRXD1000/Phoenix
13
  - VAGOsolutions/SauerkrautLM-7b-v1-mistral
14
  - malteos/hermeo-7b
 
 
 
 
15
  ---
16
 
17
  # Wiedervereinigung-7b-dpo
18
 
19
  ![image/png](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b/resolve/main/Wiedervereinigung-7b.png)
20
 
21
- Some of the best german models with 7b parameters as a dare_ties merge.
 
22
  Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
23
- Hence the name. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
 
25
  Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
26
  * [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
@@ -59,27 +88,6 @@ parameters:
59
  dtype: bfloat16
60
  ```
61
 
62
- ## mt-bench-de
63
-
64
- Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
65
-
66
- ```json
67
- {
68
- "first_turn": 7.3,
69
- "second_turn": 6.925,
70
- "categories": {
71
- "writing": 8.425,
72
- "roleplay": 8.6,
73
- "reasoning": 5.4,
74
- "math": 4.35,
75
- "coding": 4.3,
76
- "extraction": 7.975,
77
- "stem": 8.5,
78
- "humanities": 9.35
79
- },
80
- "average": 7.1125
81
- }
82
- ```
83
 
84
  ## 💻 Usage
85
 
 
12
  - DRXD1000/Phoenix
13
  - VAGOsolutions/SauerkrautLM-7b-v1-mistral
14
  - malteos/hermeo-7b
15
+ license: apache-2.0
16
+ language:
17
+ - de
18
+ - en
19
  ---
20
 
21
  # Wiedervereinigung-7b-dpo
22
 
23
  ![image/png](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b/resolve/main/Wiedervereinigung-7b.png)
24
 
25
+ This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
26
+ It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
27
  Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
28
+ Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
29
+ using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
30
+
31
+ ## mt-bench-de
32
+
33
+ Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
34
+
35
+ ```json
36
+ {
37
+ "first_turn": 7.3,
38
+ "second_turn": 6.925,
39
+ "categories": {
40
+ "writing": 8.425,
41
+ "roleplay": 8.6,
42
+ "reasoning": 5.4,
43
+ "math": 4.35,
44
+ "coding": 4.3,
45
+ "extraction": 7.975,
46
+ "stem": 8.5,
47
+ "humanities": 9.35
48
+ },
49
+ "average": 7.1125
50
+ }
51
+ ```
52
+
53
 
54
  Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
55
  * [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
 
88
  dtype: bfloat16
89
  ```
90
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
 
92
  ## 💻 Usage
93