Update README.md
Browse files
README.md
CHANGED
@@ -19,12 +19,16 @@ language:
|
|
19 |
---
|
20 |
|
21 |
# Wiedervereinigung-7b-dpo
|
22 |
-
|
23 |

|
24 |
|
25 |
This is a dpo aligned merge of our favourite german models, scoring 7.11 on the mt-bench-de average.
|
26 |
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
|
27 |
-
Therefore the name, no nationalist ideas involved :-).
|
|
|
|
|
|
|
|
|
28 |
|
29 |
Kudos to the authors of the original models at [DiscoResearch](https://huggingface.co/DiscoResearch) and [VAGOsolutions](https://huggingface.co/VAGOsolutions), [Malte Ostendorff](https://huggingface.co/malteos)
|
30 |
and [Matthias Uhlig](https://huggingface.co/DRXD1000). We are your fan club.
|
@@ -53,8 +57,23 @@ Is the merged model alone already good? Well, of course. But it is even better w
|
|
53 |
}
|
54 |
```
|
55 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
|
57 |
-
Wiedervereinigung-7b
|
58 |
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
|
59 |
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
|
60 |
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
|
@@ -101,7 +120,7 @@ import transformers
|
|
101 |
import torch
|
102 |
|
103 |
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
|
104 |
-
messages = [{"role": "user", "content": "Was ist ein deutsches
|
105 |
|
106 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
107 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
|
|
19 |
---
|
20 |
|
21 |
# Wiedervereinigung-7b-dpo
|
22 |
+
LoneStriker/Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2
|
23 |

|
24 |
|
25 |
This is a dpo aligned merge of our favourite german models, scoring 7.11 on the mt-bench-de average.
|
26 |
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
|
27 |
+
Therefore the name, no nationalist ideas involved :-).
|
28 |
+
|
29 |
+
To improve result quality they are dpo-trained with a german translation of slimorca dpo using hermeo-7B for reject results.
|
30 |
+
|
31 |
+
If you are gpu-poor like me you can now use [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory) to train with german datasets.
|
32 |
|
33 |
Kudos to the authors of the original models at [DiscoResearch](https://huggingface.co/DiscoResearch) and [VAGOsolutions](https://huggingface.co/VAGOsolutions), [Malte Ostendorff](https://huggingface.co/malteos)
|
34 |
and [Matthias Uhlig](https://huggingface.co/DRXD1000). We are your fan club.
|
|
|
57 |
}
|
58 |
```
|
59 |
|
60 |
+
## Other Versions
|
61 |
+
|
62 |
+
A big thank you to [LoneStriker](https://huggingface.co/LoneStriker) for the quantized models.
|
63 |
+
|
64 |
+
| Name | Quant method | Bits |
|
65 |
+
| ---- | ---- | ---- |
|
66 |
+
[Wiedervereinigung-7b-dpo-GPTQ](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-GPTQ)| Unquantized | 16 |
|
67 |
+
[Wiedervereinigung-7b-dpo-GPTQ](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-GPTQ)| GPTQ | 4 |
|
68 |
+
[Wiedervereinigung-7b-dpo-AWQ](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-AWQ)| AWQ | 4 |
|
69 |
+
[Wiedervereinigung-7b-dpo-GGUF](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-GGUF)| GGUF | 3-8 |
|
70 |
+
[Wiedervereinigung-7b-dpo-8.0bpw-h8-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-8.0bpw-h8-exl2)| EXL2 | 8 |
|
71 |
+
[Wiedervereinigung-7b-dpo-6.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-6.0bpw-h6-exl2)| EXL2 | 6 |
|
72 |
+
[Wiedervereinigung-7b-dpo-5.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-5.0bpw-h6-exl2)| EXL2 | 5 |
|
73 |
+
[Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2)| EXL2 | 4 |
|
74 |
+
[Wiedervereinigung-7b-dpo-3.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-3.0bpw-h6-exl2)| EXL2 | 3 |
|
75 |
|
76 |
+
Wiedervereinigung-7b is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
|
77 |
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
|
78 |
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
|
79 |
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
|
|
|
120 |
import torch
|
121 |
|
122 |
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
|
123 |
+
messages = [{"role": "user", "content": "Was ist ein deutsches Large Language Model?"}]
|
124 |
|
125 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
126 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|