johannhartmann commited on
Commit
2cf6771
·
verified ·
1 Parent(s): 90c6620

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -4
README.md CHANGED
@@ -19,12 +19,16 @@ language:
19
  ---
20
 
21
  # Wiedervereinigung-7b-dpo
22
-
23
  ![image/png](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b/resolve/main/Wiedervereinigung-7b.png)
24
 
25
  This is a dpo aligned merge of our favourite german models, scoring 7.11 on the mt-bench-de average.
26
  Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
27
- Therefore the name, no nationalist ideas involved :-). To improve result quality they are dpo-trained with a german translation of intel-orca-dpo we added support for 10 german datasets to [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory).
 
 
 
 
28
 
29
  Kudos to the authors of the original models at [DiscoResearch](https://huggingface.co/DiscoResearch) and [VAGOsolutions](https://huggingface.co/VAGOsolutions), [Malte Ostendorff](https://huggingface.co/malteos)
30
  and [Matthias Uhlig](https://huggingface.co/DRXD1000). We are your fan club.
@@ -53,8 +57,23 @@ Is the merged model alone already good? Well, of course. But it is even better w
53
  }
54
  ```
55
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
 
57
- Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
58
  * [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
59
  * [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
60
  * [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
@@ -101,7 +120,7 @@ import transformers
101
  import torch
102
 
103
  model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
104
- messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
105
 
106
  tokenizer = AutoTokenizer.from_pretrained(model)
107
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
 
19
  ---
20
 
21
  # Wiedervereinigung-7b-dpo
22
+ LoneStriker/Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2
23
  ![image/png](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b/resolve/main/Wiedervereinigung-7b.png)
24
 
25
  This is a dpo aligned merge of our favourite german models, scoring 7.11 on the mt-bench-de average.
26
  Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
27
+ Therefore the name, no nationalist ideas involved :-).
28
+
29
+ To improve result quality they are dpo-trained with a german translation of slimorca dpo using hermeo-7B for reject results.
30
+
31
+ If you are gpu-poor like me you can now use [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory) to train with german datasets.
32
 
33
  Kudos to the authors of the original models at [DiscoResearch](https://huggingface.co/DiscoResearch) and [VAGOsolutions](https://huggingface.co/VAGOsolutions), [Malte Ostendorff](https://huggingface.co/malteos)
34
  and [Matthias Uhlig](https://huggingface.co/DRXD1000). We are your fan club.
 
57
  }
58
  ```
59
 
60
+ ## Other Versions
61
+
62
+ A big thank you to [LoneStriker](https://huggingface.co/LoneStriker) for the quantized models.
63
+
64
+ | Name | Quant method | Bits |
65
+ | ---- | ---- | ---- |
66
+ [Wiedervereinigung-7b-dpo-GPTQ](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-GPTQ)| Unquantized | 16 |
67
+ [Wiedervereinigung-7b-dpo-GPTQ](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-GPTQ)| GPTQ | 4 |
68
+ [Wiedervereinigung-7b-dpo-AWQ](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-AWQ)| AWQ | 4 |
69
+ [Wiedervereinigung-7b-dpo-GGUF](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-GGUF)| GGUF | 3-8 |
70
+ [Wiedervereinigung-7b-dpo-8.0bpw-h8-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-8.0bpw-h8-exl2)| EXL2 | 8 |
71
+ [Wiedervereinigung-7b-dpo-6.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-6.0bpw-h6-exl2)| EXL2 | 6 |
72
+ [Wiedervereinigung-7b-dpo-5.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-5.0bpw-h6-exl2)| EXL2 | 5 |
73
+ [Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2)| EXL2 | 4 |
74
+ [Wiedervereinigung-7b-dpo-3.0bpw-h6-exl2](https://huggingface.co/LoneStriker/Wiedervereinigung-7b-dpo-3.0bpw-h6-exl2)| EXL2 | 3 |
75
 
76
+ Wiedervereinigung-7b is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
77
  * [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
78
  * [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
79
  * [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
 
120
  import torch
121
 
122
  model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
123
+ messages = [{"role": "user", "content": "Was ist ein deutsches Large Language Model?"}]
124
 
125
  tokenizer = AutoTokenizer.from_pretrained(model)
126
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)