cstr commited on
Commit
34325e8
·
verified ·
1 Parent(s): 8e3d75e

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +15 -18
README.md CHANGED
@@ -3,38 +3,35 @@ tags:
3
  - merge
4
  - mergekit
5
  - lazymergekit
6
- - mlabonne/OrpoLlama-3-8B
7
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
8
  base_model:
9
- - mlabonne/OrpoLlama-3-8B
10
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
11
  ---
12
 
13
  # llama3-discolm-orpo-t2
14
 
15
  llama3-discolm-orpo-t2 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
- * [mlabonne/OrpoLlama-3-8B](https://huggingface.co/mlabonne/OrpoLlama-3-8B)
17
  * [DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental](https://huggingface.co/DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental)
18
 
19
  ## 🧩 Configuration
20
 
21
  ```yaml
22
- slices:
23
- - sources:
24
- - model: mlabonne/OrpoLlama-3-8B
25
- layer_range: [0, 32]
26
- - model: DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
27
- layer_range: [0, 32]
28
- merge_method: slerp
29
- base_model: mlabonne/OrpoLlama-3-8B
30
- parameters:
31
- t:
32
- - filter: self_attn
33
- value: [1, 0.7, 0.5, 0.3, 0.1]
34
- - filter: mlp
35
- value: [0, 0.3, 0.5, 0.7, 0.9]
36
- - value: 0.5
37
  dtype: bfloat16
 
38
  ```
39
 
40
  ## 💻 Usage
 
3
  - merge
4
  - mergekit
5
  - lazymergekit
6
+ - meta-llama/Meta-Llama-3-8B
7
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
8
  base_model:
9
+ - meta-llama/Meta-Llama-3-8B
10
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
11
  ---
12
 
13
  # llama3-discolm-orpo-t2
14
 
15
  llama3-discolm-orpo-t2 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
+ * [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
17
  * [DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental](https://huggingface.co/DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental)
18
 
19
  ## 🧩 Configuration
20
 
21
  ```yaml
22
+ models:
23
+ - layer_range: [0, 40]
24
+ model: meta-llama/Meta-Llama-3-8B
25
+ parameters:
26
+ weight: 0.2
27
+ - layer_range: [0, 40]
28
+ model: DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
29
+ parameters:
30
+ weight: 0.8
31
+ merge_method: task_arithmetic
32
+ base_model: meta-llama/Meta-Llama-3-8B
 
 
 
 
33
  dtype: bfloat16
34
+ random_seed: 0
35
  ```
36
 
37
  ## 💻 Usage