Hjgugugjhuhjggg commited on
Commit
f8e35de
·
verified ·
1 Parent(s): 6519673

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
  base_model:
 
3
  - bunnycore/Llama-3.2-3B-ProdigyPlusPlus
 
4
  - chuanli11/Llama-3.2-3B-Instruct-uncensored
5
- - meta-llama/Llama-3.2-3B-Instruct
6
  - huihui-ai/Llama-3.2-3B-Instruct-abliterated
7
- - meta-llama/Llama-3.2-3B
8
  library_name: transformers
9
  tags:
10
  - mergekit
@@ -18,15 +18,15 @@ This is a merge of pre-trained language models created using [mergekit](https://
18
  ## Merge Details
19
  ### Merge Method
20
 
21
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [bunnycore/Llama-3.2-3B-ProdigyPlusPlus](https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus) as a base.
22
 
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
26
- * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
27
  * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
28
- * [huihui-ai/Llama-3.2-3B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.2-3B-Instruct-abliterated)
29
  * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
 
30
 
31
  ### Configuration
32
 
@@ -34,18 +34,14 @@ The following YAML configuration was used to produce this model:
34
 
35
  ```yaml
36
  base_model:
37
- model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
38
- merge_method: ties
 
 
 
39
  merge_method_sequence:
40
  - dare_ties
41
- - slerp
42
  - ties
43
- models:
44
- - model: meta-llama/Llama-3.2-3B-Instruct
45
- - model: meta-llama/Llama-3.2-3B
46
- - model: chuanli11/Llama-3.2-3B-Instruct-uncensored
47
- - model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
48
- - model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
49
  parameters:
50
  batch_size: 32
51
  density: 0.5
@@ -72,22 +68,38 @@ parameters:
72
  - 0
73
  - value: 0.5
74
  weight: 0.5
75
- slice:
76
- - models:
77
- - meta-llama/Llama-3.2-3B-Instruct
78
- - bunnycore/Llama-3.2-3B-ProdigyPlusPlus
79
- - models:
80
- - meta-llama/Llama-3.2-3B-Instruct
81
- - meta-llama/Llama-3.2-3B
82
- - models:
83
- - chuanli11/Llama-3.2-3B-Instruct-uncensored
84
- - huihui-ai/Llama-3.2-3B-Instruct-abliterated
85
- sources:
86
- - model: meta-llama/Llama-3.2-3B-Instruct
87
- - model: meta-llama/Llama-3.2-3B
88
- - model: chuanli11/Llama-3.2-3B-Instruct-uncensored
89
- - model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
90
- - model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
  tokenizer_source: union
92
 
93
  ```
 
1
  ---
2
  base_model:
3
+ - meta-llama/Llama-3.2-3B-Instruct
4
  - bunnycore/Llama-3.2-3B-ProdigyPlusPlus
5
+ - meta-llama/Llama-3.2-3B
6
  - chuanli11/Llama-3.2-3B-Instruct-uncensored
 
7
  - huihui-ai/Llama-3.2-3B-Instruct-abliterated
 
8
  library_name: transformers
9
  tags:
10
  - mergekit
 
18
  ## Merge Details
19
  ### Merge Method
20
 
21
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [huihui-ai/Llama-3.2-3B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.2-3B-Instruct-abliterated) as a base.
22
 
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
 
26
  * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
27
+ * [bunnycore/Llama-3.2-3B-ProdigyPlusPlus](https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus)
28
  * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
29
+ * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
30
 
31
  ### Configuration
32
 
 
34
 
35
  ```yaml
36
  base_model:
37
+ model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
38
+ layer_range:
39
+ - 0
40
+ - 28
41
+ merge_method: dare_ties
42
  merge_method_sequence:
43
  - dare_ties
 
44
  - ties
 
 
 
 
 
 
45
  parameters:
46
  batch_size: 32
47
  density: 0.5
 
68
  - 0
69
  - value: 0.5
70
  weight: 0.5
71
+ slices:
72
+ - sources:
73
+ - density: 0.5
74
+ layer_range:
75
+ - 0
76
+ - 28
77
+ model: meta-llama/Llama-3.2-3B-Instruct
78
+ weight: 0.5
79
+ - density: 0.5
80
+ layer_range:
81
+ - 0
82
+ - 28
83
+ model: meta-llama/Llama-3.2-3B
84
+ weight: 0.5
85
+ - density: 0.5
86
+ layer_range:
87
+ - 0
88
+ - 28
89
+ model: chuanli11/Llama-3.2-3B-Instruct-uncensored
90
+ weight: 0.5
91
+ - density: 0.5
92
+ layer_range:
93
+ - 0
94
+ - 28
95
+ model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
96
+ weight: 0.5
97
+ - density: 0.5
98
+ layer_range:
99
+ - 0
100
+ - 28
101
+ model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
102
+ weight: 0.5
103
  tokenizer_source: union
104
 
105
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bunnycore/Llama-3.2-3B-ProdigyPlusPlus",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -33,7 +33,7 @@
33
  },
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
- "torch_dtype": "float16",
37
  "transformers_version": "4.45.1",
38
  "use_cache": true,
39
  "vocab_size": 128256
 
1
  {
2
+ "_name_or_path": "huihui-ai/Llama-3.2-3B-Instruct-abliterated",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
33
  },
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
+ "torch_dtype": "bfloat16",
37
  "transformers_version": "4.45.1",
38
  "use_cache": true,
39
  "vocab_size": 128256
mergekit_config.yml CHANGED
@@ -1,16 +1,12 @@
1
  base_model:
2
- model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
3
- merge_method: ties
 
 
 
4
  merge_method_sequence:
5
  - dare_ties
6
- - slerp
7
  - ties
8
- models:
9
- - model: meta-llama/Llama-3.2-3B-Instruct
10
- - model: meta-llama/Llama-3.2-3B
11
- - model: chuanli11/Llama-3.2-3B-Instruct-uncensored
12
- - model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
13
- - model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
14
  parameters:
15
  batch_size: 32
16
  density: 0.5
@@ -37,20 +33,36 @@ parameters:
37
  - 0
38
  - value: 0.5
39
  weight: 0.5
40
- slice:
41
- - models:
42
- - meta-llama/Llama-3.2-3B-Instruct
43
- - bunnycore/Llama-3.2-3B-ProdigyPlusPlus
44
- - models:
45
- - meta-llama/Llama-3.2-3B-Instruct
46
- - meta-llama/Llama-3.2-3B
47
- - models:
48
- - chuanli11/Llama-3.2-3B-Instruct-uncensored
49
- - huihui-ai/Llama-3.2-3B-Instruct-abliterated
50
- sources:
51
- - model: meta-llama/Llama-3.2-3B-Instruct
52
- - model: meta-llama/Llama-3.2-3B
53
- - model: chuanli11/Llama-3.2-3B-Instruct-uncensored
54
- - model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
55
- - model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
  tokenizer_source: union
 
1
  base_model:
2
+ model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
3
+ layer_range:
4
+ - 0
5
+ - 28
6
+ merge_method: dare_ties
7
  merge_method_sequence:
8
  - dare_ties
 
9
  - ties
 
 
 
 
 
 
10
  parameters:
11
  batch_size: 32
12
  density: 0.5
 
33
  - 0
34
  - value: 0.5
35
  weight: 0.5
36
+ slices:
37
+ - sources:
38
+ - density: 0.5
39
+ layer_range:
40
+ - 0
41
+ - 28
42
+ model: meta-llama/Llama-3.2-3B-Instruct
43
+ weight: 0.5
44
+ - density: 0.5
45
+ layer_range:
46
+ - 0
47
+ - 28
48
+ model: meta-llama/Llama-3.2-3B
49
+ weight: 0.5
50
+ - density: 0.5
51
+ layer_range:
52
+ - 0
53
+ - 28
54
+ model: chuanli11/Llama-3.2-3B-Instruct-uncensored
55
+ weight: 0.5
56
+ - density: 0.5
57
+ layer_range:
58
+ - 0
59
+ - 28
60
+ model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
61
+ weight: 0.5
62
+ - density: 0.5
63
+ layer_range:
64
+ - 0
65
+ - 28
66
+ model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
67
+ weight: 0.5
68
  tokenizer_source: union
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a912a44fc0022a36e49beb59c5a6f2d3db36dcb5803079caab04900f2854c167
3
- size 4998794808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:890b30c2fcdc8d81e207bcad23b051bb66cb68123b37b057e6f065223669b88a
3
+ size 4998794968
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:186db3644ffe2e65b859971d8f6b06677e7078da924701171e8ba81a486fe336
3
- size 2214738976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd563d25a47422404bd7102fb0c065c572032403559e7ce84fa64c0654469ed4
3
+ size 2214739072