Hjgugugjhuhjggg commited on
Commit
34194bc
·
verified ·
1 Parent(s): f2f826f

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
  base_model:
3
- - bunnycore/Llama-3.2-3B-ProdigyPlusPlus
4
  - huihui-ai/Llama-3.2-3B-Instruct-abliterated
5
- - chuanli11/Llama-3.2-3B-Instruct-uncensored
6
  - meta-llama/Llama-3.2-3B-Instruct
 
7
  - meta-llama/Llama-3.2-3B
 
8
  library_name: transformers
9
  tags:
10
  - mergekit
@@ -18,14 +18,14 @@ This is a merge of pre-trained language models created using [mergekit](https://
18
  ## Merge Details
19
  ### Merge Method
20
 
21
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored) as a base.
22
 
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
26
- * [bunnycore/Llama-3.2-3B-ProdigyPlusPlus](https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus)
27
  * [huihui-ai/Llama-3.2-3B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.2-3B-Instruct-abliterated)
28
  * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
 
29
  * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
30
 
31
  ### Configuration
@@ -34,72 +34,23 @@ The following YAML configuration was used to produce this model:
34
 
35
  ```yaml
36
  base_model:
37
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
38
- layer_range:
39
- - 0
40
- - 28
41
- merge_method: ties
42
  merge_method_sequence:
43
  - dare_ties
44
  - ties
 
 
 
 
 
 
45
  parameters:
46
- batch_size: 32
47
  density: 0.5
48
  int8_mask: true
49
- layer_range:
50
- - 0
51
- - 28
52
- model.embed_tokens.weight.t: 1.0
53
  normalize: false
54
- t:
55
- - filter: self_attn
56
- value:
57
- - 0
58
- - 0.5
59
- - 0.3
60
- - 0.7
61
- - 1
62
- - filter: mlp
63
- value:
64
- - 1
65
- - 0.5
66
- - 0.7
67
- - 0.3
68
- - 0
69
- - value: 0.5
70
  weight: 0.5
71
- slices:
72
- - sources:
73
- - density: 0.5
74
- layer_range:
75
- - 0
76
- - 28
77
- model: meta-llama/Llama-3.2-3B-Instruct
78
- weight: 0.5
79
- - density: 0.5
80
- layer_range:
81
- - 0
82
- - 28
83
- model: meta-llama/Llama-3.2-3B
84
- weight: 0.5
85
- - density: 0.5
86
- layer_range:
87
- - 0
88
- - 28
89
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
90
- weight: 0.5
91
- - density: 0.5
92
- layer_range:
93
- - 0
94
- - 28
95
- model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
96
- weight: 0.5
97
- - density: 0.5
98
- layer_range:
99
- - 0
100
- - 28
101
- model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
102
- weight: 0.5
103
  tokenizer_source: union
104
 
105
  ```
 
1
  ---
2
  base_model:
 
3
  - huihui-ai/Llama-3.2-3B-Instruct-abliterated
 
4
  - meta-llama/Llama-3.2-3B-Instruct
5
+ - chuanli11/Llama-3.2-3B-Instruct-uncensored
6
  - meta-llama/Llama-3.2-3B
7
+ - bunnycore/Llama-3.2-3B-ProdigyPlusPlus
8
  library_name: transformers
9
  tags:
10
  - mergekit
 
18
  ## Merge Details
19
  ### Merge Method
20
 
21
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [bunnycore/Llama-3.2-3B-ProdigyPlusPlus](https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus) as a base.
22
 
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
 
26
  * [huihui-ai/Llama-3.2-3B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.2-3B-Instruct-abliterated)
27
  * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
28
+ * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
29
  * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
30
 
31
  ### Configuration
 
34
 
35
  ```yaml
36
  base_model:
37
+ model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
38
+ dtype: float16
39
+ merge_method: dare_ties
 
 
40
  merge_method_sequence:
41
  - dare_ties
42
  - ties
43
+ models:
44
+ - model: meta-llama/Llama-3.2-3B-Instruct
45
+ - model: meta-llama/Llama-3.2-3B
46
+ - model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
47
+ - model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
48
+ - model: chuanli11/Llama-3.2-3B-Instruct-uncensored
49
  parameters:
 
50
  density: 0.5
51
  int8_mask: true
 
 
 
 
52
  normalize: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  weight: 0.5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
  tokenizer_source: union
55
 
56
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "chuanli11/Llama-3.2-3B-Instruct-uncensored",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -33,7 +33,7 @@
33
  },
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
- "torch_dtype": "bfloat16",
37
  "transformers_version": "4.45.1",
38
  "use_cache": true,
39
  "vocab_size": 128256
 
1
  {
2
+ "_name_or_path": "bunnycore/Llama-3.2-3B-ProdigyPlusPlus",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
33
  },
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
+ "torch_dtype": "float16",
37
  "transformers_version": "4.45.1",
38
  "use_cache": true,
39
  "vocab_size": 128256
mergekit_config.yml CHANGED
@@ -1,68 +1,19 @@
1
  base_model:
2
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
3
- layer_range:
4
- - 0
5
- - 28
6
- merge_method: ties
7
  merge_method_sequence:
8
  - dare_ties
9
  - ties
 
 
 
 
 
 
10
  parameters:
11
- batch_size: 32
12
  density: 0.5
13
  int8_mask: true
14
- layer_range:
15
- - 0
16
- - 28
17
- model.embed_tokens.weight.t: 1.0
18
  normalize: false
19
- t:
20
- - filter: self_attn
21
- value:
22
- - 0
23
- - 0.5
24
- - 0.3
25
- - 0.7
26
- - 1
27
- - filter: mlp
28
- value:
29
- - 1
30
- - 0.5
31
- - 0.7
32
- - 0.3
33
- - 0
34
- - value: 0.5
35
  weight: 0.5
36
- slices:
37
- - sources:
38
- - density: 0.5
39
- layer_range:
40
- - 0
41
- - 28
42
- model: meta-llama/Llama-3.2-3B-Instruct
43
- weight: 0.5
44
- - density: 0.5
45
- layer_range:
46
- - 0
47
- - 28
48
- model: meta-llama/Llama-3.2-3B
49
- weight: 0.5
50
- - density: 0.5
51
- layer_range:
52
- - 0
53
- - 28
54
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
55
- weight: 0.5
56
- - density: 0.5
57
- layer_range:
58
- - 0
59
- - 28
60
- model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
61
- weight: 0.5
62
- - density: 0.5
63
- layer_range:
64
- - 0
65
- - 28
66
- model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
67
- weight: 0.5
68
  tokenizer_source: union
 
1
  base_model:
2
+ model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
3
+ dtype: float16
4
+ merge_method: dare_ties
 
 
5
  merge_method_sequence:
6
  - dare_ties
7
  - ties
8
+ models:
9
+ - model: meta-llama/Llama-3.2-3B-Instruct
10
+ - model: meta-llama/Llama-3.2-3B
11
+ - model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
12
+ - model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
13
+ - model: chuanli11/Llama-3.2-3B-Instruct-uncensored
14
  parameters:
 
15
  density: 0.5
16
  int8_mask: true
 
 
 
 
17
  normalize: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  weight: 0.5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  tokenizer_source: union
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28f7de1dd4d1c949758eb83255f3bfdbae8d108cef7c956d784822f2e53671f4
3
- size 4998794944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:866455a0ace665a9134bb7ee83c04d47aee612e8c565a07098d1517f22539580
3
+ size 4998794808
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3b20e109c5d1a0f6c42f1b2d003432b4687bc2cd0844225e13d2436e29af335
3
- size 2214739072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7de1213796b137872b5bee15617d3cb6d175b112f07d3c5c95bc4fc077dedf6
3
+ size 2214738976
special_tokens_map.json CHANGED
@@ -12,5 +12,12 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
 
 
 
 
 
 
 
15
  }
16
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|eot_id|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
  }
23
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
- size 17209920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65ff5472d095ccd9332d9e723153d7bc7226cb6be9c1bffda738b5ba2e71bf26
3
+ size 17210084
tokenizer_config.json CHANGED
@@ -2053,10 +2053,15 @@
2053
  "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
 
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
 
 
 
 
2061
  "tokenizer_class": "PreTrainedTokenizerFast"
2062
  }
 
2053
  "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
+ "max_length": null,
2057
  "model_input_names": [
2058
  "input_ids",
2059
  "attention_mask"
2060
  ],
2061
  "model_max_length": 131072,
2062
+ "pad_to_multiple_of": null,
2063
+ "pad_token": "<|eot_id|>",
2064
+ "pad_token_type_id": 0,
2065
+ "padding_side": "left",
2066
  "tokenizer_class": "PreTrainedTokenizerFast"
2067
  }