Upload folder using huggingface_hub
Browse files- README.md +13 -62
- config.json +2 -2
- mergekit_config.yml +9 -58
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- special_tokens_map.json +7 -0
- tokenizer.json +2 -2
- tokenizer_config.json +5 -0
README.md
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
---
|
2 |
base_model:
|
3 |
-
- bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
4 |
- huihui-ai/Llama-3.2-3B-Instruct-abliterated
|
5 |
-
- chuanli11/Llama-3.2-3B-Instruct-uncensored
|
6 |
- meta-llama/Llama-3.2-3B-Instruct
|
|
|
7 |
- meta-llama/Llama-3.2-3B
|
|
|
8 |
library_name: transformers
|
9 |
tags:
|
10 |
- mergekit
|
@@ -18,14 +18,14 @@ This is a merge of pre-trained language models created using [mergekit](https://
|
|
18 |
## Merge Details
|
19 |
### Merge Method
|
20 |
|
21 |
-
This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [
|
22 |
|
23 |
### Models Merged
|
24 |
|
25 |
The following models were included in the merge:
|
26 |
-
* [bunnycore/Llama-3.2-3B-ProdigyPlusPlus](https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus)
|
27 |
* [huihui-ai/Llama-3.2-3B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.2-3B-Instruct-abliterated)
|
28 |
* [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
|
|
|
29 |
* [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
|
30 |
|
31 |
### Configuration
|
@@ -34,72 +34,23 @@ The following YAML configuration was used to produce this model:
|
|
34 |
|
35 |
```yaml
|
36 |
base_model:
|
37 |
-
model:
|
38 |
-
|
39 |
-
|
40 |
-
- 28
|
41 |
-
merge_method: ties
|
42 |
merge_method_sequence:
|
43 |
- dare_ties
|
44 |
- ties
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
parameters:
|
46 |
-
batch_size: 32
|
47 |
density: 0.5
|
48 |
int8_mask: true
|
49 |
-
layer_range:
|
50 |
-
- 0
|
51 |
-
- 28
|
52 |
-
model.embed_tokens.weight.t: 1.0
|
53 |
normalize: false
|
54 |
-
t:
|
55 |
-
- filter: self_attn
|
56 |
-
value:
|
57 |
-
- 0
|
58 |
-
- 0.5
|
59 |
-
- 0.3
|
60 |
-
- 0.7
|
61 |
-
- 1
|
62 |
-
- filter: mlp
|
63 |
-
value:
|
64 |
-
- 1
|
65 |
-
- 0.5
|
66 |
-
- 0.7
|
67 |
-
- 0.3
|
68 |
-
- 0
|
69 |
-
- value: 0.5
|
70 |
weight: 0.5
|
71 |
-
slices:
|
72 |
-
- sources:
|
73 |
-
- density: 0.5
|
74 |
-
layer_range:
|
75 |
-
- 0
|
76 |
-
- 28
|
77 |
-
model: meta-llama/Llama-3.2-3B-Instruct
|
78 |
-
weight: 0.5
|
79 |
-
- density: 0.5
|
80 |
-
layer_range:
|
81 |
-
- 0
|
82 |
-
- 28
|
83 |
-
model: meta-llama/Llama-3.2-3B
|
84 |
-
weight: 0.5
|
85 |
-
- density: 0.5
|
86 |
-
layer_range:
|
87 |
-
- 0
|
88 |
-
- 28
|
89 |
-
model: chuanli11/Llama-3.2-3B-Instruct-uncensored
|
90 |
-
weight: 0.5
|
91 |
-
- density: 0.5
|
92 |
-
layer_range:
|
93 |
-
- 0
|
94 |
-
- 28
|
95 |
-
model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
|
96 |
-
weight: 0.5
|
97 |
-
- density: 0.5
|
98 |
-
layer_range:
|
99 |
-
- 0
|
100 |
-
- 28
|
101 |
-
model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
102 |
-
weight: 0.5
|
103 |
tokenizer_source: union
|
104 |
|
105 |
```
|
|
|
1 |
---
|
2 |
base_model:
|
|
|
3 |
- huihui-ai/Llama-3.2-3B-Instruct-abliterated
|
|
|
4 |
- meta-llama/Llama-3.2-3B-Instruct
|
5 |
+
- chuanli11/Llama-3.2-3B-Instruct-uncensored
|
6 |
- meta-llama/Llama-3.2-3B
|
7 |
+
- bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
8 |
library_name: transformers
|
9 |
tags:
|
10 |
- mergekit
|
|
|
18 |
## Merge Details
|
19 |
### Merge Method
|
20 |
|
21 |
+
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [bunnycore/Llama-3.2-3B-ProdigyPlusPlus](https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus) as a base.
|
22 |
|
23 |
### Models Merged
|
24 |
|
25 |
The following models were included in the merge:
|
|
|
26 |
* [huihui-ai/Llama-3.2-3B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.2-3B-Instruct-abliterated)
|
27 |
* [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
|
28 |
+
* [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
|
29 |
* [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
|
30 |
|
31 |
### Configuration
|
|
|
34 |
|
35 |
```yaml
|
36 |
base_model:
|
37 |
+
model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
38 |
+
dtype: float16
|
39 |
+
merge_method: dare_ties
|
|
|
|
|
40 |
merge_method_sequence:
|
41 |
- dare_ties
|
42 |
- ties
|
43 |
+
models:
|
44 |
+
- model: meta-llama/Llama-3.2-3B-Instruct
|
45 |
+
- model: meta-llama/Llama-3.2-3B
|
46 |
+
- model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
|
47 |
+
- model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
48 |
+
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
|
49 |
parameters:
|
|
|
50 |
density: 0.5
|
51 |
int8_mask: true
|
|
|
|
|
|
|
|
|
52 |
normalize: false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
53 |
weight: 0.5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
54 |
tokenizer_source: union
|
55 |
|
56 |
```
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -33,7 +33,7 @@
|
|
33 |
},
|
34 |
"rope_theta": 500000.0,
|
35 |
"tie_word_embeddings": true,
|
36 |
-
"torch_dtype": "
|
37 |
"transformers_version": "4.45.1",
|
38 |
"use_cache": true,
|
39 |
"vocab_size": 128256
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "bunnycore/Llama-3.2-3B-ProdigyPlusPlus",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
33 |
},
|
34 |
"rope_theta": 500000.0,
|
35 |
"tie_word_embeddings": true,
|
36 |
+
"torch_dtype": "float16",
|
37 |
"transformers_version": "4.45.1",
|
38 |
"use_cache": true,
|
39 |
"vocab_size": 128256
|
mergekit_config.yml
CHANGED
@@ -1,68 +1,19 @@
|
|
1 |
base_model:
|
2 |
-
model:
|
3 |
-
|
4 |
-
|
5 |
-
- 28
|
6 |
-
merge_method: ties
|
7 |
merge_method_sequence:
|
8 |
- dare_ties
|
9 |
- ties
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
parameters:
|
11 |
-
batch_size: 32
|
12 |
density: 0.5
|
13 |
int8_mask: true
|
14 |
-
layer_range:
|
15 |
-
- 0
|
16 |
-
- 28
|
17 |
-
model.embed_tokens.weight.t: 1.0
|
18 |
normalize: false
|
19 |
-
t:
|
20 |
-
- filter: self_attn
|
21 |
-
value:
|
22 |
-
- 0
|
23 |
-
- 0.5
|
24 |
-
- 0.3
|
25 |
-
- 0.7
|
26 |
-
- 1
|
27 |
-
- filter: mlp
|
28 |
-
value:
|
29 |
-
- 1
|
30 |
-
- 0.5
|
31 |
-
- 0.7
|
32 |
-
- 0.3
|
33 |
-
- 0
|
34 |
-
- value: 0.5
|
35 |
weight: 0.5
|
36 |
-
slices:
|
37 |
-
- sources:
|
38 |
-
- density: 0.5
|
39 |
-
layer_range:
|
40 |
-
- 0
|
41 |
-
- 28
|
42 |
-
model: meta-llama/Llama-3.2-3B-Instruct
|
43 |
-
weight: 0.5
|
44 |
-
- density: 0.5
|
45 |
-
layer_range:
|
46 |
-
- 0
|
47 |
-
- 28
|
48 |
-
model: meta-llama/Llama-3.2-3B
|
49 |
-
weight: 0.5
|
50 |
-
- density: 0.5
|
51 |
-
layer_range:
|
52 |
-
- 0
|
53 |
-
- 28
|
54 |
-
model: chuanli11/Llama-3.2-3B-Instruct-uncensored
|
55 |
-
weight: 0.5
|
56 |
-
- density: 0.5
|
57 |
-
layer_range:
|
58 |
-
- 0
|
59 |
-
- 28
|
60 |
-
model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
|
61 |
-
weight: 0.5
|
62 |
-
- density: 0.5
|
63 |
-
layer_range:
|
64 |
-
- 0
|
65 |
-
- 28
|
66 |
-
model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
67 |
-
weight: 0.5
|
68 |
tokenizer_source: union
|
|
|
1 |
base_model:
|
2 |
+
model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
3 |
+
dtype: float16
|
4 |
+
merge_method: dare_ties
|
|
|
|
|
5 |
merge_method_sequence:
|
6 |
- dare_ties
|
7 |
- ties
|
8 |
+
models:
|
9 |
+
- model: meta-llama/Llama-3.2-3B-Instruct
|
10 |
+
- model: meta-llama/Llama-3.2-3B
|
11 |
+
- model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
|
12 |
+
- model: bunnycore/Llama-3.2-3B-ProdigyPlusPlus
|
13 |
+
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
|
14 |
parameters:
|
|
|
15 |
density: 0.5
|
16 |
int8_mask: true
|
|
|
|
|
|
|
|
|
17 |
normalize: false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
weight: 0.5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
tokenizer_source: union
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:866455a0ace665a9134bb7ee83c04d47aee612e8c565a07098d1517f22539580
|
3 |
+
size 4998794808
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7de1213796b137872b5bee15617d3cb6d175b112f07d3c5c95bc4fc077dedf6
|
3 |
+
size 2214738976
|
special_tokens_map.json
CHANGED
@@ -12,5 +12,12 @@
|
|
12 |
"normalized": false,
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
}
|
16 |
}
|
|
|
12 |
"normalized": false,
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<|eot_id|>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
}
|
23 |
}
|
tokenizer.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65ff5472d095ccd9332d9e723153d7bc7226cb6be9c1bffda738b5ba2e71bf26
|
3 |
+
size 17210084
|
tokenizer_config.json
CHANGED
@@ -2053,10 +2053,15 @@
|
|
2053 |
"chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
|
|
2056 |
"model_input_names": [
|
2057 |
"input_ids",
|
2058 |
"attention_mask"
|
2059 |
],
|
2060 |
"model_max_length": 131072,
|
|
|
|
|
|
|
|
|
2061 |
"tokenizer_class": "PreTrainedTokenizerFast"
|
2062 |
}
|
|
|
2053 |
"chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
2056 |
+
"max_length": null,
|
2057 |
"model_input_names": [
|
2058 |
"input_ids",
|
2059 |
"attention_mask"
|
2060 |
],
|
2061 |
"model_max_length": 131072,
|
2062 |
+
"pad_to_multiple_of": null,
|
2063 |
+
"pad_token": "<|eot_id|>",
|
2064 |
+
"pad_token_type_id": 0,
|
2065 |
+
"padding_side": "left",
|
2066 |
"tokenizer_class": "PreTrainedTokenizerFast"
|
2067 |
}
|