Add files using upload-large-folder tool
Browse files- README.md +51 -29
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +41 -41
- tokenizer_config.json +1 -0
README.md
CHANGED
@@ -1,4 +1,19 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
base_model: mlx-community/Llama-3.2-3B-Instruct
|
3 |
language:
|
4 |
- en
|
@@ -12,14 +27,6 @@ language:
|
|
12 |
library_name: transformers
|
13 |
license: llama3.2
|
14 |
pipeline_tag: text-generation
|
15 |
-
tags:
|
16 |
-
- facebook
|
17 |
-
- meta
|
18 |
-
- pytorch
|
19 |
-
- llama
|
20 |
-
- llama-3
|
21 |
-
- mlx
|
22 |
-
- mlx
|
23 |
extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Version\
|
24 |
\ Release Date: September 25, 2024\n\n“Agreement” means the terms and conditions\
|
25 |
\ for use, reproduction, distribution and modification of the Llama Materials set\
|
@@ -146,7 +153,7 @@ extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Vers
|
|
146 |
\ overburden, interfere with or impair the proper working, integrity, operation\
|
147 |
\ or appearance of a website or computer system\n 7. Engage in any action, or\
|
148 |
\ facilitate any action, to intentionally circumvent or remove usage restrictions\
|
149 |
-
\ or other safety measures, or to enable functionality disabled by Meta
|
150 |
\ in, promote, incite, facilitate, or assist in the planning or development of activities\
|
151 |
\ that present a risk of death or bodily harm to individuals, including use of Llama\
|
152 |
\ 3.2 related to the following:\n 8. Military, warfare, nuclear industries or\
|
@@ -167,7 +174,7 @@ extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Vers
|
|
167 |
\ without consent, authorization, or legal right\n 18. Representing that the\
|
168 |
\ use of Llama 3.2 or outputs are human-generated\n 19. Generating or facilitating\
|
169 |
\ false online engagement, including fake reviews and other means of fake online\
|
170 |
-
\ engagement
|
171 |
\ of your AI system 5. Interact with third party tools, models, or software designed\
|
172 |
\ to generate unlawful content or engage in unlawful or harmful conduct and/or represent\
|
173 |
\ that the outputs of such tools, models, or software are associated with Meta or\
|
@@ -207,31 +214,46 @@ extra_gated_description: The information you provide will be collected, stored,
|
|
207 |
and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
|
208 |
extra_gated_button_content: Submit
|
209 |
---
|
|
|
|
|
|
|
|
|
|
|
210 |
|
211 |
-
|
|
|
212 |
|
213 |
-
|
214 |
-
|
215 |
-
|
|
|
216 |
|
217 |
-
##
|
|
|
|
|
|
|
218 |
|
219 |
-
|
220 |
-
|
221 |
-
|
|
|
222 |
|
223 |
-
|
224 |
-
|
|
|
|
|
225 |
|
226 |
-
|
|
|
|
|
227 |
|
228 |
-
|
|
|
|
|
229 |
|
230 |
-
|
231 |
-
|
232 |
-
prompt = tokenizer.apply_chat_template(
|
233 |
-
messages, add_generation_prompt=True
|
234 |
-
)
|
235 |
|
236 |
-
|
237 |
-
|
|
|
|
1 |
---
|
2 |
+
tags:
|
3 |
+
- facebook
|
4 |
+
- meta
|
5 |
+
- pytorch
|
6 |
+
- llama
|
7 |
+
- llama-3
|
8 |
+
- mlx
|
9 |
+
- mlx
|
10 |
+
- reasoning
|
11 |
+
- llama
|
12 |
+
- deepseek
|
13 |
+
- ollama
|
14 |
+
- chain-of-thoughts
|
15 |
+
- small-llm
|
16 |
+
- edge
|
17 |
base_model: mlx-community/Llama-3.2-3B-Instruct
|
18 |
language:
|
19 |
- en
|
|
|
27 |
library_name: transformers
|
28 |
license: llama3.2
|
29 |
pipeline_tag: text-generation
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Version\
|
31 |
\ Release Date: September 25, 2024\n\n“Agreement” means the terms and conditions\
|
32 |
\ for use, reproduction, distribution and modification of the Llama Materials set\
|
|
|
153 |
\ overburden, interfere with or impair the proper working, integrity, operation\
|
154 |
\ or appearance of a website or computer system\n 7. Engage in any action, or\
|
155 |
\ facilitate any action, to intentionally circumvent or remove usage restrictions\
|
156 |
+
\ or other safety measures, or to enable functionality disabled by Meta \n2. Engage\
|
157 |
\ in, promote, incite, facilitate, or assist in the planning or development of activities\
|
158 |
\ that present a risk of death or bodily harm to individuals, including use of Llama\
|
159 |
\ 3.2 related to the following:\n 8. Military, warfare, nuclear industries or\
|
|
|
174 |
\ without consent, authorization, or legal right\n 18. Representing that the\
|
175 |
\ use of Llama 3.2 or outputs are human-generated\n 19. Generating or facilitating\
|
176 |
\ false online engagement, including fake reviews and other means of fake online\
|
177 |
+
\ engagement \n4. Fail to appropriately disclose to end users any known dangers\
|
178 |
\ of your AI system 5. Interact with third party tools, models, or software designed\
|
179 |
\ to generate unlawful content or engage in unlawful or harmful conduct and/or represent\
|
180 |
\ that the outputs of such tools, models, or software are associated with Meta or\
|
|
|
214 |
and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
|
215 |
extra_gated_button_content: Submit
|
216 |
---
|
217 |
+
# ReasonableLlama-3B: A Fine-Tuned Reasoning Model
|
218 |
+
|
219 |
+
HF: https://huggingface.co/adeelahmad/ReasonableLlama3-3B-Jr
|
220 |
+
Ollama: https://ollama.com/adeelahmad/ReasonableLLAMA-Jr-3b
|
221 |
+
|
222 |
|
223 |
+
Welcome to **ReasonableLlama-3B**, a cutting-edge reasoning model built on the foundation of LLaMA-3B. This model has been carefully fine-tuned to enhance its capabilities in logical
|
224 |
+
thinking, problem-solving, and creative analysis.
|
225 |
|
226 |
+
## Overview
|
227 |
+
- **Model Name**: ReasonableLlama-3B
|
228 |
+
- **Base Architecture**: LLaMA-3B (Large Language Model with 3B parameters)
|
229 |
+
- **Purpose**: Designed for tasks requiring advanced reasoning, problem-solving, and creative thinking
|
230 |
|
231 |
+
## Features
|
232 |
+
- **Advanced Reasoning**: Excels in logical analysis, problem-solving, and decision-making.
|
233 |
+
- **Creative Thinking**: Generates innovative solutions and ideas.
|
234 |
+
- **Curriculum-Based Fine-Tuning**: Trained on high-quality datasets to enhance reasoning abilities.
|
235 |
|
236 |
+
## Technical Details
|
237 |
+
- **Parameter Count**: 3B parameters
|
238 |
+
- **Training Process**: Fine-tuned using state-of-the-art techniques for reasoning tasks
|
239 |
+
- **Specialization**: Optimized for specific reasoning workflows and scenarios
|
240 |
|
241 |
+
## Use Cases
|
242 |
+
- **Research**: Facilitates complex problem-solving and theoretical analysis.
|
243 |
+
- **Education**: Assists in creating educational examples and problem sets.
|
244 |
+
- **Problem Solving**: Helps generate innovative solutions across various domains.
|
245 |
|
246 |
+
## Installation and Usage
|
247 |
+
- **Integration**: Can be integrated into existing systems via APIs or local setup.
|
248 |
+
- **Inputs**: Supports text and images, leveraging Ollama's versatile capabilities.
|
249 |
|
250 |
+
## Limitations
|
251 |
+
- **Scope**: Limited to single-step reasoning; multi-hop reasoning is a current focus area.
|
252 |
+
- **Data Bias**: Caution with dataset provenance as it may reflect historical biases.
|
253 |
|
254 |
+
## Contributing
|
255 |
+
Contributions welcome! Fork the project, submit issues, and pull requests on GitHub. Your insights can help shape future improvements.
|
|
|
|
|
|
|
256 |
|
257 |
+
## Citations
|
258 |
+
- Special thanks to LLaMA's developers for providing a strong foundation.
|
259 |
+
- Acknowledgments to the community contributing to open-source AI advancements.
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ff29640561515b67419537c05cc772061ccbcbd64a91a9253a7e26340e08cb4
|
3 |
+
size 5368478677
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06086cfa2c959311846c8f528db9264fccbf4fc389c236147ea5b6bfc878ef67
|
3 |
+
size 1057050040
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "model-00001-of-00002.safetensors",
|
@@ -94,24 +94,24 @@
|
|
94 |
"model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
95 |
"model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
96 |
"model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
97 |
-
"model.layers.18.input_layernorm.weight": "model-
|
98 |
-
"model.layers.18.mlp.down_proj.weight": "model-
|
99 |
-
"model.layers.18.mlp.gate_proj.weight": "model-
|
100 |
-
"model.layers.18.mlp.up_proj.weight": "model-
|
101 |
-
"model.layers.18.post_attention_layernorm.weight": "model-
|
102 |
"model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
103 |
-
"model.layers.18.self_attn.o_proj.weight": "model-
|
104 |
"model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
105 |
-
"model.layers.18.self_attn.v_proj.weight": "model-
|
106 |
-
"model.layers.19.input_layernorm.weight": "model-
|
107 |
-
"model.layers.19.mlp.down_proj.weight": "model-
|
108 |
-
"model.layers.19.mlp.gate_proj.weight": "model-
|
109 |
-
"model.layers.19.mlp.up_proj.weight": "model-
|
110 |
-
"model.layers.19.post_attention_layernorm.weight": "model-
|
111 |
-
"model.layers.19.self_attn.k_proj.weight": "model-
|
112 |
-
"model.layers.19.self_attn.o_proj.weight": "model-
|
113 |
-
"model.layers.19.self_attn.q_proj.weight": "model-
|
114 |
-
"model.layers.19.self_attn.v_proj.weight": "model-
|
115 |
"model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
116 |
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
117 |
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
@@ -121,33 +121,33 @@
|
|
121 |
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
122 |
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
123 |
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
124 |
-
"model.layers.20.input_layernorm.weight": "model-
|
125 |
-
"model.layers.20.mlp.down_proj.weight": "model-
|
126 |
-
"model.layers.20.mlp.gate_proj.weight": "model-
|
127 |
-
"model.layers.20.mlp.up_proj.weight": "model-
|
128 |
-
"model.layers.20.post_attention_layernorm.weight": "model-
|
129 |
-
"model.layers.20.self_attn.k_proj.weight": "model-
|
130 |
-
"model.layers.20.self_attn.o_proj.weight": "model-
|
131 |
-
"model.layers.20.self_attn.q_proj.weight": "model-
|
132 |
-
"model.layers.20.self_attn.v_proj.weight": "model-
|
133 |
-
"model.layers.21.input_layernorm.weight": "model-
|
134 |
-
"model.layers.21.mlp.down_proj.weight": "model-
|
135 |
-
"model.layers.21.mlp.gate_proj.weight": "model-
|
136 |
-
"model.layers.21.mlp.up_proj.weight": "model-
|
137 |
-
"model.layers.21.post_attention_layernorm.weight": "model-
|
138 |
-
"model.layers.21.self_attn.k_proj.weight": "model-
|
139 |
-
"model.layers.21.self_attn.o_proj.weight": "model-
|
140 |
-
"model.layers.21.self_attn.q_proj.weight": "model-
|
141 |
-
"model.layers.21.self_attn.v_proj.weight": "model-
|
142 |
"model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
143 |
-
"model.layers.22.mlp.down_proj.weight": "model-
|
144 |
-
"model.layers.22.mlp.gate_proj.weight": "model-
|
145 |
"model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
|
146 |
"model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
147 |
-
"model.layers.22.self_attn.k_proj.weight": "model-
|
148 |
-
"model.layers.22.self_attn.o_proj.weight": "model-
|
149 |
-
"model.layers.22.self_attn.q_proj.weight": "model-
|
150 |
-
"model.layers.22.self_attn.v_proj.weight": "model-
|
151 |
"model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
152 |
"model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
|
153 |
"model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 6425499648
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "model-00001-of-00002.safetensors",
|
|
|
94 |
"model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
95 |
"model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
96 |
"model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
97 |
+
"model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
98 |
+
"model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
99 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
100 |
+
"model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
101 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
102 |
"model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
103 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
104 |
"model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
105 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
106 |
+
"model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
107 |
+
"model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
108 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
109 |
+
"model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
110 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
111 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
112 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
113 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
114 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
115 |
"model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
116 |
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
117 |
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
121 |
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
122 |
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
123 |
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
124 |
+
"model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
125 |
+
"model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
126 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
127 |
+
"model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
128 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
129 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
130 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
131 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
132 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
133 |
+
"model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
134 |
+
"model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
135 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
136 |
+
"model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
137 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
138 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
139 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
140 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
141 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
142 |
"model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
143 |
+
"model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
144 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
145 |
"model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
|
146 |
"model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
147 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
148 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
149 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
150 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
151 |
"model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
152 |
"model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
|
153 |
"model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
|
tokenizer_config.json
CHANGED
@@ -2053,6 +2053,7 @@
|
|
2053 |
"chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"Cutting Knowledge Date: December 2023\n\nYou are a AI Assistant ReasonableLLaMa Junior, specialized in advanced reasoning and problem-solving. Break down the User Query, understand the context and what user needs, and address all constraints, challenges and gaps.\n Instructions and Response template within [[RESPONSE-FORMAT]]...[[/RESPONSE-FORMAT]]:\n\n[[RESPONSE-FORMAT]]\n<think>\n [Perform deep analytical thinking about the problem, thinking out loud when vocalizing your approch]\n [- When Assistant is stuck, I will Frame it as \\\"I am considering approch A and approch B, I am leaning towards A because ...\\\"\n - When needed I would ask clarification questions rather than heading down a wrong path]\n</think>\n\nFinal, response without any placeholders (including relevent reasoning) that directly addresses all aspects of the problem, further include verified implementation plan, cli commands or relevent code aginst the User's query validating that the response is complete and accurate.\n\n[[/RESPONSE-FORMAT]]\n\nRemember: DO NOT REPEAT THE TEMPLATE VARIABLES WITHIN [square-brackets] or [[RESPONSE-FORMAT]]/[[/RESPONSE-FORMAT]]. Your response start with the tag <think> and ends with the final response outside of the the think tag.\n\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\n\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\n\" }}\n{{- \"Today Date: \" + date_string + \"\n\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and tools is not none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {%- set first_user_message = \"\" %}\n {%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\n\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if message.tool_calls|length != 1 %}\n {%- for tool_call in message.tool_calls %}\n {%- set function_call = tool_call.function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + function_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- function_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endfor %}\n {%- else %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\n\n\" }}\n {%- if message.content is mapping or (message.content is iterable and message.content is not string) %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}\n{%- endif %}",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
|
|
2056 |
"model_input_names": [
|
2057 |
"input_ids",
|
2058 |
"attention_mask"
|
|
|
2053 |
"chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"Cutting Knowledge Date: December 2023\n\nYou are a AI Assistant ReasonableLLaMa Junior, specialized in advanced reasoning and problem-solving. Break down the User Query, understand the context and what user needs, and address all constraints, challenges and gaps.\n Instructions and Response template within [[RESPONSE-FORMAT]]...[[/RESPONSE-FORMAT]]:\n\n[[RESPONSE-FORMAT]]\n<think>\n [Perform deep analytical thinking about the problem, thinking out loud when vocalizing your approch]\n [- When Assistant is stuck, I will Frame it as \\\"I am considering approch A and approch B, I am leaning towards A because ...\\\"\n - When needed I would ask clarification questions rather than heading down a wrong path]\n</think>\n\nFinal, response without any placeholders (including relevent reasoning) that directly addresses all aspects of the problem, further include verified implementation plan, cli commands or relevent code aginst the User's query validating that the response is complete and accurate.\n\n[[/RESPONSE-FORMAT]]\n\nRemember: DO NOT REPEAT THE TEMPLATE VARIABLES WITHIN [square-brackets] or [[RESPONSE-FORMAT]]/[[/RESPONSE-FORMAT]]. Your response start with the tag <think> and ends with the final response outside of the the think tag.\n\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\n\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\n\" }}\n{{- \"Today Date: \" + date_string + \"\n\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and tools is not none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {%- set first_user_message = \"\" %}\n {%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\n\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if message.tool_calls|length != 1 %}\n {%- for tool_call in message.tool_calls %}\n {%- set function_call = tool_call.function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + function_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- function_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endfor %}\n {%- else %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\n\n\" }}\n {%- if message.content is mapping or (message.content is iterable and message.content is not string) %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}\n{%- endif %}",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
2056 |
+
"extra_special_tokens": {},
|
2057 |
"model_input_names": [
|
2058 |
"input_ids",
|
2059 |
"attention_mask"
|