adeelahmad commited on
Commit
adc3dab
·
verified ·
1 Parent(s): 1f1a796

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -1,4 +1,19 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  base_model: mlx-community/Llama-3.2-3B-Instruct
3
  language:
4
  - en
@@ -12,14 +27,6 @@ language:
12
  library_name: transformers
13
  license: llama3.2
14
  pipeline_tag: text-generation
15
- tags:
16
- - facebook
17
- - meta
18
- - pytorch
19
- - llama
20
- - llama-3
21
- - mlx
22
- - mlx
23
  extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Version\
24
  \ Release Date: September 25, 2024\n\n“Agreement” means the terms and conditions\
25
  \ for use, reproduction, distribution and modification of the Llama Materials set\
@@ -146,7 +153,7 @@ extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Vers
146
  \ overburden, interfere with or impair the proper working, integrity, operation\
147
  \ or appearance of a website or computer system\n 7. Engage in any action, or\
148
  \ facilitate any action, to intentionally circumvent or remove usage restrictions\
149
- \ or other safety measures, or to enable functionality disabled by Meta \n2. Engage\
150
  \ in, promote, incite, facilitate, or assist in the planning or development of activities\
151
  \ that present a risk of death or bodily harm to individuals, including use of Llama\
152
  \ 3.2 related to the following:\n 8. Military, warfare, nuclear industries or\
@@ -167,7 +174,7 @@ extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Vers
167
  \ without consent, authorization, or legal right\n 18. Representing that the\
168
  \ use of Llama 3.2 or outputs are human-generated\n 19. Generating or facilitating\
169
  \ false online engagement, including fake reviews and other means of fake online\
170
- \ engagement \n4. Fail to appropriately disclose to end users any known dangers\
171
  \ of your AI system 5. Interact with third party tools, models, or software designed\
172
  \ to generate unlawful content or engage in unlawful or harmful conduct and/or represent\
173
  \ that the outputs of such tools, models, or software are associated with Meta or\
@@ -207,31 +214,46 @@ extra_gated_description: The information you provide will be collected, stored,
207
  and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
208
  extra_gated_button_content: Submit
209
  ---
 
 
 
 
 
210
 
211
- # adeelahmad/ReasonableLlama3-3B-Jr
 
212
 
213
- The Model [adeelahmad/ReasonableLlama3-3B-Jr](https://huggingface.co/adeelahmad/ReasonableLlama3-3B-Jr) was
214
- converted to MLX format from [mlx-community/Llama-3.2-3B-Instruct](https://huggingface.co/mlx-community/Llama-3.2-3B-Instruct)
215
- using mlx-lm version **0.21.4**.
 
216
 
217
- ## Use with mlx
 
 
 
218
 
219
- ```bash
220
- pip install mlx-lm
221
- ```
 
222
 
223
- ```python
224
- from mlx_lm import load, generate
 
 
225
 
226
- model, tokenizer = load("adeelahmad/ReasonableLlama3-3B-Jr")
 
 
227
 
228
- prompt = "hello"
 
 
229
 
230
- if tokenizer.chat_template is not None:
231
- messages = [{"role": "user", "content": prompt}]
232
- prompt = tokenizer.apply_chat_template(
233
- messages, add_generation_prompt=True
234
- )
235
 
236
- response = generate(model, tokenizer, prompt=prompt, verbose=True)
237
- ```
 
 
1
  ---
2
+ tags:
3
+ - facebook
4
+ - meta
5
+ - pytorch
6
+ - llama
7
+ - llama-3
8
+ - mlx
9
+ - mlx
10
+ - reasoning
11
+ - llama
12
+ - deepseek
13
+ - ollama
14
+ - chain-of-thoughts
15
+ - small-llm
16
+ - edge
17
  base_model: mlx-community/Llama-3.2-3B-Instruct
18
  language:
19
  - en
 
27
  library_name: transformers
28
  license: llama3.2
29
  pipeline_tag: text-generation
 
 
 
 
 
 
 
 
30
  extra_gated_prompt: "### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT\n\nLlama 3.2 Version\
31
  \ Release Date: September 25, 2024\n\n“Agreement” means the terms and conditions\
32
  \ for use, reproduction, distribution and modification of the Llama Materials set\
 
153
  \ overburden, interfere with or impair the proper working, integrity, operation\
154
  \ or appearance of a website or computer system\n 7. Engage in any action, or\
155
  \ facilitate any action, to intentionally circumvent or remove usage restrictions\
156
+ \ or other safety measures, or to enable functionality disabled by Meta \n2. Engage\
157
  \ in, promote, incite, facilitate, or assist in the planning or development of activities\
158
  \ that present a risk of death or bodily harm to individuals, including use of Llama\
159
  \ 3.2 related to the following:\n 8. Military, warfare, nuclear industries or\
 
174
  \ without consent, authorization, or legal right\n 18. Representing that the\
175
  \ use of Llama 3.2 or outputs are human-generated\n 19. Generating or facilitating\
176
  \ false online engagement, including fake reviews and other means of fake online\
177
+ \ engagement \n4. Fail to appropriately disclose to end users any known dangers\
178
  \ of your AI system 5. Interact with third party tools, models, or software designed\
179
  \ to generate unlawful content or engage in unlawful or harmful conduct and/or represent\
180
  \ that the outputs of such tools, models, or software are associated with Meta or\
 
214
  and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
215
  extra_gated_button_content: Submit
216
  ---
217
+ # ReasonableLlama-3B: A Fine-Tuned Reasoning Model
218
+
219
+ HF: https://huggingface.co/adeelahmad/ReasonableLlama3-3B-Jr
220
+ Ollama: https://ollama.com/adeelahmad/ReasonableLLAMA-Jr-3b
221
+
222
 
223
+ Welcome to **ReasonableLlama-3B**, a cutting-edge reasoning model built on the foundation of LLaMA-3B. This model has been carefully fine-tuned to enhance its capabilities in logical
224
+ thinking, problem-solving, and creative analysis.
225
 
226
+ ## Overview
227
+ - **Model Name**: ReasonableLlama-3B
228
+ - **Base Architecture**: LLaMA-3B (Large Language Model with 3B parameters)
229
+ - **Purpose**: Designed for tasks requiring advanced reasoning, problem-solving, and creative thinking
230
 
231
+ ## Features
232
+ - **Advanced Reasoning**: Excels in logical analysis, problem-solving, and decision-making.
233
+ - **Creative Thinking**: Generates innovative solutions and ideas.
234
+ - **Curriculum-Based Fine-Tuning**: Trained on high-quality datasets to enhance reasoning abilities.
235
 
236
+ ## Technical Details
237
+ - **Parameter Count**: 3B parameters
238
+ - **Training Process**: Fine-tuned using state-of-the-art techniques for reasoning tasks
239
+ - **Specialization**: Optimized for specific reasoning workflows and scenarios
240
 
241
+ ## Use Cases
242
+ - **Research**: Facilitates complex problem-solving and theoretical analysis.
243
+ - **Education**: Assists in creating educational examples and problem sets.
244
+ - **Problem Solving**: Helps generate innovative solutions across various domains.
245
 
246
+ ## Installation and Usage
247
+ - **Integration**: Can be integrated into existing systems via APIs or local setup.
248
+ - **Inputs**: Supports text and images, leveraging Ollama's versatile capabilities.
249
 
250
+ ## Limitations
251
+ - **Scope**: Limited to single-step reasoning; multi-hop reasoning is a current focus area.
252
+ - **Data Bias**: Caution with dataset provenance as it may reflect historical biases.
253
 
254
+ ## Contributing
255
+ Contributions welcome! Fork the project, submit issues, and pull requests on GitHub. Your insights can help shape future improvements.
 
 
 
256
 
257
+ ## Citations
258
+ - Special thanks to LLaMA's developers for providing a strong foundation.
259
+ - Acknowledgments to the community contributing to open-source AI advancements.
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2bbdd2a6974d2df44cd93567dbde3e72deed6651b49394bbd2535675002485a1
3
- size 5368424925
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ff29640561515b67419537c05cc772061ccbcbd64a91a9253a7e26340e08cb4
3
+ size 5368478677
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea7d43db459f43a7c7b6bf291ca5b41dd1dc61abdf922053cac06e759689f088
3
- size 2466389976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06086cfa2c959311846c8f528db9264fccbf4fc389c236147ea5b6bfc878ef67
3
+ size 1057050040
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 7834785792
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
@@ -94,24 +94,24 @@
94
  "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
95
  "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
96
  "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
97
- "model.layers.18.input_layernorm.weight": "model-00002-of-00002.safetensors",
98
- "model.layers.18.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
99
- "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
100
- "model.layers.18.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
101
- "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
102
  "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
103
- "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
104
  "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
105
- "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
106
- "model.layers.19.input_layernorm.weight": "model-00002-of-00002.safetensors",
107
- "model.layers.19.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
108
- "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
109
- "model.layers.19.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
110
- "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
111
- "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
112
- "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
113
- "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
114
- "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
115
  "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
116
  "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
117
  "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
@@ -121,33 +121,33 @@
121
  "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
122
  "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
123
  "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
124
- "model.layers.20.input_layernorm.weight": "model-00002-of-00002.safetensors",
125
- "model.layers.20.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
126
- "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
127
- "model.layers.20.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
128
- "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
129
- "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
130
- "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
131
- "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
132
- "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
133
- "model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",
134
- "model.layers.21.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
135
- "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
136
- "model.layers.21.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
137
- "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
138
- "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
139
- "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
140
- "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
141
- "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
142
  "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
143
- "model.layers.22.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
144
- "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
145
  "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
146
  "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
147
- "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
148
- "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
149
- "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
150
- "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
151
  "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
152
  "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
153
  "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 6425499648
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
 
94
  "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
95
  "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
96
  "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
99
+ "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
102
  "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
104
  "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
111
+ "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
114
+ "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
115
  "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
116
  "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
117
  "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
 
121
  "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
122
  "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
123
  "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
126
+ "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
128
+ "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
135
+ "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
138
+ "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
140
+ "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
142
  "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
143
+ "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
145
  "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
146
  "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
147
+ "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
150
+ "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
151
  "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
152
  "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
153
  "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
tokenizer_config.json CHANGED
@@ -2053,6 +2053,7 @@
2053
  "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"Cutting Knowledge Date: December 2023\n\nYou are a AI Assistant ReasonableLLaMa Junior, specialized in advanced reasoning and problem-solving. Break down the User Query, understand the context and what user needs, and address all constraints, challenges and gaps.\n Instructions and Response template within [[RESPONSE-FORMAT]]...[[/RESPONSE-FORMAT]]:\n\n[[RESPONSE-FORMAT]]\n<think>\n [Perform deep analytical thinking about the problem, thinking out loud when vocalizing your approch]\n [- When Assistant is stuck, I will Frame it as \\\"I am considering approch A and approch B, I am leaning towards A because ...\\\"\n - When needed I would ask clarification questions rather than heading down a wrong path]\n</think>\n\nFinal, response without any placeholders (including relevent reasoning) that directly addresses all aspects of the problem, further include verified implementation plan, cli commands or relevent code aginst the User's query validating that the response is complete and accurate.\n\n[[/RESPONSE-FORMAT]]\n\nRemember: DO NOT REPEAT THE TEMPLATE VARIABLES WITHIN [square-brackets] or [[RESPONSE-FORMAT]]/[[/RESPONSE-FORMAT]]. Your response start with the tag <think> and ends with the final response outside of the the think tag.\n\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\n\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\n\" }}\n{{- \"Today Date: \" + date_string + \"\n\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and tools is not none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {%- set first_user_message = \"\" %}\n {%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\n\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if message.tool_calls|length != 1 %}\n {%- for tool_call in message.tool_calls %}\n {%- set function_call = tool_call.function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + function_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- function_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endfor %}\n {%- else %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\n\n\" }}\n {%- if message.content is mapping or (message.content is iterable and message.content is not string) %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}\n{%- endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
 
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
 
2053
  "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"Cutting Knowledge Date: December 2023\n\nYou are a AI Assistant ReasonableLLaMa Junior, specialized in advanced reasoning and problem-solving. Break down the User Query, understand the context and what user needs, and address all constraints, challenges and gaps.\n Instructions and Response template within [[RESPONSE-FORMAT]]...[[/RESPONSE-FORMAT]]:\n\n[[RESPONSE-FORMAT]]\n<think>\n [Perform deep analytical thinking about the problem, thinking out loud when vocalizing your approch]\n [- When Assistant is stuck, I will Frame it as \\\"I am considering approch A and approch B, I am leaning towards A because ...\\\"\n - When needed I would ask clarification questions rather than heading down a wrong path]\n</think>\n\nFinal, response without any placeholders (including relevent reasoning) that directly addresses all aspects of the problem, further include verified implementation plan, cli commands or relevent code aginst the User's query validating that the response is complete and accurate.\n\n[[/RESPONSE-FORMAT]]\n\nRemember: DO NOT REPEAT THE TEMPLATE VARIABLES WITHIN [square-brackets] or [[RESPONSE-FORMAT]]/[[/RESPONSE-FORMAT]]. Your response start with the tag <think> and ends with the final response outside of the the think tag.\n\" %}\n{%- endif %}\n\n{#- System message #}\n{{- \"<|start_header_id|>system<|end_header_id|>\n\n\" }}\n{%- if tools is not none %}\n {{- \"Environment: ipython\n\" }}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\n\" }}\n{{- \"Today Date: \" + date_string + \"\n\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and tools is not none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {%- set first_user_message = \"\" %}\n {%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\n\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\n\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\n\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if message.tool_calls|length != 1 %}\n {%- for tool_call in message.tool_calls %}\n {%- set function_call = tool_call.function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + function_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- function_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endfor %}\n {%- else %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\n\n\" }}\n {%- if message.content is mapping or (message.content is iterable and message.content is not string) %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}\n{%- endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
+ "extra_special_tokens": {},
2057
  "model_input_names": [
2058
  "input_ids",
2059
  "attention_mask"