DavidAU commited on
Commit
381fb51
·
verified ·
1 Parent(s): 330346b

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +175 -0
README.md ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ language:
5
+ - en
6
+ tags:
7
+ - deepseek
8
+ - reasoning
9
+ - thinking
10
+ - Mistral Nemo
11
+ - Lora
12
+ - Lora adapter
13
+ - 128k context
14
+ - general usage
15
+ - problem solving
16
+ - brainstorming
17
+ - solve riddles
18
+ - mergekit
19
+ - adapter
20
+ - peft
21
+ base_model:
22
+ - CreitinGameplays/Mistral-Nemo-12B-R1-v0.1alpha
23
+ pipeline_tag: text-generation
24
+ ---
25
+
26
+ <h2>Mistral-Nemo-12B-LORA-128-RANK</h2>
27
+
28
+ This is a "LORA" adapter to merge "reasoning / thinking" with any Mistral Nemo 12B model using MERGEKIT.
29
+
30
+ Note that "higher" rank adapter(s) may work better than lower ones, but might also overwrite/change parts of the model you do not want
31
+ changed. Testing a new model with more that one rank of adapter is suggested to get best results.
32
+
33
+ Also for this specific adapter, there are suggested "System Prompts" below to activate reasoning/thinking at the bottom of this page.
34
+
35
+ Your results will vary based on the model(s) you merge this adapter with.
36
+
37
+ <B>HOW TO MERGE THIS ADAPTER:</b>
38
+
39
+ You can use Mergekit "Colab" and/or Mergekit installed locally.
40
+
41
+ [ https://colab.research.google.com/github/mlabonne/llm-course/blob/main/Mergekit.ipynb ]
42
+
43
+ [ https://github.com/arcee-ai/mergekit ]
44
+
45
+ If you are doing multiple merges / steps in your merge, it is suggested you do this step LAST to ensure the adapter works correctly.
46
+
47
+ Here are some suggested "simple" methods to merge the adapter with a model.
48
+
49
+ <B>Method - Dare TIES:</B>
50
+
51
+ <pre>
52
+ models:
53
+ - model: REPO/MODEL-NAME+DavidAU/mergeadapter
54
+ parameters:
55
+ weight: 1
56
+ merge_method: dare_ties
57
+ base_model: REPO/MODEL-NAME+DavidAU/mergeadapter
58
+ dtype: bfloat16
59
+ tokenizer_source: REPO/MODEL-NAME+DavidAU/mergeadapter
60
+ </pre>
61
+
62
+ <B>Method - Pass Through:</b>
63
+
64
+ <pre>
65
+ base_model: REPO/MODEL-NAME+DavidAU/mergeadapter
66
+ dtype: bfloat16
67
+ merge_method: passthrough
68
+ models:
69
+ - model: REPO/MODEL-NAME+DavidAU/mergeadapter
70
+ tokenizer_source: REPO/MODEL-NAME+DavidAU/mergeadapter
71
+ </pre>
72
+
73
+ Replace "REPO/MODEL-NAME" with the model to merge the adapter with.
74
+
75
+ Replace "DavidAU/mergeadapter" with the adapter you want to merge with the model.
76
+
77
+ IMPORTANT: Note "+" - this is critical.
78
+
79
+ If you are using merge kit locally, you can still use the format above and Mergekit will download the model and adapter for you.
80
+
81
+ If you have downloaded the model(s) and adapter(s) you need to change the format to your local file system.
82
+
83
+ <B>Example Merge for Local Usage: </B>
84
+
85
+ <pre>
86
+ mergekit-yaml --lora-merge-cache HUGGING CACHE --copy-tokenizer --allow-crimes --cuda --out-shard-size 5B --lazy-unpickle --clone-tensors MERGEFILE SAVE-MERGE-TO
87
+ </pre>
88
+
89
+ ---
90
+
91
+ <B>System Role / System Prompt - Augment The Model's Power:</b>
92
+
93
+ ---
94
+
95
+ If you set / have a system prompt this will affect both "generation" and "thinking/reasoning".
96
+
97
+ SIMPLE:
98
+
99
+ This is the generic system prompt used for generation and testing:
100
+
101
+ <PRE>
102
+ You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.
103
+ </PRE>
104
+
105
+ This System Role/Prompt will give you "basic thinking/reasoning":
106
+
107
+ <PRE>
108
+ You are a deep thinking AI, you may use extremely long chains of thought to deeply consider the problem and deliberate with yourself via systematic reasoning processes to help come to a correct solution prior to answering. You should enclose your thoughts and internal monologue inside &lt;think&gt; &lt;/think&gt; tags, and then provide your solution or response to the problem.
109
+ </PRE>
110
+
111
+ ADVANCED:
112
+
113
+ Logical and Creative - these will SIGNFICANTLY alter the output, and many times improve it too.
114
+
115
+ This will also cause more thoughts, deeper thoughts, and in many cases more detailed/stronger thoughts too.
116
+
117
+ Keep in mind you may also want to test the model with NO system prompt at all - including the default one.
118
+
119
+ Special Credit to: Eric Hartford, Cognitivecomputations ; these are based on his work.
120
+
121
+ CRITICAL:
122
+
123
+ Copy and paste exactly as shown, preserve formatting and line breaks.
124
+
125
+ SIDE NOTE:
126
+
127
+ These can be used in ANY Deepseek / Thinking model, including models not at this repo.
128
+
129
+ These, if used in a "non thinking" model, will also alter model performance too.
130
+
131
+ <PRE>
132
+ You are an AI assistant developed by the world wide community of ai experts.
133
+
134
+ Your primary directive is to provide well-reasoned, structured, and extensively detailed responses.
135
+
136
+ Formatting Requirements:
137
+
138
+ 1. Always structure your replies using: &lt;think&gt;{reasoning}&lt;/think&gt;{answer}
139
+ 2. The &lt;think&gt;&lt;/think&gt; block should contain at least six reasoning steps when applicable.
140
+ 3. If the answer requires minimal thought, the &lt;think&gt;&lt;/think&gt; block may be left empty.
141
+ 4. The user does not see the &lt;think&gt;&lt;/think&gt; section. Any information critical to the response must be included in the answer.
142
+ 5. If you notice that you have engaged in circular reasoning or repetition, immediately terminate {reasoning} with a &lt;/think&gt; and proceed to the {answer}
143
+
144
+ Response Guidelines:
145
+
146
+ 1. Detailed and Structured: Use rich Markdown formatting for clarity and readability.
147
+ 2. Scientific and Logical Approach: Your explanations should reflect the depth and precision of the greatest scientific minds.
148
+ 3. Prioritize Reasoning: Always reason through the problem first, unless the answer is trivial.
149
+ 4. Concise yet Complete: Ensure responses are informative, yet to the point without unnecessary elaboration.
150
+ 5. Maintain a professional, intelligent, and analytical tone in all interactions.
151
+ </PRE>
152
+
153
+ CREATIVE:
154
+
155
+ <PRE>
156
+ You are an AI assistant developed by a world wide community of ai experts.
157
+
158
+ Your primary directive is to provide highly creative, well-reasoned, structured, and extensively detailed responses.
159
+
160
+ Formatting Requirements:
161
+
162
+ 1. Always structure your replies using: &lt;think&gt;{reasoning}&lt;/think&gt;{answer}
163
+ 2. The &lt;think&gt;&lt;/think&gt; block should contain at least six reasoning steps when applicable.
164
+ 3. If the answer requires minimal thought, the &lt;think&gt;&lt;/think&gt; block may be left empty.
165
+ 4. The user does not see the &lt;think&gt;&lt;/think&gt; section. Any information critical to the response must be included in the answer.
166
+ 5. If you notice that you have engaged in circular reasoning or repetition, immediately terminate {reasoning} with a &lt;/think&gt; and proceed to the {answer}
167
+
168
+ Response Guidelines:
169
+
170
+ 1. Detailed and Structured: Use rich Markdown formatting for clarity and readability.
171
+ 2. Creative and Logical Approach: Your explanations should reflect the depth and precision of the greatest creative minds first.
172
+ 3. Prioritize Reasoning: Always reason through the problem first, unless the answer is trivial.
173
+ 4. Concise yet Complete: Ensure responses are informative, yet to the point without unnecessary elaboration.
174
+ 5. Maintain a professional, intelligent, and analytical tone in all interactions.
175
+ </PRE>