Text Generation
GGUF
English
Chinese
MOE
Qwen 2.5 MOE
Mixture of Experts
6X1.5B
deepseek
reasoning
thinking
creative
128k context
general usage
problem solving
brainstorming
solve riddles
story generation
plot generation
storytelling
fiction story
story
writing
fiction
Qwen 2.5
mergekit
Inference Endpoints
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -77,7 +77,12 @@ This model is also mastered in Float 32, which helped overall model generation a
|
|
77 |
|
78 |
Temp of .4 to .8 is suggested, however it will still operate at much higher temps like 1.8, 2.6 etc.
|
79 |
|
80 |
-
|
|
|
|
|
|
|
|
|
|
|
81 |
|
82 |
Quants uploaded: Q4_K_S, Q8_0
|
83 |
|
|
|
77 |
|
78 |
Temp of .4 to .8 is suggested, however it will still operate at much higher temps like 1.8, 2.6 etc.
|
79 |
|
80 |
+
Depending on your prompt change temp SLOWLY: IE: .41,.42,.43 ... etc etc.
|
81 |
+
|
82 |
+
Likewise, because these are small models, it may do a tonne of "thinking"/"reasoning" and then "forget" to finish a / the task(s). In
|
83 |
+
this case, prompt the model to "Complete the task XYZ with the 'reasoning plan' above" .
|
84 |
+
|
85 |
+
Also set context limit at 4k minimum, 8K+ suggested.
|
86 |
|
87 |
Quants uploaded: Q4_K_S, Q8_0
|
88 |
|