Text Generation
GGUF
English
Chinese
MOE
Qwen 2.5 MOE
Mixture of Experts
6X1.5B
deepseek
reasoning
thinking
creative
128k context
general usage
problem solving
brainstorming
solve riddles
story generation
plot generation
storytelling
fiction story
story
writing
fiction
Qwen 2.5
mergekit
Inference Endpoints
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -57,7 +57,7 @@ In Lmstudio the "Jinja Template" should load by default.
|
|
57 |
|
58 |
In other apps - use the Deepseek Tokenizer.
|
59 |
|
60 |
-
Sometimes this model will output Chinese Characters/Symbols (with an English prompt) - regen to clear.
|
61 |
|
62 |
Sometimes it will work great, other times it will give "so/so" answers and then sometimes it will bat it out of the park, and past the "state line."
|
63 |
|
@@ -67,13 +67,14 @@ It is all over the map.
|
|
67 |
|
68 |
Four examples below so you have some idea what this model can do.
|
69 |
|
70 |
-
Keep in mind this model is six 1.5B parameters models working together, and will not have the power of
|
71 |
|
72 |
-
Also, this model has 4/6 experts activated.
|
73 |
|
74 |
You may want to set 6/6 experts for best results.
|
75 |
|
76 |
-
This model is also mastered in Float 32, which helped overall model generation and addressed some model issues
|
|
|
77 |
|
78 |
Temp of .4 to .8 is suggested, however it will still operate at much higher temps like 1.8, 2.6 etc.
|
79 |
|
|
|
57 |
|
58 |
In other apps - use the Deepseek Tokenizer.
|
59 |
|
60 |
+
Sometimes this model will output/think in Chinese Characters/Symbols (with an English prompt) - regen to clear.
|
61 |
|
62 |
Sometimes it will work great, other times it will give "so/so" answers and then sometimes it will bat it out of the park, and past the "state line."
|
63 |
|
|
|
67 |
|
68 |
Four examples below so you have some idea what this model can do.
|
69 |
|
70 |
+
Keep in mind this model is six 1.5B parameters models working together, and will not have the power of a 14B or 32B reasoning/thinking model.
|
71 |
|
72 |
+
Also, this model has 4/6 experts activated by default.
|
73 |
|
74 |
You may want to set 6/6 experts for best results.
|
75 |
|
76 |
+
This model is also mastered in Float 32, which helped overall model generation and addressed some model generation issues
|
77 |
+
and oddly seemed to add some new ones (? - Chinese Char/Symb thinking.).
|
78 |
|
79 |
Temp of .4 to .8 is suggested, however it will still operate at much higher temps like 1.8, 2.6 etc.
|
80 |
|