DavidAU commited on
Commit
7d71d3b
·
verified ·
1 Parent(s): 31ee94e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -2
README.md CHANGED
@@ -31,7 +31,9 @@ pipeline_tag: text-generation
31
 
32
  <H2>Qwen2.5-MOE-6x1.5B-DeepSeek-Reasoning-e32-gguf</H2>
33
 
34
- This is a highly experimental Qwen2.5 MOE (Mixture of Experts) model comprised of SIX Qwen 2.5 1.5B models.
 
 
35
 
36
  It includes the following models:
37
 
@@ -55,7 +57,7 @@ if your AI/LLM app can not access the "Jinja Template".
55
 
56
  In Lmstudio the "Jinja Template" should load by default.
57
 
58
- In other apps - use the Deepseek Tokenizer.
59
 
60
  Sometimes this model will output/think in Chinese Characters/Symbols (with an English prompt) - regen to clear.
61
 
@@ -82,6 +84,10 @@ Depending on your prompt change temp SLOWLY: IE: .41,.42,.43 ... etc etc.
82
 
83
  Likewise, because these are small models, it may do a tonne of "thinking"/"reasoning" and then "forget" to finish a / the task(s). In
84
  this case, prompt the model to "Complete the task XYZ with the 'reasoning plan' above" .
 
 
 
 
85
 
86
  Also set context limit at 4k minimum, 8K+ suggested.
87
 
 
31
 
32
  <H2>Qwen2.5-MOE-6x1.5B-DeepSeek-Reasoning-e32-gguf</H2>
33
 
34
+ This is a highly experimental Qwen2.5 MOE (Mixture of Experts) model comprised of SIX Qwen 2.5 1.5B models creating a 8.71B model.
35
+
36
+ This model can be used for all use cases, and is also (mostly) uncensored.
37
 
38
  It includes the following models:
39
 
 
57
 
58
  In Lmstudio the "Jinja Template" should load by default.
59
 
60
+ In other apps - use the Deepseek Tokenizer and/or "Jinja Template".
61
 
62
  Sometimes this model will output/think in Chinese Characters/Symbols (with an English prompt) - regen to clear.
63
 
 
84
 
85
  Likewise, because these are small models, it may do a tonne of "thinking"/"reasoning" and then "forget" to finish a / the task(s). In
86
  this case, prompt the model to "Complete the task XYZ with the 'reasoning plan' above" .
87
+
88
+ Likewise it may function better if you breakdown the reasoning/thinking task(s) into smaller pieces :
89
+
90
+ "IE: Instead of asking for 6 plots FOR theme XYZ, ASK IT for ONE plot for theme XYZ at a time".
91
 
92
  Also set context limit at 4k minimum, 8K+ suggested.
93