Text Generation
GGUF
English
mixture of experts
reasoning
thinking
thoughts
deepseek
Mixture of Experts
context 128k
4x8B
Llama 3.1 MOE
creative
creative writing
general usage
problem solving
brainstorming
solve riddles
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
story
writing
fiction
roleplaying
swearing
horror
nsfw
llama 3.1
Not-For-All-Audiences
mergekit
Inference Endpoints
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -36,10 +36,10 @@ pipeline_tag: text-generation
|
|
36 |
|
37 |
<h2>DeepSeek-MOE-4X8B-R1-Distill-Llama-3.1-Deep-Thinker-Uncensored-24B-GGUF</h2>
|
38 |
|
39 |
-
This as a 4X8B, Mixture of Experts model with all 4 experts activated, all with Deepseek Reasoning tech installed giving you a 32B parameter model in only 24B model size.
|
40 |
-
|
41 |
<img src="blackroot.jpg" style="float:right; width:300px; height:300px; padding:5px;">
|
42 |
|
|
|
|
|
43 |
This model is a Deepseek model with "Distilled" components of "thinking/reasoning" fused into it.
|
44 |
|
45 |
This model contains 4 Llama fine tunes, and can be used for creative and non-creative use.
|
|
|
36 |
|
37 |
<h2>DeepSeek-MOE-4X8B-R1-Distill-Llama-3.1-Deep-Thinker-Uncensored-24B-GGUF</h2>
|
38 |
|
|
|
|
|
39 |
<img src="blackroot.jpg" style="float:right; width:300px; height:300px; padding:5px;">
|
40 |
|
41 |
+
This as a 4X8B, Mixture of Experts model with all 4 experts activated, all with Deepseek Reasoning tech installed giving you a 32B parameter model in only 24B model size.
|
42 |
+
|
43 |
This model is a Deepseek model with "Distilled" components of "thinking/reasoning" fused into it.
|
44 |
|
45 |
This model contains 4 Llama fine tunes, and can be used for creative and non-creative use.
|