Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ hasnonname's trashpanda baby is getting a sequel. More JLLM-ish than ever, too.
|
|
19 |
|
20 |
<p><b>Context/instruct template</b>: Mistral V7 or V3 Tekken, <b>but for some godforsaken reason, we found that this is (arguably) better with Llama 3 context/instruct.</b> It's funny, stupid and insane, we don't know why this is the case. Trying out Llama 3 instruct/context on base MS24B told us it was coherent for it too in 4/5 responses, but not better than Mistral ones. As to why V1 seems to do better than base MS24B, we don't really know.</p>
|
21 |
|
22 |
-
<p><b>Samplers</b>: temperature at 0.8 - 1.
|
23 |
|
24 |
A virt-io derivative prompt worked best during our testing, but feel free to use what you like.
|
25 |
|
|
|
19 |
|
20 |
<p><b>Context/instruct template</b>: Mistral V7 or V3 Tekken, <b>but for some godforsaken reason, we found that this is (arguably) better with Llama 3 context/instruct.</b> It's funny, stupid and insane, we don't know why this is the case. Trying out Llama 3 instruct/context on base MS24B told us it was coherent for it too in 4/5 responses, but not better than Mistral ones. As to why V1 seems to do better than base MS24B, we don't really know.</p>
|
21 |
|
22 |
+
<p><b>Samplers</b>: temperature at 0.8 - 1.25, min_p at 0.05, top_a at 0.2, smoothing_factor at 0.2. Some optional settings include repetition_penalty at 1.03 or DRY if you have access to it.</p>
|
23 |
|
24 |
A virt-io derivative prompt worked best during our testing, but feel free to use what you like.
|
25 |
|