922CA commited on
Commit
cea6ee5
·
1 Parent(s): f417a5c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ ---
4
+ # LLilmonix3b-v0.4a:
5
+ * Experimental [Marx-3b-v2](https://huggingface.co/acrastt/Marx-3B-V2) fine-tuned for Monika character from DDLC
6
+ * Trained on a dataset of ~600 items (dialogue scraped from game, reddit, and Twitter augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn each into snippets of multi-turn chat dialogue between Player and Monika + manually crafted test dataset of 12 items)
7
+ * Trained to run on smaller devices
8
+ * [QLoras (hf and GGML)](https://huggingface.co/922-CA/monika-lm-lora-tests/tree/main/LLilmonix3b-v0.4-Marxv2)
9
+
10
+ ### USAGE
11
+ This is meant to be mainly a chat model with limited RP ability.
12
+
13
+ For best results: replace "Human" and "Assistant" with "Player" and "Monika" like so:
14
+
15
+ \nPlayer: (prompt)\nMonika:
16
+
17
+ Tested primarily with koboldcpp 1.41+, with q8_0 GGML, temperature at 0.7, and the rest left as is on chat mode:
18
+
19
+ ### HYPERPARAMS
20
+ * Trained for 2 epochs
21
+ * rank: 32
22
+ * lora alpha: 64
23
+ * lora dropout: 0.5
24
+ * lr: 2e-4
25
+ * batch size: 2
26
+ * warmup ratio: 0.1
27
+ * grad steps: 4
28
+
29
+ ### WARNINGS AND DISCLAIMERS
30
+ Note that aside from formatting and other minor edits, generated portion of dataset used is mostly as is generated by LM. In addition, the is meant to be a smaller version of the larger Monika models. As such, this version may not reflect perfectly the characteristics of Monika.
31
+
32
+ Additionally, this is still yet another test, particularly where we use one of our earlier fine tunes to generate a more in-character dataset for the target character.
33
+
34
+ Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk.