Doctor-Shotgun commited on
Commit
e699dfd
·
1 Parent(s): 1342776

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -13
README.md CHANGED
@@ -6,11 +6,10 @@ base_model: deepseekai/deepseek-llm-67b-base
6
  model-index:
7
  - name: workspace/volume/limarp-deepseek-qlora-out
8
  results: []
 
 
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
15
  <details><summary>See axolotl config</summary>
16
 
@@ -87,21 +86,63 @@ fsdp_config:
87
 
88
  </details><br>
89
 
90
- # workspace/volume/limarp-deepseek-qlora-out
91
 
92
- This model was trained from scratch on the None dataset.
93
 
94
  ## Model description
95
 
96
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
 
98
  ## Intended uses & limitations
99
 
100
- More information needed
101
 
102
  ## Training and evaluation data
103
 
104
- More information needed
105
 
106
  ## Training procedure
107
 
@@ -119,10 +160,6 @@ The following hyperparameters were used during training:
119
  - lr_scheduler_warmup_steps: 10
120
  - num_epochs: 2
121
 
122
- ### Training results
123
-
124
-
125
-
126
  ### Framework versions
127
 
128
  - Transformers 4.36.2
@@ -147,4 +184,4 @@ The following `bitsandbytes` quantization config was used during training:
147
  ### Framework versions
148
 
149
 
150
- - PEFT 0.6.0
 
6
  model-index:
7
  - name: workspace/volume/limarp-deepseek-qlora-out
8
  results: []
9
+ datasets:
10
+ - lemonilia/LimaRP
11
  ---
12
 
 
 
 
13
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
14
  <details><summary>See axolotl config</summary>
15
 
 
86
 
87
  </details><br>
88
 
89
+ # limarp-deepseek-67b-qlora
90
 
91
+ This model is an unofficial Deepseek 67B training on the LimaRP dataset.
92
 
93
  ## Model description
94
 
95
+ Prompt format is the [extended Alpaca format](https://github.com/tatsu-lab/stanford_alpaca):
96
+
97
+ ```
98
+ ### Instruction:
99
+ Character's Persona: {bot character description}
100
+ User's Persona: {user character description}
101
+ Scenario: {what happens in the story}
102
+ Play the role of Character. You must engage in a roleplaying chat with User below this line. Do not write dialogues and narration for User.
103
+ ### Input:
104
+ User: {utterance}
105
+ ### Response:
106
+ Character: {utterance}
107
+ ### Input
108
+ User: {utterance}
109
+ ### Response:
110
+ Character: {utterance}
111
+ (etc.)
112
+ ```
113
+
114
+ Inspired by the previously named "Roleplay" preset in SillyTavern, with this version of LimaRP it is possible to append a length modifier to the response instruction sequence, like this:
115
+
116
+ ```
117
+ ### Input
118
+ User: {utterance}
119
+
120
+ ### Response: (length = medium)
121
+ Character: {utterance}
122
+ ```
123
+
124
+ This has an immediately noticeable effect on bot responses. The lengths using during training are:
125
+ `micro`, `tiny`, `short`, `medium`, `long`, `massive`, `huge`, `enormous`, `humongous`, `unlimited`.
126
+ **The recommended starting length is medium**. Keep in mind that the AI can ramble or impersonate
127
+ the user with very long messages.
128
+
129
+ The length control effect is reproducible, but the messages will not necessarily follow
130
+ lengths very precisely, rather follow certain ranges on average, as seen in this table
131
+ with data from tests made with one reply at the beginning of the conversation:
132
+
133
+ ![lengths](https://i.imgur.com/2WXGgaV.png)
134
+
135
+ Response length control appears to work well also deep into the conversation. **By omitting
136
+ the modifier, the model will choose the most appropriate response length** (although it might
137
+ not necessarily be what the user desires).
138
 
139
  ## Intended uses & limitations
140
 
141
+ The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model.
142
 
143
  ## Training and evaluation data
144
 
145
+ For more details about LimaRP, see the dataset page.
146
 
147
  ## Training procedure
148
 
 
160
  - lr_scheduler_warmup_steps: 10
161
  - num_epochs: 2
162
 
 
 
 
 
163
  ### Framework versions
164
 
165
  - Transformers 4.36.2
 
184
  ### Framework versions
185
 
186
 
187
+ - PEFT 0.6.0