lemonilia commited on
Commit
e87a881
·
verified ·
1 Parent(s): 2782e9b

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ pipeline_tag: conversational
6
+ tags:
7
+ - not-for-all-audiences
8
+ ---
9
+
10
+ Base: [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
11
+
12
+ Test model, do **not** use.
13
+
14
+ It employs a different prompting format than the base's, and _not_ Alpaca. Not intended for
15
+ public consumption yet, so no information will be given here in that regard.
16
+
17
+ It's unlikely that the model will produce the intended outputs without the specific format it's
18
+ been trained on.
19
+
20
+ # Dataset
21
+ Similar to LimaRP, but more niche. Flexible training sample length (from 4k to 32k tokens,
22
+ at least). Might or might not be released in the future.
23
+
24
+ # Training details
25
+ ## Hardware
26
+ 1x NVidia RTX 3090 24GB
27
+
28
+ ## Software
29
+ [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl)
30
+
31
+ ## Training hyperparameters
32
+ - load_in_4bit: true
33
+ - adapter: qlora
34
+ - sequence_len: 16384
35
+ - sample_packing: true
36
+ - pad_to_sequence_len: false
37
+ - gradient_accumulation_steps: 4
38
+ - micro_batch_size: 1
39
+ - eval_batch_size: 1
40
+ - num_epochs: 2
41
+ - optimizer: adamw_bnb_8bit
42
+ - lr_scheduler: constant
43
+ - learning_rate: 0.000085
44
+ - weight_decay: 0.05
45
+ - train_on_inputs: true
46
+ - bf16: true
47
+ - fp16: false
48
+ - tf32: true
49
+ - lora_r: 20
50
+ - lora_alpha: 16
51
+ - lora_dropout: 0.1
52
+ - lora_target_linear: true