invisietch commited on
Commit
fc0a47e
·
verified ·
1 Parent(s): 1487689

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -26
README.md CHANGED
@@ -1,40 +1,100 @@
1
  ---
2
- base_model: []
3
  library_name: transformers
4
  tags:
5
- - mergekit
6
- - merge
7
-
 
 
 
8
  ---
9
- # MiS-Firefly-v0.2-22B
10
 
11
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
- ## Merge Details
14
- ### Merge Method
15
 
16
- This model was merged using the SLERP merge method.
17
 
18
- ### Models Merged
 
 
19
 
20
- The following models were included in the merge:
21
- * /mnt/models/checkpoint-404
22
- * /mnt/models/checkpoint-808
23
 
24
- ### Configuration
25
 
26
- The following YAML configuration was used to produce this model:
 
27
 
28
- ```yaml
29
- dtype: bfloat16
30
- models:
31
- - model: /mnt/models/checkpoint-808
32
- - model: /mnt/models/checkpoint-404
33
- merge_method: slerp
34
- base_model: /mnt/models/checkpoint-808
35
- parameters:
36
- t:
37
- - value: [0, 0, 0.25, 0.35, 0.4, 0.45, 0.4, 0.35, 0.25, 0, 0]
38
- embed_slerp: true
39
 
40
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  library_name: transformers
3
  tags:
4
+ - not-for-all-audiences
5
+ - axolotl
6
+ - qlora
7
+ language:
8
+ - en
9
+ license: other
10
  ---
 
11
 
12
+ <div align="center">
13
+ <b style="font-size: 36px;">MiS-Firefly-v0.2-22B (GGUF)</b>
14
+
15
+ <img src="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B-GGUF/resolve/main/header.png" style="width:60%">
16
+
17
+ <b>HF</b> :
18
+ <a href="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B">FP16</a>
19
+ &vert;
20
+ <b>GGUF</b> :
21
+ <a href="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B-GGUF">Static GGUF</a>
22
+ </div>
23
+
24
+ # Model Details
25
+
26
+ **This is a fix for the quantization issue in Firefly v0.1.**
27
+
28
+ Firefly is a Mistral Small 22B finetune designed for creative writing and roleplay. The model is largely uncensored and should support
29
+ context up to 32,768 tokens.
30
+
31
+ The model has been tested in various roleplay scenarios up to 16k context, as well as in a role as an assistant. It shows a broad
32
+ competency &amp; coherence across various scenarios.
33
+
34
+ Special thanks to <a href="https://huggingface.co/SicariusSicariiStuff">SicariusSicariiStuff</a> for bouncing ideas back &amp; forth on
35
+ training, and <a href="https://huggingface.co/SytanSD">SytanSD</a> for quants.
36
+
37
+ ## KNOWN QUANTIZATION ISSUE
38
+
39
+ Some quants seem to have issues with misspelling complicated names.
40
+
41
+ This doesn't happen at fp16 or q8_0 even with very weird names and multiple swipes meaning something's getting lost in quant.
42
+
43
+ Suggested workarounds:
44
+
45
+ - If you can, run q8_0 (I'm told this fits on a 4090 with flash attention), I haven't seen the issue in ~900 messages on q8.
46
+ - If not, try some lower quants (ideally imatrix), I haven't tested them all but it appears to happen on Q6_K the most and less often on the 6.5bpw EXL2. If you find one where this doesn't happen, tell me.
47
+ - If none of that works, use a simpler name.
48
+
49
+ I'll try resolving it with a light merge ASAP, it seems like the wrong weight is just getting truncated in quantization causing these issues.
50
 
51
+ # Feedback
 
52
 
53
+ I appreciate all feedback on any of my models, you can use:
54
 
55
+ * [My Discord server](https://discord.gg/AJwZuu7Ncx) - requires Discord.
56
+ * [The Community tab](https://huggingface.co/invisietch/MiS-Firefly-v0.1-22B/discussions) - requires HF login.
57
+ * Discord DMs to **invisietch**.
58
 
59
+ Your feedback is how I improve these models for future versions.
 
 
60
 
61
+ # Disclaimer
62
 
63
+ This model is extensively uncensored. It can generate explicit, disturbing or offensive responses. Use responsibly. I am not responsible for
64
+ your use of this model.
65
 
66
+ This model is a finetune of Mistral Small 22B (2409) and usage must follow the terms of Mistral's license. By downloading this model, you
67
+ agree not to use it for commercial purposes unless you have a valid Mistral commercial license. See [the base model card](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409)
68
+ for more details.
69
+
70
+ # Prompting Format
71
+
72
+ I'd recommend Mistral v2v3 prompting format:
 
 
 
 
73
 
74
  ```
75
+ <s>[INST] User message here.[/INST] Bot response here</s>[INST] User message 2 here.
76
+ ```
77
+
78
+ # Sampler Settings
79
+
80
+ I'm running the following sampler settings but this is an RC and they may not be optimal.
81
+
82
+ - **Temperature:** Dynamic 0.7-1.1
83
+ - **Min-P:** 0.07
84
+ - **Rep Pen:** 1.08
85
+ - **Rep Pen Range:** 1536
86
+ - **XTC:** 0.1/0.15
87
+
88
+ If you get completely incoherent responses, feel free to use these as a starting point.
89
+
90
+ # Training Strategy
91
+
92
+ I started with a finetune of Mistral Small 22B which had been trained on the Gutenberg dataset: [nbeerbower/Mistral-Small-Gutenberg-Doppel-22B](https://huggingface.co/nbeerbower/Mistral-Small-Gutenberg-Doppel-22B).
93
+
94
+ The first stage of my training was a single epoch at low LR over a 474 million token text completion dataset.
95
+
96
+ I followed this up with a coherence, decensorship & roleplay finetune over a 172 million token instruct dataset over two epochs.
97
+
98
+ Total training time was about 32hrs on 4x Nvidia A100 80GB.
99
+
100
+ <img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>