mradermacher commited on
Commit
eaa7278
·
verified ·
1 Parent(s): a068937

auto-patch README.md

Browse files
Files changed (1) hide show
  1. README.md +172 -0
README.md CHANGED
@@ -1,6 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  <!-- ### quantize_version: 2 -->
2
  <!-- ### output_tensor_quantised: 1 -->
3
  <!-- ### convert_type: hf -->
4
  <!-- ### vocab_type: -->
5
  <!-- ### tags: nicoboss -->
6
  weighted/imatrix quants of https://huggingface.co/LLaMAX/LLaMAX2-7B-X-CSQA
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: LLaMAX/LLaMAX2-7B-X-CSQA
3
+ language:
4
+ - af
5
+ - am
6
+ - ar
7
+ - hy
8
+ - as
9
+ - ast
10
+ - az
11
+ - be
12
+ - bn
13
+ - bs
14
+ - bg
15
+ - my
16
+ - ca
17
+ - ceb
18
+ - zho
19
+ - hr
20
+ - cs
21
+ - da
22
+ - nl
23
+ - en
24
+ - et
25
+ - tl
26
+ - fi
27
+ - fr
28
+ - ff
29
+ - gl
30
+ - lg
31
+ - ka
32
+ - de
33
+ - el
34
+ - gu
35
+ - ha
36
+ - he
37
+ - hi
38
+ - hu
39
+ - is
40
+ - ig
41
+ - id
42
+ - ga
43
+ - it
44
+ - ja
45
+ - jv
46
+ - kea
47
+ - kam
48
+ - kn
49
+ - kk
50
+ - km
51
+ - ko
52
+ - ky
53
+ - lo
54
+ - lv
55
+ - ln
56
+ - lt
57
+ - luo
58
+ - lb
59
+ - mk
60
+ - ms
61
+ - ml
62
+ - mt
63
+ - mi
64
+ - mr
65
+ - mn
66
+ - ne
67
+ - ns
68
+ - no
69
+ - ny
70
+ - oc
71
+ - or
72
+ - om
73
+ - ps
74
+ - fa
75
+ - pl
76
+ - pt
77
+ - pa
78
+ - ro
79
+ - ru
80
+ - sr
81
+ - sn
82
+ - sd
83
+ - sk
84
+ - sl
85
+ - so
86
+ - ku
87
+ - es
88
+ - sw
89
+ - sv
90
+ - tg
91
+ - ta
92
+ - te
93
+ - th
94
+ - tr
95
+ - uk
96
+ - umb
97
+ - ur
98
+ - uz
99
+ - vi
100
+ - cy
101
+ - wo
102
+ - xh
103
+ - yo
104
+ - zu
105
+ library_name: transformers
106
+ license: mit
107
+ quantized_by: mradermacher
108
+ tags:
109
+ - Multilingual
110
+ ---
111
+ ## About
112
+
113
  <!-- ### quantize_version: 2 -->
114
  <!-- ### output_tensor_quantised: 1 -->
115
  <!-- ### convert_type: hf -->
116
  <!-- ### vocab_type: -->
117
  <!-- ### tags: nicoboss -->
118
  weighted/imatrix quants of https://huggingface.co/LLaMAX/LLaMAX2-7B-X-CSQA
119
+
120
+ <!-- provided-files -->
121
+ static quants are available at https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-GGUF
122
+ ## Usage
123
+
124
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
125
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
126
+ more details, including on how to concatenate multi-part files.
127
+
128
+ ## Provided Quants
129
+
130
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
131
+
132
+ | Link | Type | Size/GB | Notes |
133
+ |:-----|:-----|--------:|:------|
134
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ1_S.gguf) | i1-IQ1_S | 1.6 | for the desperate |
135
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ1_M.gguf) | i1-IQ1_M | 1.8 | mostly desperate |
136
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.0 | |
137
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.1 | |
138
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ2_S.gguf) | i1-IQ2_S | 2.3 | |
139
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q2_K_S.gguf) | i1-Q2_K_S | 2.4 | very low quality |
140
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ2_M.gguf) | i1-IQ2_M | 2.5 | |
141
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q2_K.gguf) | i1-Q2_K | 2.6 | IQ3_XXS probably better |
142
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 2.7 | lower quality |
143
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ3_XS.gguf) | i1-IQ3_XS | 2.9 | |
144
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ3_S.gguf) | i1-IQ3_S | 3.0 | beats Q3_K* |
145
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.0 | IQ3_XS probably better |
146
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ3_M.gguf) | i1-IQ3_M | 3.2 | |
147
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q3_K_M.gguf) | i1-Q3_K_M | 3.4 | IQ3_S probably better |
148
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q3_K_L.gguf) | i1-Q3_K_L | 3.7 | IQ3_M probably better |
149
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ4_XS.gguf) | i1-IQ4_XS | 3.7 | |
150
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-IQ4_NL.gguf) | i1-IQ4_NL | 3.9 | prefer IQ4_XS |
151
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q4_0.gguf) | i1-Q4_0 | 3.9 | fast, low quality |
152
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.0 | optimal size/speed/quality |
153
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q4_K_M.gguf) | i1-Q4_K_M | 4.2 | fast, recommended |
154
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q4_1.gguf) | i1-Q4_1 | 4.3 | |
155
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q5_K_S.gguf) | i1-Q5_K_S | 4.8 | |
156
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q5_K_M.gguf) | i1-Q5_K_M | 4.9 | |
157
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-X-CSQA-i1-GGUF/resolve/main/LLaMAX2-7B-X-CSQA.i1-Q6_K.gguf) | i1-Q6_K | 5.6 | practically like static Q6_K |
158
+
159
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
160
+ types (lower is better):
161
+
162
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
163
+
164
+ And here are Artefact2's thoughts on the matter:
165
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
166
+
167
+ ## FAQ / Model Request
168
+
169
+ See https://huggingface.co/mradermacher/model_requests for some answers to
170
+ questions you might have and/or if you want some other model quantized.
171
+
172
+ ## Thanks
173
+
174
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
175
+ me use its servers and providing upgrades to my workstation to enable
176
+ this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
177
+
178
+ <!-- end -->