Commit
·
fa51f55
1
Parent(s):
0373904
Update README.md
Browse files
README.md
CHANGED
@@ -27,12 +27,15 @@ It uses a mixture of the following datasets:
|
|
27 |
- GGML 30B 4-bit: [https://huggingface.co/gozfarb/llama-30b-supercot-ggml](https://huggingface.co/gozfarb/llama-30b-supercot-ggml)
|
28 |
- 30B (unquantized): [https://huggingface.co/ausboss/llama-30b-supercot](https://huggingface.co/ausboss/llama-30b-supercot)
|
29 |
- 30B 4-bit 128g CUDA: [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda)
|
|
|
|
|
30 |
|
31 |
#### 13B
|
|
|
32 |
- 13B (unquantized): [https://huggingface.co/ausboss/llama-13b-supercot](https://huggingface.co/ausboss/llama-13b-supercot)
|
33 |
- 13B 4-bit 128g CUDA: [https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g](https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g)
|
34 |
- 13B 4-bit 128g TRITON: [https://huggingface.co/TheYuriLover/llama-13b-SuperCOT-4bit-TRITON](https://huggingface.co/TheYuriLover/llama-13b-SuperCOT-4bit-TRITON)
|
35 |
-
-
|
36 |
|
37 |
(Thanks to all the awesome anons with supercomputers)
|
38 |
|
|
|
27 |
- GGML 30B 4-bit: [https://huggingface.co/gozfarb/llama-30b-supercot-ggml](https://huggingface.co/gozfarb/llama-30b-supercot-ggml)
|
28 |
- 30B (unquantized): [https://huggingface.co/ausboss/llama-30b-supercot](https://huggingface.co/ausboss/llama-30b-supercot)
|
29 |
- 30B 4-bit 128g CUDA: [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda)
|
30 |
+
- 30B 4-bit 128g TRITON: N/A
|
31 |
+
- 30B 4-bit CUDA (no groupsize, better VRAM): [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-cuda)
|
32 |
|
33 |
#### 13B
|
34 |
+
- GGML 13B 4-bit: [https://huggingface.co/gozfarb/llama-13b-supercot-ggml](https://huggingface.co/gozfarb/llama-13b-supercot-ggml)
|
35 |
- 13B (unquantized): [https://huggingface.co/ausboss/llama-13b-supercot](https://huggingface.co/ausboss/llama-13b-supercot)
|
36 |
- 13B 4-bit 128g CUDA: [https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g](https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g)
|
37 |
- 13B 4-bit 128g TRITON: [https://huggingface.co/TheYuriLover/llama-13b-SuperCOT-4bit-TRITON](https://huggingface.co/TheYuriLover/llama-13b-SuperCOT-4bit-TRITON)
|
38 |
+
- 13B 4-bit CUDA (no groupsize, better VRAM): N/A
|
39 |
|
40 |
(Thanks to all the awesome anons with supercomputers)
|
41 |
|