Upload folder using huggingface_hub
Browse files- Qwen2.5-3B-Q2_K.gguf +2 -2
- Qwen2.5-3B-Q3_K_L.gguf +2 -2
- Qwen2.5-3B-Q3_K_M.gguf +2 -2
- Qwen2.5-3B-Q3_K_S.gguf +2 -2
- Qwen2.5-3B-Q4_0.gguf +2 -2
- Qwen2.5-3B-Q4_K_M.gguf +2 -2
- Qwen2.5-3B-Q4_K_S.gguf +2 -2
- Qwen2.5-3B-Q5_0.gguf +2 -2
- Qwen2.5-3B-Q5_K_M.gguf +2 -2
- Qwen2.5-3B-Q5_K_S.gguf +2 -2
- Qwen2.5-3B-Q6_K.gguf +2 -2
- Qwen2.5-3B-Q8_0.gguf +2 -2
- README.md +20 -26
Qwen2.5-3B-Q2_K.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7514076ac1ea0d62a2cc8b0985d223056137ecb90bf897df936d5cb2488fd182
|
3 |
+
size 1274753280
|
Qwen2.5-3B-Q3_K_L.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3ee4ce06df941d1e48c33d452f488af49ba1a5da36a9f218970a71f8af0899d9
|
3 |
+
size 1707389184
|
Qwen2.5-3B-Q3_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b6c08f5c12075cb7705d1a1f2620b6b03aeb3d479265531ecc2e47bc9861246
|
3 |
+
size 1590472960
|
Qwen2.5-3B-Q3_K_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:952f6770250b6f6aff8a5a8c09216e8c50c7d1fc03d07bfe2232be3fccc16913
|
3 |
+
size 1454354688
|
Qwen2.5-3B-Q4_0.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71252fe8f31368e45f4d5e586c53dba9d57e3f036e2f5451d243b8fa69864902
|
3 |
+
size 1822847232
|
Qwen2.5-3B-Q4_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2851fd1cca4b0bdbfc17608deedaef98d6fdee8bbc60f348455e0070ce4d3a56
|
3 |
+
size 1929900288
|
Qwen2.5-3B-Q4_K_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:282d95058b77661e79cd084ea6eadd9eef0e549afb97776f00fed9689061df11
|
3 |
+
size 1834381568
|
Qwen2.5-3B-Q5_0.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24ee1065e37544d28b55dd7fe40db22de60a14a6ca2e0341f3eeebf4fa1b9754
|
3 |
+
size 2169663744
|
Qwen2.5-3B-Q5_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ff612358ef4f432f705d55914242d32fe3c36e4bb2366fcb1979cbaae8f8a54
|
3 |
+
size 2224812288
|
Qwen2.5-3B-Q5_K_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ece7ecdcacac14b413068857ead8d7178533df38b0a3393b93eb5c15077c3d1d
|
3 |
+
size 2169663744
|
Qwen2.5-3B-Q6_K.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc3dde5b4ed32f5a33d1ed127d3615c0fec5ff692007c722b8969e1ff0fb4b1a
|
3 |
+
size 2538156288
|
Qwen2.5-3B-Q8_0.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:166504c09d1f2b9d0393b858817bf13982ccc50d1e97810752e14d403fb32685
|
3 |
+
size 3285473536
|
README.md
CHANGED
@@ -1,12 +1,12 @@
|
|
1 |
---
|
2 |
-
|
3 |
-
license_name: qwen-research
|
4 |
-
license_link: https://huggingface.co/Qwen/Qwen2.5-3B/blob/main/LICENSE
|
5 |
language:
|
6 |
- en
|
7 |
-
|
8 |
-
|
9 |
tags:
|
|
|
|
|
10 |
- TensorBlock
|
11 |
- GGUF
|
12 |
---
|
@@ -22,13 +22,12 @@ tags:
|
|
22 |
</div>
|
23 |
</div>
|
24 |
|
25 |
-
##
|
26 |
|
27 |
-
This repo contains GGUF format model files for [
|
28 |
|
29 |
The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
|
30 |
|
31 |
-
|
32 |
<div style="text-align: left; margin: 20px 0;">
|
33 |
<a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
|
34 |
Run them on the TensorBlock client using your local machine ↗
|
@@ -37,31 +36,26 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
|
|
37 |
|
38 |
## Prompt template
|
39 |
|
40 |
-
|
41 |
```
|
42 |
-
|
43 |
-
{system_prompt}<|im_end|>
|
44 |
-
<|im_start|>user
|
45 |
-
{prompt}<|im_end|>
|
46 |
-
<|im_start|>assistant
|
47 |
```
|
48 |
|
49 |
## Model file specification
|
50 |
|
51 |
| Filename | Quant type | File Size | Description |
|
52 |
| -------- | ---------- | --------- | ----------- |
|
53 |
-
| [Qwen2.5-3B-Q2_K.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q2_K.gguf) | Q2_K | 1.
|
54 |
-
| [Qwen2.5-3B-Q3_K_S.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q3_K_S.gguf) | Q3_K_S | 1.
|
55 |
-
| [Qwen2.5-3B-Q3_K_M.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q3_K_M.gguf) | Q3_K_M | 1.
|
56 |
-
| [Qwen2.5-3B-Q3_K_L.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q3_K_L.gguf) | Q3_K_L | 1.
|
57 |
-
| [Qwen2.5-3B-Q4_0.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q4_0.gguf) | Q4_0 | 1.
|
58 |
-
| [Qwen2.5-3B-Q4_K_S.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q4_K_S.gguf) | Q4_K_S | 1.
|
59 |
-
| [Qwen2.5-3B-Q4_K_M.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q4_K_M.gguf) | Q4_K_M | 1.
|
60 |
-
| [Qwen2.5-3B-Q5_0.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q5_0.gguf) | Q5_0 | 2.
|
61 |
-
| [Qwen2.5-3B-Q5_K_S.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q5_K_S.gguf) | Q5_K_S | 2.
|
62 |
-
| [Qwen2.5-3B-Q5_K_M.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q5_K_M.gguf) | Q5_K_M | 2.
|
63 |
-
| [Qwen2.5-3B-Q6_K.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q6_K.gguf) | Q6_K | 2.
|
64 |
-
| [Qwen2.5-3B-Q8_0.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q8_0.gguf) | Q8_0 | 3.
|
65 |
|
66 |
|
67 |
## Downloading instruction
|
|
|
1 |
---
|
2 |
+
base_model: unsloth/Qwen2.5-3B
|
|
|
|
|
3 |
language:
|
4 |
- en
|
5 |
+
library_name: transformers
|
6 |
+
license: other
|
7 |
tags:
|
8 |
+
- unsloth
|
9 |
+
- transformers
|
10 |
- TensorBlock
|
11 |
- GGUF
|
12 |
---
|
|
|
22 |
</div>
|
23 |
</div>
|
24 |
|
25 |
+
## unsloth/Qwen2.5-3B - GGUF
|
26 |
|
27 |
+
This repo contains GGUF format model files for [unsloth/Qwen2.5-3B](https://huggingface.co/unsloth/Qwen2.5-3B).
|
28 |
|
29 |
The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
|
30 |
|
|
|
31 |
<div style="text-align: left; margin: 20px 0;">
|
32 |
<a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
|
33 |
Run them on the TensorBlock client using your local machine ↗
|
|
|
36 |
|
37 |
## Prompt template
|
38 |
|
|
|
39 |
```
|
40 |
+
|
|
|
|
|
|
|
|
|
41 |
```
|
42 |
|
43 |
## Model file specification
|
44 |
|
45 |
| Filename | Quant type | File Size | Description |
|
46 |
| -------- | ---------- | --------- | ----------- |
|
47 |
+
| [Qwen2.5-3B-Q2_K.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q2_K.gguf) | Q2_K | 1.275 GB | smallest, significant quality loss - not recommended for most purposes |
|
48 |
+
| [Qwen2.5-3B-Q3_K_S.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q3_K_S.gguf) | Q3_K_S | 1.454 GB | very small, high quality loss |
|
49 |
+
| [Qwen2.5-3B-Q3_K_M.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q3_K_M.gguf) | Q3_K_M | 1.590 GB | very small, high quality loss |
|
50 |
+
| [Qwen2.5-3B-Q3_K_L.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q3_K_L.gguf) | Q3_K_L | 1.707 GB | small, substantial quality loss |
|
51 |
+
| [Qwen2.5-3B-Q4_0.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q4_0.gguf) | Q4_0 | 1.823 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
|
52 |
+
| [Qwen2.5-3B-Q4_K_S.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q4_K_S.gguf) | Q4_K_S | 1.834 GB | small, greater quality loss |
|
53 |
+
| [Qwen2.5-3B-Q4_K_M.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q4_K_M.gguf) | Q4_K_M | 1.930 GB | medium, balanced quality - recommended |
|
54 |
+
| [Qwen2.5-3B-Q5_0.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q5_0.gguf) | Q5_0 | 2.170 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
|
55 |
+
| [Qwen2.5-3B-Q5_K_S.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q5_K_S.gguf) | Q5_K_S | 2.170 GB | large, low quality loss - recommended |
|
56 |
+
| [Qwen2.5-3B-Q5_K_M.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q5_K_M.gguf) | Q5_K_M | 2.225 GB | large, very low quality loss - recommended |
|
57 |
+
| [Qwen2.5-3B-Q6_K.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q6_K.gguf) | Q6_K | 2.538 GB | very large, extremely low quality loss |
|
58 |
+
| [Qwen2.5-3B-Q8_0.gguf](https://huggingface.co/tensorblock/Qwen2.5-3B-GGUF/blob/main/Qwen2.5-3B-Q8_0.gguf) | Q8_0 | 3.285 GB | very large, extremely low quality loss - not recommended |
|
59 |
|
60 |
|
61 |
## Downloading instruction
|