tsqn commited on
Commit
14016d0
·
verified ·
1 Parent(s): 44b8b77

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -1,36 +1,37 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- wizardlm-13b-v1.2-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wizardlm-13b-v1.2-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
37
+ *.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,51 +1,51 @@
1
- ---
2
- license: llama2
3
- base_model: WizardLMTeam/WizardLM-13B-V1.2
4
- tags:
5
- - llama-cpp
6
- - gguf-my-repo
7
- ---
8
-
9
- # tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF
10
- This model was converted to GGUF format from [`WizardLMTeam/WizardLM-13B-V1.2`](https://huggingface.co/WizardLMTeam/WizardLM-13B-V1.2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
11
- Refer to the [original model card](https://huggingface.co/WizardLMTeam/WizardLM-13B-V1.2) for more details on the model.
12
-
13
- ## Use with llama.cpp
14
- Install llama.cpp through brew (works on Mac and Linux)
15
-
16
- ```bash
17
- brew install llama.cpp
18
-
19
- ```
20
- Invoke the llama.cpp server or the CLI.
21
-
22
- ### CLI:
23
- ```bash
24
- llama-cli --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -p "The meaning to life and the universe is"
25
- ```
26
-
27
- ### Server:
28
- ```bash
29
- llama-server --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -c 2048
30
- ```
31
-
32
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
33
-
34
- Step 1: Clone llama.cpp from GitHub.
35
- ```
36
- git clone https://github.com/ggerganov/llama.cpp
37
- ```
38
-
39
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
40
- ```
41
- cd llama.cpp && LLAMA_CURL=1 make
42
- ```
43
-
44
- Step 3: Run inference through the main binary.
45
- ```
46
- ./llama-cli --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -p "The meaning to life and the universe is"
47
- ```
48
- or
49
- ```
50
- ./llama-server --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -c 2048
51
- ```
 
1
+ ---
2
+ license: llama2
3
+ base_model: WizardLMTeam/WizardLM-13B-V1.2
4
+ tags:
5
+ - llama-cpp
6
+ - gguf-my-repo
7
+ ---
8
+
9
+ # tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF
10
+ This model was converted to GGUF format from [`WizardLMTeam/WizardLM-13B-V1.2`](https://huggingface.co/WizardLMTeam/WizardLM-13B-V1.2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
11
+ Refer to the [original model card](https://huggingface.co/WizardLMTeam/WizardLM-13B-V1.2) for more details on the model.
12
+
13
+ ## Use with llama.cpp
14
+ Install llama.cpp through brew (works on Mac and Linux)
15
+
16
+ ```bash
17
+ brew install llama.cpp
18
+
19
+ ```
20
+ Invoke the llama.cpp server or the CLI.
21
+
22
+ ### CLI:
23
+ ```bash
24
+ llama-cli --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -p "The meaning to life and the universe is"
25
+ ```
26
+
27
+ ### Server:
28
+ ```bash
29
+ llama-server --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -c 2048
30
+ ```
31
+
32
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
33
+
34
+ Step 1: Clone llama.cpp from GitHub.
35
+ ```
36
+ git clone https://github.com/ggerganov/llama.cpp
37
+ ```
38
+
39
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
40
+ ```
41
+ cd llama.cpp && LLAMA_CURL=1 make
42
+ ```
43
+
44
+ Step 3: Run inference through the main binary.
45
+ ```
46
+ ./llama-cli --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -p "The meaning to life and the universe is"
47
+ ```
48
+ or
49
+ ```
50
+ ./llama-server --hf-repo tsqn/WizardLM-13B-V1.2-Q4_K_M-GGUF --hf-file wizardlm-13b-v1.2-q4_k_m.gguf -c 2048
51
+ ```
wizardlm-13b-v1.2-q6_k-00001-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64f071e114bf20c5edf35c58b7a4b0706b53e8d379d5dfbf827dd8d2ee8513a0
3
+ size 1975081824
wizardlm-13b-v1.2-q6_k-00002-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd78a5271b2e756b793c4abb47f00dd35358c8c378cc99dabec4d132df022f22
3
+ size 1980850304
wizardlm-13b-v1.2-q6_k-00003-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7600b3aad58b6045c05c56d43e2d8137b9d52f354504bce8a53b8ee796306e0f
3
+ size 1980829792
wizardlm-13b-v1.2-q6_k-00004-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:244764c2bbd1d3a501081474f4484a3d9d8953131dcb9d2f8a5ef87b0541725b
3
+ size 1965777056
wizardlm-13b-v1.2-q6_k-00005-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66c084821cc49100bc951d8c5c3261c54d3e5b2dbbb6d27659a62858c4bb141e
3
+ size 1995882528
wizardlm-13b-v1.2-q6_k-00006-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09cc985bf4fb279091502e728046a8989e54df25f0f68577169603aa423dc369
3
+ size 780719808