v000000 commited on
Commit
6b49cd3
·
verified ·
1 Parent(s): c41b1de

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -42
README.md CHANGED
@@ -5,49 +5,72 @@ tags:
5
  - mergekit
6
  - merge
7
  - llama-cpp
8
- - gguf-my-repo
9
  ---
10
 
11
- # v000000/SwallowMaid-8B-L3-SPPO-abliterated-Q8_0-GGUF
12
- This model was converted to GGUF format from [`v000000/SwallowMaid-8B-L3-SPPO-abliterated`](https://huggingface.co/v000000/SwallowMaid-8B-L3-SPPO-abliterated) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/v000000/SwallowMaid-8B-L3-SPPO-abliterated) for more details on the model.
14
 
15
- ## Use with llama.cpp
16
- Install llama.cpp through brew (works on Mac and Linux)
17
-
18
- ```bash
19
- brew install llama.cpp
20
-
21
- ```
22
- Invoke the llama.cpp server or the CLI.
23
-
24
- ### CLI:
25
- ```bash
26
- llama-cli --hf-repo v000000/SwallowMaid-8B-L3-SPPO-abliterated-Q8_0-GGUF --hf-file swallowmaid-8b-l3-sppo-abliterated-q8_0.gguf -p "The meaning to life and the universe is"
27
- ```
28
-
29
- ### Server:
30
- ```bash
31
- llama-server --hf-repo v000000/SwallowMaid-8B-L3-SPPO-abliterated-Q8_0-GGUF --hf-file swallowmaid-8b-l3-sppo-abliterated-q8_0.gguf -c 2048
32
- ```
33
-
34
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
35
-
36
- Step 1: Clone llama.cpp from GitHub.
37
- ```
38
- git clone https://github.com/ggerganov/llama.cpp
39
- ```
40
-
41
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
42
- ```
43
- cd llama.cpp && LLAMA_CURL=1 make
44
- ```
45
-
46
- Step 3: Run inference through the main binary.
47
- ```
48
- ./llama-cli --hf-repo v000000/SwallowMaid-8B-L3-SPPO-abliterated-Q8_0-GGUF --hf-file swallowmaid-8b-l3-sppo-abliterated-q8_0.gguf -p "The meaning to life and the universe is"
49
- ```
50
- or
51
- ```
52
- ./llama-server --hf-repo v000000/SwallowMaid-8B-L3-SPPO-abliterated-Q8_0-GGUF --hf-file swallowmaid-8b-l3-sppo-abliterated-q8_0.gguf -c 2048
53
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  - mergekit
6
  - merge
7
  - llama-cpp
8
+ - llama
9
  ---
10
 
11
+ This model was converted to GGUF format from [`v000000/SwallowMaid-8B-L3-SPPO-abliterated`](https://huggingface.co/v000000/SwallowMaid-8B-L3-SPPO-abliterated) using llama.cpp
 
12
  Refer to the [original model card](https://huggingface.co/v000000/SwallowMaid-8B-L3-SPPO-abliterated) for more details on the model.
13
 
14
+ # merge
15
+
16
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
+
18
+ ## Merge Details
19
+ ### Merge Method
20
+
21
+ This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method.
22
+
23
+ ### Models Merged
24
+
25
+ The following models were included in the merge:
26
+ * [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
27
+ * [UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3](https://huggingface.co/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3)
28
+ * [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS)
29
+ * [maldv/llama-3-fantasy-writer-8b](https://huggingface.co/maldv/llama-3-fantasy-writer-8b)
30
+ * [tokyotech-llm/Llama-3-Swallow-8B-v0.1](https://huggingface.co/tokyotech-llm/Llama-3-Swallow-8B-v0.1)
31
+ * [Nitral-AI/Hathor_Respawn-L3-8B-v0.8](https://huggingface.co/Nitral-AI/Hathor_Respawn-L3-8B-v0.8)
32
+
33
+ ### Configuration
34
+
35
+ The following YAML configuration was used to produce this model:
36
+
37
+ ```yaml
38
+ # Part 3, Apply abliteration (SwallowMaid-8B)
39
+ models:
40
+ - model: sppo-rpmix-part2+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
41
+ parameters:
42
+ weight: 1.0
43
+ merge_method: linear
44
+ dtype: float32
45
+
46
+ # Part 2, infuse 35% swallow+rpmix to SPPO-Iter3 (sppo-rpmix-part2)
47
+ models:
48
+ - model: UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
49
+ parameters:
50
+ weight: 1.0
51
+ - model: rpmix-part1
52
+ parameters:
53
+ weight: 0.35
54
+ merge_method: task_arithmetic
55
+ base_model: UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
56
+ parameters:
57
+ normalize: false
58
+ dtype: float32
59
+
60
+ # Part 1, linear merge rpmix (rpmix-part1)
61
+ models:
62
+ - model: Nitral-AI/Hathor_Respawn-L3-8B-v0.8
63
+ parameters:
64
+ weight: 0.6
65
+ - model: maldv/llama-3-fantasy-writer-8b
66
+ parameters:
67
+ weight: 0.1
68
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
69
+ parameters:
70
+ weight: 0.4
71
+ - model: tokyotech-llm/Llama-3-Swallow-8B-v0.1
72
+ parameters:
73
+ weight: 0.15
74
+ merge_method: linear
75
+ dtype: float32
76
+ ```