grimjim commited on
Commit
79afa4f
·
1 Parent(s): 03fecc4

Initial release

Browse files
.gitattributes CHANGED
@@ -4,6 +4,7 @@
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
 
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gguf filter=lfs diff=lfs merge=lfs -text
8
  *.gz filter=lfs diff=lfs merge=lfs -text
9
  *.h5 filter=lfs diff=lfs merge=lfs -text
10
  *.joblib filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - grimjim/SauerHuatuoSkywork-o1-Llama-3.1-8B
4
+ library_name: transformers
5
+ pipeline_tag: text-generation
6
+ quanted_by: grimjim
7
+ license: llama3.1
8
+ model-index:
9
+ - name: SauerHuatuoSkywork-o1-Llama-3.1-8B
10
+ results:
11
+ - task:
12
+ type: text-generation
13
+ name: Text Generation
14
+ dataset:
15
+ name: IFEval (0-Shot)
16
+ type: wis-k/instruction-following-eval
17
+ split: train
18
+ args:
19
+ num_few_shot: 0
20
+ metrics:
21
+ - type: inst_level_strict_acc and prompt_level_strict_acc
22
+ value: 52.19
23
+ name: averaged accuracy
24
+ source:
25
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B
26
+ name: Open LLM Leaderboard
27
+ - task:
28
+ type: text-generation
29
+ name: Text Generation
30
+ dataset:
31
+ name: BBH (3-Shot)
32
+ type: SaylorTwift/bbh
33
+ split: test
34
+ args:
35
+ num_few_shot: 3
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 32.09
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: MATH Lvl 5 (4-Shot)
48
+ type: lighteval/MATH-Hard
49
+ split: test
50
+ args:
51
+ num_few_shot: 4
52
+ metrics:
53
+ - type: exact_match
54
+ value: 16.99
55
+ name: exact match
56
+ source:
57
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: GPQA (0-shot)
64
+ type: Idavidrein/gpqa
65
+ split: train
66
+ args:
67
+ num_few_shot: 0
68
+ metrics:
69
+ - type: acc_norm
70
+ value: 9.51
71
+ name: acc_norm
72
+ source:
73
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B
74
+ name: Open LLM Leaderboard
75
+ - task:
76
+ type: text-generation
77
+ name: Text Generation
78
+ dataset:
79
+ name: MuSR (0-shot)
80
+ type: TAUR-Lab/MuSR
81
+ args:
82
+ num_few_shot: 0
83
+ metrics:
84
+ - type: acc_norm
85
+ value: 15.79
86
+ name: acc_norm
87
+ source:
88
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B
89
+ name: Open LLM Leaderboard
90
+ - task:
91
+ type: text-generation
92
+ name: Text Generation
93
+ dataset:
94
+ name: MMLU-PRO (5-shot)
95
+ type: TIGER-Lab/MMLU-Pro
96
+ config: main
97
+ split: test
98
+ args:
99
+ num_few_shot: 5
100
+ metrics:
101
+ - type: acc
102
+ value: 33.23
103
+ name: accuracy
104
+ source:
105
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B
106
+ name: Open LLM Leaderboard
107
+ ---
108
+ # SauerHuatuoSkywork-o1-Llama-3.1-8B
109
+
110
+ This repo contains GGUF quants of a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
111
+
112
+ An experiment to hybridize a relatively high scoring Llama 3.1 8B model with o1 reasoning capabilities.
113
+
114
+ Although IFEval benched lower than the SauerkrautLM mode, every other benchmark improved from the addition of the o1 merge at low weight.
115
+
116
+ Made with Llama.
117
+
118
+ ## Merge Details
119
+ ### Merge Method
120
+
121
+ This model was merged using the SLERP merge method.
122
+
123
+ ### Models Merged
124
+
125
+ The following models were included in the merge:
126
+ * [grimjim/HuatuoSkywork-o1-Llama-3.1-8B](https://huggingface.co/grimjim/HuatuoSkywork-o1-Llama-3.1-8B)
127
+ * [VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct)
128
+
129
+ ### Configuration
130
+
131
+ The following YAML configuration was used to produce this model:
132
+
133
+ ```yaml
134
+ models:
135
+ - model: grimjim/HuatuoSkywork-o1-Llama-3.1-8B
136
+ - model: VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
137
+ merge_method: slerp
138
+ base_model: grimjim/HuatuoSkywork-o1-Llama-3.1-8B
139
+ parameters:
140
+ t:
141
+ - value: 0.96
142
+ dtype: bfloat16
143
+
144
+ ```
145
+
146
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
147
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/grimjim__SauerHuatuoSkywork-o1-Llama-3.1-8B-details)!
148
+ Summarized results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/contents/viewer/default/train?q=grimjim%2FSauerHuatuoSkywork-o1-Llama-3.1-8B&sort[column]=Average%20%E2%AC%86%EF%B8%8F&sort[direction]=desc)!
149
+
150
+ | Metric |Value (%)|
151
+ |-------------------|--------:|
152
+ |**Average** | 26.63|
153
+ |IFEval (0-Shot) | 52.19|
154
+ |BBH (3-Shot) | 32.09|
155
+ |MATH Lvl 5 (4-Shot)| 16.99|
156
+ |GPQA (0-shot) | 9.51|
157
+ |MuSR (0-shot) | 15.79|
158
+ |MMLU-PRO (5-shot) | 33.23|
159
+
SauerHuatuoSkywork-o1-Llama-3.1-8B.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acacf0b9db7f7cd0e9bb704452f02c04185b0a965d1223a29bb146d5aa7166f5
3
+ size 4920739488
SauerHuatuoSkywork-o1-Llama-3.1-8B.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31cebf5bf3c9c94398e50d8b189b5585e27eda13661ffb8234dfc70ac72d441b
3
+ size 5732992672
SauerHuatuoSkywork-o1-Llama-3.1-8B.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d18b92f64215e5d77ea172c8d1d86e06d167845bb7de9044f04cab88ad2c7558
3
+ size 6596011680
SauerHuatuoSkywork-o1-Llama-3.1-8B.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d24ef119724ded747ce0250885cd0578f78c962fe8454a5592143ee8a2073ea0
3
+ size 8540776096