eval_name
stringlengths 12
111
| Precision
stringclasses 3
values | Type
stringclasses 6
values | T
stringclasses 6
values | Weight type
stringclasses 2
values | Architecture
stringclasses 52
values | Model
stringlengths 355
689
| fullname
stringlengths 4
102
| Model sha
stringlengths 0
40
| Average ⬆️
float64 1.03
52
| Hub License
stringclasses 26
values | Hub ❤️
int64 0
5.9k
| #Params (B)
int64 -1
140
| Available on the hub
bool 2
classes | MoE
bool 2
classes | Flagged
bool 2
classes | Chat Template
bool 2
classes | CO₂ cost (kg)
float64 0.03
107
| IFEval Raw
float64 0
0.9
| IFEval
float64 0
90
| BBH Raw
float64 0.27
0.75
| BBH
float64 0.81
63.5
| MATH Lvl 5 Raw
float64 0
0.51
| MATH Lvl 5
float64 0
50.7
| GPQA Raw
float64 0.22
0.44
| GPQA
float64 0
24.9
| MUSR Raw
float64 0.29
0.6
| MUSR
float64 0
38.5
| MMLU-PRO Raw
float64 0.1
0.73
| MMLU-PRO
float64 0
70
| Merged
bool 2
classes | Official Providers
bool 2
classes | Upload To Hub Date
stringclasses 424
values | Submission Date
stringclasses 169
values | Generation
int64 0
10
| Base Model
stringlengths 4
102
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Xclbr7_Hyena-12b_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/Xclbr7/Hyena-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Xclbr7/Hyena-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Xclbr7__Hyena-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Xclbr7/Hyena-12b | 9dd5eb77ce8e0e05e260ae4d812631fb980527fa | 20.437243 | apache-2.0 | 1 | 12 | true | false | false | false | 1.859893 | 0.340446 | 34.044557 | 0.545718 | 34.665649 | 0.093656 | 9.365559 | 0.297819 | 6.375839 | 0.398427 | 11.070052 | 0.343916 | 27.101803 | false | false | 2024-09-19 | 2024-09-19 | 1 | Xclbr7/Arcanum-12b |
Xclbr7_caliburn-12b_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/Xclbr7/caliburn-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Xclbr7/caliburn-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Xclbr7__caliburn-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Xclbr7/caliburn-12b | f76fa67c7ca8bf7e75540baf55972ba52a46630b | 22.808395 | mit | 0 | 12 | true | false | false | false | 1.85622 | 0.357631 | 35.763109 | 0.551863 | 35.636841 | 0.10423 | 10.422961 | 0.336409 | 11.521253 | 0.429188 | 13.781771 | 0.36752 | 29.724439 | false | false | 2024-09-14 | 2024-09-14 | 0 | Xclbr7/caliburn-12b |
Xclbr7_caliburn-v2-12b_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/Xclbr7/caliburn-v2-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Xclbr7/caliburn-v2-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Xclbr7__caliburn-v2-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Xclbr7/caliburn-v2-12b | fa736b3b852298dd8c047ac6dcc620161df4a79b | 20.953525 | mit | 0 | 12 | true | false | false | false | 1.632194 | 0.296682 | 29.668169 | 0.514143 | 30.387967 | 0.10423 | 10.422961 | 0.326342 | 10.178971 | 0.437031 | 14.128906 | 0.378408 | 30.934176 | false | false | 2024-09-16 | 2024-09-16 | 0 | Xclbr7/caliburn-v2-12b |
Yash21_TinyYi-7B-Test_float16 | float16 | 🟢 pretrained | 🟢 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/Yash21/TinyYi-7B-Test" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Yash21/TinyYi-7B-Test</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Yash21__TinyYi-7B-Test-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Yash21/TinyYi-7B-Test | 7750e5de73fbcf1dcc0832b4cdabaa9713c20475 | 4.495167 | apache-2.0 | 0 | 6 | true | false | false | false | 0.763109 | 0.185649 | 18.564852 | 0.29098 | 2.267966 | 0 | 0 | 0.264262 | 1.901566 | 0.336448 | 3.222656 | 0.109126 | 1.013963 | true | false | 2024-01-06 | 2024-07-03 | 0 | Yash21/TinyYi-7B-Test |
Youlln_1PARAMMYL-8B-ModelStock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/1PARAMMYL-8B-ModelStock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/1PARAMMYL-8B-ModelStock</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__1PARAMMYL-8B-ModelStock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/1PARAMMYL-8B-ModelStock | 4ce556da5ccd1ecac8d0f3e1e94d1982f11b910d | 26.283975 | 0 | 8 | false | false | false | false | 0.892522 | 0.537134 | 53.713369 | 0.521584 | 31.799951 | 0.147281 | 14.728097 | 0.323826 | 9.8434 | 0.440938 | 14.283854 | 0.400017 | 33.33518 | false | false | 2024-09-20 | 2024-09-20 | 1 | Youlln/1PARAMMYL-8B-ModelStock (Merge) |
|
Youlln_2PRYMMAL-Yi1.5-6B-SLERP_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/2PRYMMAL-Yi1.5-6B-SLERP" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/2PRYMMAL-Yi1.5-6B-SLERP</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__2PRYMMAL-Yi1.5-6B-SLERP-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/2PRYMMAL-Yi1.5-6B-SLERP | b776bd3ce6784b96ff928b1d5ad51b2991909f2c | 18.979223 | apache-2.0 | 0 | 6 | true | false | false | false | 1.383177 | 0.282594 | 28.259352 | 0.466475 | 24.495644 | 0.112538 | 11.253776 | 0.307047 | 7.606264 | 0.475604 | 18.150521 | 0.316988 | 24.109781 | true | false | 2024-09-22 | 2024-09-23 | 1 | Youlln/2PRYMMAL-Yi1.5-6B-SLERP (Merge) |
Youlln_3PRYMMAL-PHI3-3B-SLERP_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/3PRYMMAL-PHI3-3B-SLERP" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/3PRYMMAL-PHI3-3B-SLERP</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__3PRYMMAL-PHI3-3B-SLERP-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/3PRYMMAL-PHI3-3B-SLERP | 9396bcf1709ac8360a95a746482520fab4295706 | 24.849214 | apache-2.0 | 0 | 3 | true | false | false | false | 1.078439 | 0.36555 | 36.555007 | 0.542183 | 35.827668 | 0.154079 | 15.407855 | 0.326342 | 10.178971 | 0.464844 | 17.772135 | 0.400183 | 33.35365 | true | false | 2024-09-23 | 2024-09-23 | 1 | Youlln/3PRYMMAL-PHI3-3B-SLERP (Merge) |
Youlln_4PRYMMAL-GEMMA2-9B-SLERP_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/4PRYMMAL-GEMMA2-9B-SLERP" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/4PRYMMAL-GEMMA2-9B-SLERP</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__4PRYMMAL-GEMMA2-9B-SLERP-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/4PRYMMAL-GEMMA2-9B-SLERP | 7dac3b4ab4298113ae3103d63bb284e1ac8bf4d4 | 23.550238 | apache-2.0 | 0 | 9 | true | false | false | false | 2.85266 | 0.271377 | 27.137661 | 0.592253 | 42.064172 | 0.082326 | 8.232628 | 0.330537 | 10.738255 | 0.467198 | 17.466406 | 0.420961 | 35.662308 | true | false | 2024-09-23 | 2024-09-23 | 1 | Youlln/4PRYMMAL-GEMMA2-9B-SLERP (Merge) |
Youlln_ECE-PRYMMAL-0.5B-FT-V3_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-0.5B-FT-V3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-0.5B-FT-V3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-0.5B-FT-V3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-0.5B-FT-V3 | d542b4d53888fcc8e96c32892d47ec51afc9edc9 | 4.342504 | apache-2.0 | 0 | 0 | true | false | false | false | 0.573135 | 0.164191 | 16.419101 | 0.309313 | 3.616883 | 0 | 0 | 0.25755 | 1.006711 | 0.364448 | 3.222656 | 0.116107 | 1.789672 | false | false | 2024-10-16 | 2024-10-16 | 1 | Youlln/ECE-PRYMMAL-0.5B-FT-V3 (Merge) |
Youlln_ECE-PRYMMAL-0.5B-FT-V3-MUSR_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-0.5B-FT-V3-MUSR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-0.5B-FT-V3-MUSR</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-0.5B-FT-V3-MUSR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-0.5B-FT-V3-MUSR | 221dc80a1acd6f7dda0644699e6d61b90a5a0a05 | 5.551291 | apache-2.0 | 0 | 0 | true | false | false | false | 1.029985 | 0.15335 | 15.334978 | 0.304115 | 5.062186 | 0.024924 | 2.492447 | 0.249161 | 0 | 0.366031 | 3.253906 | 0.164478 | 7.164229 | false | false | 2024-10-21 | 2024-10-21 | 1 | Youlln/ECE-PRYMMAL-0.5B-FT-V3-MUSR (Merge) |
Youlln_ECE-PRYMMAL-0.5B-FT-V4-MUSR_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-0.5B-FT-V4-MUSR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-0.5B-FT-V4-MUSR</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-0.5B-FT-V4-MUSR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-0.5B-FT-V4-MUSR | f5b268d63bb10f05a229da4f2ee9cb0882c93971 | 4.198599 | apache-2.0 | 0 | 0 | true | false | false | false | 0.948691 | 0.113757 | 11.375705 | 0.303836 | 4.949092 | 0.011329 | 1.132931 | 0.270134 | 2.684564 | 0.352885 | 1.477344 | 0.132148 | 3.571956 | false | false | 2024-10-21 | 2024-10-21 | 1 | Youlln/ECE-PRYMMAL-0.5B-FT-V4-MUSR (Merge) |
Youlln_ECE-PRYMMAL-0.5B-SLERP-V2_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-0.5B-SLERP-V2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-0.5B-SLERP-V2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-0.5B-SLERP-V2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-0.5B-SLERP-V2 | 5e87669abcdc042774a63b94a13880f1acd6e15d | 4.614607 | apache-2.0 | 0 | 0 | true | false | false | false | 0.654997 | 0.161193 | 16.119341 | 0.293477 | 1.917561 | 0 | 0 | 0.274329 | 3.243848 | 0.383115 | 5.35599 | 0.109458 | 1.050901 | false | false | 2024-10-22 | 2024-10-22 | 1 | Youlln/ECE-PRYMMAL-0.5B-SLERP-V2 (Merge) |
Youlln_ECE-PRYMMAL-0.5B-SLERP-V3_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-0.5B-SLERP-V3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-0.5B-SLERP-V3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-0.5B-SLERP-V3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-0.5B-SLERP-V3 | 94bfab3b1f41458427e5f8598ceb3ec731ba1bd6 | 3.663014 | apache-2.0 | 0 | 0 | true | false | false | false | 0.637698 | 0.167014 | 16.701352 | 0.293838 | 2.319605 | 0 | 0 | 0.251678 | 0.223714 | 0.354125 | 1.765625 | 0.10871 | 0.96779 | false | false | 2024-10-22 | 2024-10-22 | 0 | Youlln/ECE-PRYMMAL-0.5B-SLERP-V3 |
Youlln_ECE-PRYMMAL-YL-1B-SLERP-V1_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-YL-1B-SLERP-V1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1 | b5cd268edb0cc5c2c6ab2c49c950e611b2b8138c | 16.404998 | apache-2.0 | 0 | 1 | true | false | false | false | 0.595341 | 0.325108 | 32.510849 | 0.420851 | 18.279511 | 0.090634 | 9.063444 | 0.291107 | 5.480984 | 0.426583 | 11.589583 | 0.293551 | 21.505615 | false | false | 2024-11-08 | 2024-11-08 | 0 | Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1 |
Youlln_ECE-PRYMMAL-YL-1B-SLERP-V2_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-YL-1B-SLERP-V2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2 | 3559f643c8d5774135a1cd8daea78fef31035679 | 16.404998 | apache-2.0 | 0 | 1 | true | false | false | false | 0.604628 | 0.325108 | 32.510849 | 0.420851 | 18.279511 | 0.090634 | 9.063444 | 0.291107 | 5.480984 | 0.426583 | 11.589583 | 0.293551 | 21.505615 | false | false | 2024-11-08 | 2024-11-08 | 0 | Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2 |
Youlln_ECE-PRYMMAL-YL-7B-SLERP-V4_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL-YL-7B-SLERP-V4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL-YL-7B-SLERP-V4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL-YL-7B-SLERP-V4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL-YL-7B-SLERP-V4 | 4939b9e24be6f03d5df1e9bb7dc1b4fd5d59404a | 10.416375 | apache-2.0 | 0 | 7 | true | false | false | false | 0.770157 | 0.25097 | 25.096965 | 0.376973 | 13.157437 | 0.026435 | 2.643505 | 0.265101 | 2.013423 | 0.37449 | 7.011198 | 0.213182 | 12.575724 | false | false | 2024-11-06 | 2024-11-06 | 0 | Youlln/ECE-PRYMMAL-YL-7B-SLERP-V4 |
Youlln_ECE-PRYMMAL0.5-FT_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL0.5-FT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL0.5-FT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL0.5-FT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL0.5-FT | 56b9fd5f26e5b6379fe4aa62e0f66b87b5c6f8e8 | 5.19551 | apache-2.0 | 0 | 0 | true | false | false | false | 0.503391 | 0.185073 | 18.507338 | 0.313209 | 5.1516 | 0 | 0 | 0.255872 | 0.782998 | 0.330125 | 1.432292 | 0.147689 | 5.298833 | false | false | 2024-10-02 | 2024-10-02 | 1 | Youlln/ECE-PRYMMAL0.5-FT (Merge) |
Youlln_ECE-PRYMMAL0.5B-Youri_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL0.5B-Youri" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL0.5B-Youri</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL0.5B-Youri-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL0.5B-Youri | 1477d3deff98f35f523aa222bc0442278d464566 | 3.505274 | 1 | 0 | false | false | false | false | 0.655846 | 0.144632 | 14.46318 | 0.281736 | 1.501296 | 0 | 0 | 0.243289 | 0 | 0.369656 | 4.007031 | 0.109541 | 1.060136 | false | false | 2024-10-07 | 2024-10-07 | 1 | Youlln/ECE-PRYMMAL0.5B-Youri (Merge) |
|
Youlln_ECE-PRYMMAL1B-FT-V1_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-PRYMMAL1B-FT-V1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-PRYMMAL1B-FT-V1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-PRYMMAL1B-FT-V1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-PRYMMAL1B-FT-V1 | d0fc3a6e93f91c8d586eb25c9f2a4ea4ca99e9f4 | 11.923327 | apache-2.0 | 0 | 1 | true | false | false | false | 0.737596 | 0.214375 | 21.437453 | 0.403265 | 16.189386 | 0.068731 | 6.873112 | 0.278523 | 3.803132 | 0.341656 | 3.873698 | 0.274269 | 19.36318 | false | false | 2024-10-12 | 2024-10-12 | 1 | Youlln/ECE-PRYMMAL1B-FT-V1 (Merge) |
Youlln_ECE-Qwen0.5B-FT-V2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE-Qwen0.5B-FT-V2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE-Qwen0.5B-FT-V2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE-Qwen0.5B-FT-V2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE-Qwen0.5B-FT-V2 | c87da3f19ab74854fca30f9ca71ce5c4884ef629 | 7.48657 | apache-2.0 | 0 | 0 | true | false | false | false | 0.5233 | 0.252593 | 25.259312 | 0.328971 | 7.632148 | 0.015106 | 1.510574 | 0.266779 | 2.237136 | 0.306281 | 0.885156 | 0.166556 | 7.395095 | false | false | 2024-10-11 | 2024-10-11 | 1 | Youlln/ECE-Qwen0.5B-FT-V2 (Merge) |
Youlln_ECE.EIFFEIL.ia-0.5B-SLERP_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/Youlln/ECE.EIFFEIL.ia-0.5B-SLERP" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Youlln/ECE.EIFFEIL.ia-0.5B-SLERP</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Youlln__ECE.EIFFEIL.ia-0.5B-SLERP-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Youlln/ECE.EIFFEIL.ia-0.5B-SLERP | e376ce416af881eefa778d2566d15d9a6d29e7d9 | 8.716673 | apache-2.0 | 0 | 0 | true | false | false | false | 0.604059 | 0.25614 | 25.614037 | 0.330567 | 8.405356 | 0.05287 | 5.287009 | 0.265101 | 2.013423 | 0.310219 | 0.94401 | 0.190326 | 10.0362 | true | false | 2024-10-14 | 2024-10-14 | 1 | Youlln/ECE.EIFFEIL.ia-0.5B-SLERP (Merge) |
YoungPanda_qwenqwen_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2MoeForCausalLM | <a target="_blank" href="https://huggingface.co/YoungPanda/qwenqwen" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">YoungPanda/qwenqwen</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/YoungPanda__qwenqwen-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | YoungPanda/qwenqwen | 3b5d9b63076acc8988b8f7e9734cf1d78bb39c25 | 4.443749 | 0 | 14 | false | false | false | true | 7.122669 | 0.126397 | 12.639685 | 0.337899 | 8.19478 | 0.015106 | 1.510574 | 0.25 | 0 | 0.343365 | 2.453906 | 0.116772 | 1.863549 | false | false | 2024-09-12 | 0 | Removed |
||
Yuma42_KangalKhan-RawRuby-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/Yuma42/KangalKhan-RawRuby-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Yuma42/KangalKhan-RawRuby-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Yuma42__KangalKhan-RawRuby-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Yuma42/KangalKhan-RawRuby-7B | 54f56d4c6889eaf43fdd5f7d6dcef3c2ebe51929 | 20.440737 | apache-2.0 | 7 | 7 | true | false | false | true | 0.653719 | 0.547675 | 54.767461 | 0.475473 | 26.387284 | 0.063444 | 6.344411 | 0.287752 | 5.033557 | 0.394958 | 7.636458 | 0.302277 | 22.475251 | true | false | 2024-02-17 | 2024-06-26 | 1 | Yuma42/KangalKhan-RawRuby-7B (Merge) |
ZeroXClem_L3-Aspire-Heart-Matrix-8B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/ZeroXClem/L3-Aspire-Heart-Matrix-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZeroXClem/L3-Aspire-Heart-Matrix-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZeroXClem__L3-Aspire-Heart-Matrix-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZeroXClem/L3-Aspire-Heart-Matrix-8B | d63917595e911b077cff38109c74622c3ec41704 | 25.752283 | apache-2.0 | 3 | 8 | true | false | false | true | 0.787701 | 0.483353 | 48.335306 | 0.538421 | 34.307547 | 0.179003 | 17.900302 | 0.324664 | 9.955257 | 0.418708 | 13.071875 | 0.378491 | 30.94341 | true | false | 2024-11-21 | 2024-11-22 | 1 | ZeroXClem/L3-Aspire-Heart-Matrix-8B (Merge) |
ZeroXClem_Qwen-2.5-Aether-SlerpFusion-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/ZeroXClem/Qwen-2.5-Aether-SlerpFusion-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZeroXClem/Qwen-2.5-Aether-SlerpFusion-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZeroXClem__Qwen-2.5-Aether-SlerpFusion-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZeroXClem/Qwen-2.5-Aether-SlerpFusion-7B | 23992e1be9f77d767181dc7bcb42176395f42c30 | 29.589629 | apache-2.0 | 2 | 7 | true | false | false | true | 0.676354 | 0.62616 | 62.61597 | 0.546224 | 36.011209 | 0.241692 | 24.169184 | 0.298658 | 6.487696 | 0.417781 | 11.289323 | 0.43268 | 36.964391 | true | false | 2024-11-13 | 2024-11-20 | 1 | ZeroXClem/Qwen-2.5-Aether-SlerpFusion-7B (Merge) |
ZeroXClem_Qwen2.5-7B-HomerAnvita-NerdMix_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/ZeroXClem/Qwen2.5-7B-HomerAnvita-NerdMix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZeroXClem/Qwen2.5-7B-HomerAnvita-NerdMix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZeroXClem__Qwen2.5-7B-HomerAnvita-NerdMix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZeroXClem/Qwen2.5-7B-HomerAnvita-NerdMix | cd87a9d7c9a9c8950af84e1f4c72fff5d4625d8a | 34.169056 | apache-2.0 | 4 | 7 | true | false | false | true | 0.784193 | 0.770765 | 77.07649 | 0.554132 | 36.579206 | 0.295317 | 29.531722 | 0.319631 | 9.284116 | 0.439052 | 14.414844 | 0.443152 | 38.127955 | true | false | 2024-11-21 | 2024-11-21 | 1 | ZeroXClem/Qwen2.5-7B-HomerAnvita-NerdMix (Merge) |
ZeroXClem_Qwen2.5-7B-HomerCreative-Mix_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/ZeroXClem/Qwen2.5-7B-HomerCreative-Mix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZeroXClem/Qwen2.5-7B-HomerCreative-Mix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZeroXClem__Qwen2.5-7B-HomerCreative-Mix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZeroXClem/Qwen2.5-7B-HomerCreative-Mix | 6849553db73428ca67823a06f5cfeea660f77df8 | 34.353368 | apache-2.0 | 5 | 7 | true | false | false | true | 0.729571 | 0.783504 | 78.350443 | 0.554807 | 36.770722 | 0.323263 | 32.326284 | 0.299497 | 6.599553 | 0.434958 | 13.769792 | 0.444731 | 38.303413 | true | false | 2024-11-21 | 2024-11-21 | 1 | ZeroXClem/Qwen2.5-7B-HomerCreative-Mix (Merge) |
ZeroXClem_Qwen2.5-7B-Qandora-CySec_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/ZeroXClem/Qwen2.5-7B-Qandora-CySec" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZeroXClem/Qwen2.5-7B-Qandora-CySec</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZeroXClem__Qwen2.5-7B-Qandora-CySec-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZeroXClem/Qwen2.5-7B-Qandora-CySec | 6c8b513dbc61a9f704210d26124244f19f3bc4cc | 30.95351 | apache-2.0 | 3 | 7 | true | false | false | true | 0.68208 | 0.677317 | 67.73173 | 0.549002 | 36.264898 | 0.228852 | 22.885196 | 0.300336 | 6.711409 | 0.428604 | 13.408854 | 0.448471 | 38.718972 | true | false | 2024-11-12 | 2024-11-12 | 1 | ZeroXClem/Qwen2.5-7B-Qandora-CySec (Merge) |
ZeusLabs_L3-Aethora-15B-V2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/ZeusLabs/L3-Aethora-15B-V2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZeusLabs/L3-Aethora-15B-V2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZeusLabs__L3-Aethora-15B-V2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZeusLabs/L3-Aethora-15B-V2 | 2c601f116c37dd912c89357dbdbef879a637997e | 24.673537 | cc-by-sa-4.0 | 40 | 15 | true | false | false | true | 2.377669 | 0.720806 | 72.080635 | 0.501091 | 28.968505 | 0.079305 | 7.930514 | 0.287752 | 5.033557 | 0.387083 | 6.252083 | 0.349983 | 27.775931 | false | false | 2024-06-27 | 2024-06-27 | 1 | ZeusLabs/L3-Aethora-15B-V2 (Merge) |
ZhangShenao_SELM-Llama-3-8B-Instruct-iter-3_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/ZhangShenao/SELM-Llama-3-8B-Instruct-iter-3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ZhangShenao/SELM-Llama-3-8B-Instruct-iter-3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ZhangShenao__SELM-Llama-3-8B-Instruct-iter-3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ZhangShenao/SELM-Llama-3-8B-Instruct-iter-3 | 9c95ccdeceed14a3c2881bc495101a1acca1385f | 23.652706 | mit | 5 | 8 | true | false | false | true | 0.655591 | 0.690282 | 69.028179 | 0.504609 | 29.078531 | 0.062689 | 6.268882 | 0.258389 | 1.118568 | 0.38451 | 5.497135 | 0.378324 | 30.924941 | false | false | 2024-05-25 | 2024-07-02 | 3 | meta-llama/Meta-Llama-3-8B-Instruct |
aaditya_Llama3-OpenBioLLM-70B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/aaditya/Llama3-OpenBioLLM-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">aaditya/Llama3-OpenBioLLM-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/aaditya__Llama3-OpenBioLLM-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | aaditya/Llama3-OpenBioLLM-70B | 5f79deaf38bc5f662943d304d59cb30357e8e5bd | 35.004197 | llama3 | 355 | 70 | true | false | false | true | 9.657022 | 0.759674 | 75.967433 | 0.639887 | 47.147075 | 0.19864 | 19.864048 | 0.322987 | 9.731544 | 0.441719 | 14.348177 | 0.486702 | 42.966903 | false | false | 2024-04-24 | 2024-08-30 | 2 | meta-llama/Meta-Llama-3-70B |
abacusai_Dracarys-72B-Instruct_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Dracarys-72B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Dracarys-72B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Dracarys-72B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Dracarys-72B-Instruct | 10cabc4beb57a69df51533f65e39a7ad22821370 | 42.710042 | other | 19 | 72 | true | false | false | true | 12.383464 | 0.785578 | 78.557782 | 0.694407 | 56.93552 | 0.356495 | 35.649547 | 0.39094 | 18.791946 | 0.455823 | 16.811198 | 0.545628 | 49.514258 | false | true | 2024-08-14 | 2024-08-16 | 0 | abacusai/Dracarys-72B-Instruct |
abacusai_Liberated-Qwen1.5-14B_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Liberated-Qwen1.5-14B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Liberated-Qwen1.5-14B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Liberated-Qwen1.5-14B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Liberated-Qwen1.5-14B | cc0fa5102bfee821bb5e49f082731ccb9d1fedf1 | 19.866148 | other | 21 | 14 | true | false | false | true | 4.077312 | 0.363102 | 36.310212 | 0.4948 | 28.020906 | 0.121601 | 12.160121 | 0.283557 | 4.474273 | 0.417469 | 10.316927 | 0.35123 | 27.91445 | false | true | 2024-03-05 | 2024-09-05 | 0 | abacusai/Liberated-Qwen1.5-14B |
abacusai_Llama-3-Smaug-8B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Llama-3-Smaug-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Llama-3-Smaug-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Llama-3-Smaug-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Llama-3-Smaug-8B | fe54a7d42160d3d8fcc3289c8c411fd9dd5e8357 | 18.967057 | llama2 | 87 | 8 | true | false | false | true | 0.910218 | 0.486675 | 48.667535 | 0.493071 | 27.880374 | 0.079305 | 7.930514 | 0.248322 | 0 | 0.36225 | 5.047917 | 0.318484 | 24.276005 | false | true | 2024-04-19 | 2024-07-02 | 0 | abacusai/Llama-3-Smaug-8B |
abacusai_Smaug-34B-v0.1_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Smaug-34B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Smaug-34B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Smaug-34B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Smaug-34B-v0.1 | 34d54c65a0247d5eb694973106c816d9c0ad3fc2 | 23.757347 | apache-2.0 | 60 | 34 | true | false | false | true | 11.785941 | 0.501563 | 50.156252 | 0.535779 | 34.261661 | 0 | 0 | 0.329698 | 10.626398 | 0.397875 | 8.134375 | 0.454289 | 39.365396 | false | true | 2024-01-25 | 2024-06-12 | 1 | jondurbin/bagel-34b-v0.2 |
abacusai_Smaug-72B-v0.1_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Smaug-72B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Smaug-72B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Smaug-72B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Smaug-72B-v0.1 | a1d657156f82c24b670158406378648233487011 | 29.573654 | other | 467 | 72 | true | false | false | false | 38.824997 | 0.5167 | 51.670013 | 0.599563 | 43.1251 | 0.181269 | 18.126888 | 0.323826 | 9.8434 | 0.447323 | 14.415365 | 0.46235 | 40.261155 | false | true | 2024-02-02 | 2024-06-12 | 1 | moreh/MoMo-72B-lora-1.8.7-DPO |
abacusai_Smaug-Llama-3-70B-Instruct-32K_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Smaug-Llama-3-70B-Instruct-32K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Smaug-Llama-3-70B-Instruct-32K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Smaug-Llama-3-70B-Instruct-32K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Smaug-Llama-3-70B-Instruct-32K | 33840982dc253968f32ef3a534ee0e025eb97482 | 35.022193 | llama3 | 21 | 70 | true | false | false | true | 13.303413 | 0.776111 | 77.611072 | 0.649311 | 49.07037 | 0.230363 | 23.036254 | 0.296141 | 6.152125 | 0.420792 | 12.432292 | 0.476479 | 41.831043 | false | true | 2024-06-11 | 2024-08-06 | 0 | abacusai/Smaug-Llama-3-70B-Instruct-32K |
abacusai_Smaug-Mixtral-v0.1_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Smaug-Mixtral-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Smaug-Mixtral-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Smaug-Mixtral-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Smaug-Mixtral-v0.1 | 98fdc8315906b0a8b9e7f24bad89914869fcfc20 | 22.235369 | apache-2.0 | 12 | 46 | true | true | false | true | 3.941416 | 0.555443 | 55.544289 | 0.516225 | 31.919261 | 0 | 0 | 0.301174 | 6.823266 | 0.429813 | 12.993229 | 0.335189 | 26.132166 | false | true | 2024-02-18 | 2024-08-30 | 0 | abacusai/Smaug-Mixtral-v0.1 |
abacusai_Smaug-Qwen2-72B-Instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/Smaug-Qwen2-72B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/Smaug-Qwen2-72B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__Smaug-Qwen2-72B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/Smaug-Qwen2-72B-Instruct | af015925946d0c60ef69f512c3b35f421cf8063d | 41.432226 | other | 9 | 72 | true | false | false | true | 13.257335 | 0.78253 | 78.253035 | 0.690979 | 56.266172 | 0.374622 | 37.462236 | 0.361577 | 14.876957 | 0.440073 | 15.175781 | 0.519033 | 46.559176 | false | true | 2024-06-26 | 2024-07-29 | 0 | abacusai/Smaug-Qwen2-72B-Instruct |
abacusai_bigstral-12b-32k_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/bigstral-12b-32k" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/bigstral-12b-32k</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__bigstral-12b-32k-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/bigstral-12b-32k | b78a5385ec1b04d6c97f25e9ba1dff18dc98305f | 18.072201 | apache-2.0 | 43 | 12 | true | false | false | false | 0.965279 | 0.419381 | 41.938058 | 0.470012 | 25.556902 | 0.011329 | 1.132931 | 0.292785 | 5.704698 | 0.455979 | 15.864063 | 0.264129 | 18.236554 | true | true | 2024-03-06 | 2024-09-04 | 1 | abacusai/bigstral-12b-32k (Merge) |
abacusai_bigyi-15b_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abacusai/bigyi-15b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abacusai/bigyi-15b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abacusai__bigyi-15b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abacusai/bigyi-15b | b878c15531f7aaf6cf287530f1117b1308b96dc4 | 12.976296 | other | 11 | 15 | true | false | false | false | 1.871122 | 0.209403 | 20.940327 | 0.43453 | 19.940223 | 0.024924 | 2.492447 | 0.309564 | 7.941834 | 0.353781 | 4.289323 | 0.300283 | 22.25362 | true | true | 2024-03-06 | 2024-09-17 | 1 | abacusai/bigyi-15b (Merge) |
abhishek_autotrain-0tmgq-5tpbg_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abhishek/autotrain-0tmgq-5tpbg" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abhishek/autotrain-0tmgq-5tpbg</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abhishek__autotrain-0tmgq-5tpbg-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abhishek/autotrain-0tmgq-5tpbg | a75e1fda984e009613dca3b7846c579a37ab0673 | 4.856619 | other | 0 | 0 | true | false | false | true | 0.351828 | 0.195715 | 19.571515 | 0.313451 | 4.268752 | 0 | 0 | 0.251678 | 0.223714 | 0.365042 | 3.396875 | 0.11511 | 1.678856 | false | false | 2024-11-19 | 2024-12-03 | 2 | HuggingFaceTB/SmolLM2-135M-Instruct (Merge) |
abhishek_autotrain-0tmgq-5tpbg_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abhishek/autotrain-0tmgq-5tpbg" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abhishek/autotrain-0tmgq-5tpbg</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abhishek__autotrain-0tmgq-5tpbg-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abhishek/autotrain-0tmgq-5tpbg | a75e1fda984e009613dca3b7846c579a37ab0673 | 4.837547 | other | 0 | 0 | true | false | false | true | 0.336804 | 0.195165 | 19.516549 | 0.312733 | 4.419023 | 0 | 0 | 0.259228 | 1.230425 | 0.358375 | 2.263542 | 0.114362 | 1.595745 | false | false | 2024-11-19 | 2024-12-04 | 2 | HuggingFaceTB/SmolLM2-135M-Instruct (Merge) |
abhishek_autotrain-llama3-70b-orpo-v1_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abhishek/autotrain-llama3-70b-orpo-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abhishek/autotrain-llama3-70b-orpo-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abhishek__autotrain-llama3-70b-orpo-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abhishek/autotrain-llama3-70b-orpo-v1 | 053236c6846cc561c1503ba05e2b28c94855a432 | 14.712672 | other | 4 | 70 | true | false | false | true | 10.761028 | 0.423302 | 42.330239 | 0.599799 | 41.565362 | 0.004532 | 0.453172 | 0.244128 | 0 | 0.357906 | 2.571615 | 0.112201 | 1.355644 | false | false | 2024-05-02 | 2024-08-30 | 0 | abhishek/autotrain-llama3-70b-orpo-v1 |
abhishek_autotrain-llama3-70b-orpo-v2_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abhishek/autotrain-llama3-70b-orpo-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abhishek/autotrain-llama3-70b-orpo-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abhishek__autotrain-llama3-70b-orpo-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abhishek/autotrain-llama3-70b-orpo-v2 | a2c16a8a7fa48792eb8a1f0c50e13309c2021a63 | 28.804373 | other | 3 | 70 | true | false | false | true | 12.527047 | 0.540606 | 54.060559 | 0.589947 | 39.882199 | 0.206949 | 20.694864 | 0.293624 | 5.816555 | 0.411333 | 9.95 | 0.481799 | 42.42206 | false | false | 2024-05-04 | 2024-08-21 | 0 | abhishek/autotrain-llama3-70b-orpo-v2 |
abhishek_autotrain-llama3-orpo-v2_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/abhishek/autotrain-llama3-orpo-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abhishek/autotrain-llama3-orpo-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abhishek__autotrain-llama3-orpo-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abhishek/autotrain-llama3-orpo-v2 | 1655d0683696a5de2eb9a59c339ee469297beb9c | 12.263693 | other | 3 | 8 | true | false | false | true | 0.90594 | 0.437166 | 43.716561 | 0.315938 | 4.380134 | 0.046073 | 4.607251 | 0.266779 | 2.237136 | 0.37924 | 5.104948 | 0.221825 | 13.536126 | false | false | 2024-04-22 | 2024-06-26 | 0 | abhishek/autotrain-llama3-orpo-v2 |
abhishek_autotrain-vr4a1-e5mms_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Adapter | ? | <a target="_blank" href="https://huggingface.co/abhishek/autotrain-vr4a1-e5mms" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">abhishek/autotrain-vr4a1-e5mms</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/abhishek__autotrain-vr4a1-e5mms-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | abhishek/autotrain-vr4a1-e5mms | 5206a32e0bd3067aef1ce90f5528ade7d866253f | 18.609616 | other | 0 | 16 | true | false | false | false | 1.872878 | 0.214225 | 21.422492 | 0.500062 | 28.456617 | 0.138218 | 13.821752 | 0.319631 | 9.284116 | 0.389125 | 9.040625 | 0.366689 | 29.632092 | false | false | 2024-09-05 | 2024-09-06 | 2 | meta-llama/Meta-Llama-3.1-8B |
adamo1139_Yi-34B-200K-AEZAKMI-v2_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/adamo1139/Yi-34B-200K-AEZAKMI-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">adamo1139/Yi-34B-200K-AEZAKMI-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/adamo1139__Yi-34B-200K-AEZAKMI-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | adamo1139/Yi-34B-200K-AEZAKMI-v2 | 189b42b0dae6352fbe7165255aae851961c8e678 | 23.789585 | apache-2.0 | 12 | 34 | true | false | false | true | 3.021574 | 0.455526 | 45.552578 | 0.538382 | 35.276425 | 0.054381 | 5.438066 | 0.332215 | 10.961969 | 0.388604 | 6.475521 | 0.451297 | 39.032949 | false | false | 2023-12-13 | 2024-06-26 | 0 | adamo1139/Yi-34B-200K-AEZAKMI-v2 |
adriszmar_QAIMath-Qwen2.5-7B-TIES_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/adriszmar/QAIMath-Qwen2.5-7B-TIES" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">adriszmar/QAIMath-Qwen2.5-7B-TIES</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/adriszmar__QAIMath-Qwen2.5-7B-TIES-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | adriszmar/QAIMath-Qwen2.5-7B-TIES | c89bc166dbe2a31c1fceb40ea7acdd96c5620ff5 | 5.469542 | apache-2.0 | 0 | 7 | true | false | false | false | 1.283408 | 0.174632 | 17.46322 | 0.312638 | 5.253691 | 0 | 0 | 0.244966 | 0 | 0.409594 | 9.132552 | 0.10871 | 0.96779 | true | false | 2024-10-27 | 2024-10-27 | 0 | adriszmar/QAIMath-Qwen2.5-7B-TIES |
adriszmar_QAIMath-Qwen2.5-7B-TIES_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/adriszmar/QAIMath-Qwen2.5-7B-TIES" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">adriszmar/QAIMath-Qwen2.5-7B-TIES</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/adriszmar__QAIMath-Qwen2.5-7B-TIES-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | adriszmar/QAIMath-Qwen2.5-7B-TIES | c89bc166dbe2a31c1fceb40ea7acdd96c5620ff5 | 4.963265 | apache-2.0 | 0 | 7 | true | false | false | false | 1.306316 | 0.168537 | 16.853726 | 0.312427 | 5.019151 | 0 | 0 | 0.249161 | 0 | 0.396292 | 7.169792 | 0.106632 | 0.736924 | true | false | 2024-10-27 | 2024-10-27 | 0 | adriszmar/QAIMath-Qwen2.5-7B-TIES |
ahmeda335_13_outOf_32_pruned_layers_llama3.1-8b_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/ahmeda335/13_outOf_32_pruned_layers_llama3.1-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ahmeda335/13_outOf_32_pruned_layers_llama3.1-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ahmeda335__13_outOf_32_pruned_layers_llama3.1-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ahmeda335/13_outOf_32_pruned_layers_llama3.1-8b | 248c420cc0a0bb8fce3a64a998ca0ce89613783c | 4.404259 | apache-2.0 | 0 | 5 | true | false | false | true | 0.496024 | 0.174807 | 17.480729 | 0.288326 | 1.677845 | 0 | 0 | 0.259228 | 1.230425 | 0.380323 | 4.607031 | 0.112866 | 1.429521 | false | false | 2024-10-21 | 2024-12-03 | 1 | ahmeda335/13_outOf_32_pruned_layers_llama3.1-8b (Merge) |
ai21labs_Jamba-v0.1_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | JambaForCausalLM | <a target="_blank" href="https://huggingface.co/ai21labs/Jamba-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ai21labs/Jamba-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ai21labs__Jamba-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ai21labs/Jamba-v0.1 | ce13f3fe99555a2606d1892665bb67649032ff2d | 9.142836 | apache-2.0 | 1,171 | 51 | true | true | false | true | 10.112143 | 0.202559 | 20.255921 | 0.360226 | 10.722059 | 0.011329 | 1.132931 | 0.268456 | 2.46085 | 0.359021 | 3.710937 | 0.249169 | 16.57432 | false | true | 2024-03-28 | 2024-09-16 | 0 | ai21labs/Jamba-v0.1 |
aixonlab_Aether-12b_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/aixonlab/Aether-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">aixonlab/Aether-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/aixonlab__Aether-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | aixonlab/Aether-12b | c55d08a69c74f87c18ab5afb05d46359f389c91a | 17.882297 | apache-2.0 | 1 | 12 | true | false | false | false | 1.866432 | 0.234683 | 23.468286 | 0.51794 | 30.551138 | 0.096677 | 9.667674 | 0.316275 | 8.836689 | 0.382865 | 7.991406 | 0.341007 | 26.77859 | false | false | 2024-09-24 | 2024-10-09 | 1 | Xclbr7/Arcanum-12b |
aixonlab_Grey-12b_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/aixonlab/Grey-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">aixonlab/Grey-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/aixonlab__Grey-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | aixonlab/Grey-12b | 50f56572870c49186c3679f9949a602d2d97c046 | 23.606024 | apache-2.0 | 0 | 12 | true | false | false | false | 1.468694 | 0.396799 | 39.679938 | 0.569896 | 38.746043 | 0.093656 | 9.365559 | 0.300336 | 6.711409 | 0.451635 | 16.254427 | 0.377909 | 30.878768 | false | false | 2024-10-07 | 2024-10-09 | 2 | Xclbr7/Arcanum-12b |
akjindal53244_Llama-3.1-Storm-8B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">akjindal53244/Llama-3.1-Storm-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/akjindal53244__Llama-3.1-Storm-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | akjindal53244/Llama-3.1-Storm-8B | df21b06dcf534b026dd301a44a521d7253c8b94b | 29.36525 | llama3.1 | 168 | 8 | true | false | false | true | 0.794391 | 0.803263 | 80.326312 | 0.519633 | 31.615695 | 0.162387 | 16.238671 | 0.309564 | 7.941834 | 0.402833 | 8.820833 | 0.381233 | 31.248153 | true | false | 2024-08-12 | 2024-10-27 | 0 | akjindal53244/Llama-3.1-Storm-8B |
akjindal53244_Llama-3.1-Storm-8B_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">akjindal53244/Llama-3.1-Storm-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/akjindal53244__Llama-3.1-Storm-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | akjindal53244/Llama-3.1-Storm-8B | df21b06dcf534b026dd301a44a521d7253c8b94b | 29.843219 | llama3.1 | 168 | 8 | true | false | false | true | 1.29582 | 0.805062 | 80.506168 | 0.518867 | 31.494363 | 0.166163 | 16.616314 | 0.326342 | 10.178971 | 0.402802 | 9.116927 | 0.380319 | 31.146572 | true | false | 2024-08-12 | 2024-11-26 | 0 | akjindal53244/Llama-3.1-Storm-8B |
alcholjung_llama3_medical_tuned_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Adapter | ? | <a target="_blank" href="https://huggingface.co/alcholjung/llama3_medical_tuned" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">alcholjung/llama3_medical_tuned</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/alcholjung__llama3_medical_tuned-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | alcholjung/llama3_medical_tuned | 62bd457b6fe961a42a631306577e622c83876cb6 | 11.306071 | 0 | 16 | false | false | false | false | 0.910721 | 0.010566 | 1.056641 | 0.451294 | 23.265089 | 0.002266 | 0.226586 | 0.286074 | 4.809843 | 0.466021 | 16.852604 | 0.294631 | 21.625665 | false | false | 2024-08-14 | 2024-08-14 | 0 | alcholjung/llama3_medical_tuned |
|
allenai_Llama-3.1-Tulu-3-70B_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-70B | c4280450c0cd91a2fb6f41a25c6a1662c6966b01 | 41.198857 | llama3.1 | 43 | 70 | true | false | false | true | 36.59305 | 0.829117 | 82.911674 | 0.616363 | 45.365569 | 0.382175 | 38.217523 | 0.373322 | 16.442953 | 0.494833 | 23.754167 | 0.464511 | 40.501256 | false | true | 2024-11-20 | 2024-11-27 | 1 | allenai/Llama-3.1-Tulu-3-70B (Merge) |
allenai_Llama-3.1-Tulu-3-70B_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-70B | c4280450c0cd91a2fb6f41a25c6a1662c6966b01 | 41.454527 | llama3.1 | 43 | 70 | true | false | false | true | 38.022026 | 0.837934 | 83.793446 | 0.615685 | 45.259481 | 0.382931 | 38.293051 | 0.373322 | 16.442953 | 0.498802 | 24.316927 | 0.465592 | 40.621306 | false | true | 2024-11-20 | 2024-11-27 | 1 | allenai/Llama-3.1-Tulu-3-70B (Merge) |
allenai_Llama-3.1-Tulu-3-70B-DPO_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-70B-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-70B-DPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-70B-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-70B-DPO | 6ea110f39fb660573111892a1381d3be3f826f80 | 41.204777 | llama3.1 | 7 | 70 | true | false | false | true | 36.800749 | 0.828193 | 82.819253 | 0.61462 | 45.047181 | 0.388218 | 38.821752 | 0.375839 | 16.778523 | 0.49226 | 23.399219 | 0.463265 | 40.362736 | false | true | 2024-11-20 | 2024-11-27 | 1 | allenai/Llama-3.1-Tulu-3-70B-DPO (Merge) |
allenai_Llama-3.1-Tulu-3-70B-SFT_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-70B-SFT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-70B-SFT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-70B-SFT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-70B-SFT | f58ab66db3a1c5dd805c6d3420b2b4f5aef30041 | 38.722611 | llama3.1 | 4 | 70 | true | false | false | true | 27.338327 | 0.805062 | 80.506168 | 0.595144 | 42.023984 | 0.324018 | 32.401813 | 0.344799 | 12.639821 | 0.502615 | 24.49349 | 0.462434 | 40.27039 | false | true | 2024-11-18 | 2024-11-27 | 1 | allenai/Llama-3.1-Tulu-3-70B-SFT (Merge) |
allenai_Llama-3.1-Tulu-3-8B_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-8B | 63b75e0dd6eac3725319f869716b9b70c16a6a65 | 26.034998 | llama3.1 | 95 | 8 | true | false | false | true | 0.703774 | 0.826669 | 82.666879 | 0.404983 | 16.671813 | 0.196375 | 19.637462 | 0.298658 | 6.487696 | 0.417469 | 10.45026 | 0.282663 | 20.295878 | false | true | 2024-11-20 | 2024-11-21 | 1 | allenai/Llama-3.1-Tulu-3-8B (Merge) |
allenai_Llama-3.1-Tulu-3-8B_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-8B | 50fef8756a9a4ca2010587d128aebb3a18ec897d | 25.883225 | llama3.1 | 95 | 8 | true | false | false | true | 0.701232 | 0.82547 | 82.546975 | 0.406083 | 16.858052 | 0.188822 | 18.882175 | 0.29698 | 6.263982 | 0.417469 | 10.516927 | 0.282081 | 20.231235 | false | true | 2024-11-20 | 2024-11-28 | 1 | allenai/Llama-3.1-Tulu-3-8B (Merge) |
allenai_Llama-3.1-Tulu-3-8B-DPO_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-8B-DPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-8B-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-8B-DPO | 002347006131d85678ea3865520bc9caad69869a | 25.620576 | llama3.1 | 12 | 8 | true | false | false | true | 0.670338 | 0.802938 | 80.293843 | 0.407943 | 17.426016 | 0.185801 | 18.58006 | 0.293624 | 5.816555 | 0.416135 | 10.516927 | 0.289811 | 21.090056 | false | true | 2024-11-20 | 2024-11-22 | 1 | allenai/Llama-3.1-Tulu-3-8B-DPO (Merge) |
allenai_Llama-3.1-Tulu-3-8B-RM_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForSequenceClassification | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B-RM" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-8B-RM</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-8B-RM-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-8B-RM | 76247c00745747f820f1712949b5b37901d0f9c4 | 4.235057 | llama3.1 | 7 | 8 | true | false | false | true | 0.736899 | 0.167014 | 16.701352 | 0.295004 | 2.64967 | 0 | 0 | 0.256711 | 0.894855 | 0.376417 | 4.252083 | 0.108211 | 0.912382 | false | true | 2024-11-20 | 2024-11-22 | 1 | allenai/Llama-3.1-Tulu-3-8B-RM (Merge) |
allenai_Llama-3.1-Tulu-3-8B-SFT_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B-SFT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/Llama-3.1-Tulu-3-8B-SFT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__Llama-3.1-Tulu-3-8B-SFT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/Llama-3.1-Tulu-3-8B-SFT | 4ddd761e6750e04ea3d468175f78463628bba860 | 22.534 | llama3.1 | 12 | 8 | true | false | false | true | 0.683246 | 0.74034 | 74.034008 | 0.387186 | 13.931208 | 0.114048 | 11.404834 | 0.277685 | 3.691275 | 0.426771 | 12.013021 | 0.281167 | 20.129654 | false | true | 2024-11-18 | 2024-11-22 | 1 | allenai/Llama-3.1-Tulu-3-8B-SFT (Merge) |
allenai_OLMo-1B-hf_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | OlmoForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/OLMo-1B-hf" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/OLMo-1B-hf</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__OLMo-1B-hf-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/OLMo-1B-hf | 8e995430edd24416ccfa98b5b283fa07b0c9f1a9 | 6.470278 | apache-2.0 | 17 | 1 | true | false | false | false | 0.248874 | 0.218197 | 21.819661 | 0.305195 | 3.196546 | 0.007553 | 0.755287 | 0.261745 | 1.565996 | 0.409781 | 9.55599 | 0.117354 | 1.928191 | false | true | 2024-04-12 | 2024-06-12 | 0 | allenai/OLMo-1B-hf |
allenai_OLMo-7B-Instruct-hf_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | OlmoForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/OLMo-7B-Instruct-hf" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/OLMo-7B-Instruct-hf</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/OLMo-7B-Instruct-hf | 2ea947518df93433aa71219f29b36c72ac63be95 | 10.760857 | apache-2.0 | 2 | 7 | true | false | false | true | 1.199951 | 0.347265 | 34.726526 | 0.370647 | 13.159933 | 0.008308 | 0.830816 | 0.270973 | 2.796421 | 0.376479 | 4.326563 | 0.178524 | 8.724882 | false | true | 2024-06-04 | 2024-06-27 | 0 | allenai/OLMo-7B-Instruct-hf |
allenai_OLMo-7B-hf_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | OlmoForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/OLMo-7B-hf" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/OLMo-7B-hf</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__OLMo-7B-hf-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/OLMo-7B-hf | 687d934d36a05417048d0fe7482f24f389fef6aa | 6.776151 | apache-2.0 | 12 | 6 | true | false | false | false | 0.590564 | 0.271927 | 27.192737 | 0.327913 | 5.761987 | 0.006798 | 0.679758 | 0.272651 | 3.020134 | 0.348667 | 2.083333 | 0.117271 | 1.918957 | false | true | 2024-04-12 | 2024-06-27 | 0 | allenai/OLMo-7B-hf |
allenai_OLMoE-1B-7B-0924_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | OlmoeForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/OLMoE-1B-7B-0924" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/OLMoE-1B-7B-0924</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__OLMoE-1B-7B-0924-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/OLMoE-1B-7B-0924 | 4fa3a6e09ed0e41639962f38bfba0fc532b90075 | 7.178464 | apache-2.0 | 109 | 6 | true | true | false | false | 3.076407 | 0.218471 | 21.847143 | 0.339344 | 8.308107 | 0.011329 | 1.132931 | 0.247483 | 0 | 0.348792 | 3.565625 | 0.173953 | 8.216977 | false | true | 2024-07-20 | 2024-09-30 | 0 | allenai/OLMoE-1B-7B-0924 |
allenai_OLMoE-1B-7B-0924-Instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | OlmoeForCausalLM | <a target="_blank" href="https://huggingface.co/allenai/OLMoE-1B-7B-0924-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allenai/OLMoE-1B-7B-0924-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allenai__OLMoE-1B-7B-0924-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allenai/OLMoE-1B-7B-0924-Instruct | 7f1c97f440f06ce36705e4f2b843edb5925f4498 | 13.207221 | apache-2.0 | 85 | 6 | true | true | false | true | 5.487959 | 0.465218 | 46.521784 | 0.390161 | 14.571563 | 0 | 0 | 0.267617 | 2.348993 | 0.384823 | 6.069531 | 0.187583 | 9.731457 | false | true | 2024-08-13 | 2024-09-30 | 2 | allenai/OLMoE-1B-7B-0924 |
allknowingroger_Chocolatine-24B_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Chocolatine-24B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Chocolatine-24B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Chocolatine-24B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Chocolatine-24B | 6245b82885ca4930575dbed2932ec1d32d901c0e | 21.333145 | apache-2.0 | 1 | 24 | true | false | false | false | 6.18496 | 0.195815 | 19.581488 | 0.619126 | 45.78594 | 0 | 0 | 0.325503 | 10.067114 | 0.432323 | 12.940365 | 0.456616 | 39.623966 | true | false | 2024-09-02 | 2024-09-02 | 1 | allknowingroger/Chocolatine-24B (Merge) |
allknowingroger_Gemma2Slerp1-2.6B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Gemma2Slerp1-2.6B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Gemma2Slerp1-2.6B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Gemma2Slerp1-2.6B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Gemma2Slerp1-2.6B | 2d0e85a03c55abd22963c5c3a44f180bfecebf7b | 20.594692 | 0 | 2 | false | false | false | false | 1.194263 | 0.535435 | 53.543487 | 0.434309 | 19.770255 | 0.101964 | 10.196375 | 0.283557 | 4.474273 | 0.456167 | 16.820833 | 0.268866 | 18.762928 | false | false | 2024-12-04 | 2024-12-06 | 1 | allknowingroger/Gemma2Slerp1-2.6B (Merge) |
|
allknowingroger_Gemma2Slerp1-27B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Gemma2Slerp1-27B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Gemma2Slerp1-27B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Gemma2Slerp1-27B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Gemma2Slerp1-27B | 4a5c5092f40cc161bb18ca2b9e30a653c768e062 | 36.356582 | apache-2.0 | 0 | 27 | true | false | false | false | 4.089968 | 0.718633 | 71.863323 | 0.63989 | 48.377666 | 0.249245 | 24.924471 | 0.364094 | 15.212528 | 0.476719 | 19.35651 | 0.445645 | 38.404994 | true | false | 2024-11-30 | 2024-12-06 | 1 | allknowingroger/Gemma2Slerp1-27B (Merge) |
allknowingroger_Gemma2Slerp2-2.6B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Gemma2Slerp2-2.6B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Gemma2Slerp2-2.6B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Gemma2Slerp2-2.6B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Gemma2Slerp2-2.6B | 12ca2fdb5dd866fbdc624057a176ad3d1f8c2293 | 21.28146 | 0 | 2 | false | false | false | false | 1.200011 | 0.574727 | 57.472728 | 0.430765 | 19.719839 | 0.089879 | 8.987915 | 0.305369 | 7.38255 | 0.446771 | 15.279688 | 0.269614 | 18.84604 | false | false | 2024-12-04 | 2024-12-06 | 1 | allknowingroger/Gemma2Slerp2-2.6B (Merge) |
|
allknowingroger_Gemma2Slerp2-27B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Gemma2Slerp2-27B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Gemma2Slerp2-27B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Gemma2Slerp2-27B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Gemma2Slerp2-27B | 21043f6eaf40680675461825fbdfc964f4a3c4a0 | 37.340059 | apache-2.0 | 0 | 27 | true | false | false | false | 4.411521 | 0.754553 | 75.455347 | 0.655727 | 51.090234 | 0.243202 | 24.320242 | 0.369966 | 15.995526 | 0.462083 | 16.927083 | 0.462267 | 40.251921 | true | false | 2024-11-30 | 2024-12-06 | 1 | allknowingroger/Gemma2Slerp2-27B (Merge) |
allknowingroger_Gemma2Slerp3-27B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Gemma2Slerp3-27B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Gemma2Slerp3-27B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Gemma2Slerp3-27B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Gemma2Slerp3-27B | cddd53f3b29a361be2350b76770a60b3fcc78059 | 37.166401 | apache-2.0 | 0 | 27 | true | false | false | false | 4.346639 | 0.742638 | 74.263842 | 0.649964 | 49.951521 | 0.252266 | 25.226586 | 0.354866 | 13.982103 | 0.474021 | 19.119271 | 0.464096 | 40.455083 | true | false | 2024-12-01 | 2024-12-06 | 1 | allknowingroger/Gemma2Slerp3-27B (Merge) |
allknowingroger_Gemma2Slerp4-27B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Gemma2Slerp4-27B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Gemma2Slerp4-27B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Gemma2Slerp4-27B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Gemma2Slerp4-27B | 5c89bb96e60f0297f5bf27fc10713a4dcdd54285 | 36.650032 | apache-2.0 | 0 | 27 | true | false | false | false | 4.406683 | 0.749658 | 74.965758 | 0.652958 | 50.773762 | 0.228852 | 22.885196 | 0.366611 | 15.548098 | 0.45024 | 15.179948 | 0.464927 | 40.547429 | true | false | 2024-12-01 | 2024-12-06 | 1 | allknowingroger/Gemma2Slerp4-27B (Merge) |
allknowingroger_GemmaSlerp-9B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/GemmaSlerp-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/GemmaSlerp-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__GemmaSlerp-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/GemmaSlerp-9B | 4f54819ae9c0af1f3e508f0afc88a7a734f9632d | 30.857317 | apache-2.0 | 0 | 9 | true | false | false | false | 1.60767 | 0.70432 | 70.432009 | 0.592058 | 41.556032 | 0.076284 | 7.628399 | 0.34396 | 12.527964 | 0.467323 | 17.882031 | 0.416057 | 35.117465 | true | false | 2024-10-27 | 2024-11-22 | 1 | allknowingroger/GemmaSlerp-9B (Merge) |
allknowingroger_GemmaSlerp2-9B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/GemmaSlerp2-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/GemmaSlerp2-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__GemmaSlerp2-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/GemmaSlerp2-9B | e93fb8d7fad0007e463e44365a5a82d0d6facd61 | 33.502291 | apache-2.0 | 2 | 9 | true | false | false | false | 1.785003 | 0.7281 | 72.810033 | 0.598271 | 42.541033 | 0.165408 | 16.540785 | 0.352349 | 13.646532 | 0.476719 | 19.489844 | 0.42387 | 35.98552 | true | false | 2024-10-29 | 2024-11-22 | 1 | allknowingroger/GemmaSlerp2-9B (Merge) |
allknowingroger_GemmaSlerp4-10B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/GemmaSlerp4-10B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/GemmaSlerp4-10B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__GemmaSlerp4-10B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/GemmaSlerp4-10B | e30d14d05730a83926263a7b0e4b1e002b6cd65a | 33.232197 | apache-2.0 | 2 | 10 | true | false | false | false | 1.918679 | 0.732622 | 73.262167 | 0.602786 | 43.328658 | 0.174471 | 17.44713 | 0.353188 | 13.758389 | 0.45399 | 15.482031 | 0.425033 | 36.114805 | true | false | 2024-10-30 | 2024-11-22 | 1 | allknowingroger/GemmaSlerp4-10B (Merge) |
allknowingroger_GemmaSlerp5-10B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/GemmaSlerp5-10B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/GemmaSlerp5-10B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__GemmaSlerp5-10B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/GemmaSlerp5-10B | 7e94afcde7cc1ae88105521a831abefe8126b0d1 | 34.042524 | apache-2.0 | 2 | 10 | true | false | false | false | 2.322861 | 0.735344 | 73.534444 | 0.605448 | 43.538464 | 0.197885 | 19.78852 | 0.352349 | 13.646532 | 0.460781 | 16.764323 | 0.432846 | 36.982861 | true | false | 2024-10-30 | 2024-11-22 | 1 | allknowingroger/GemmaSlerp5-10B (Merge) |
allknowingroger_GemmaStock1-27B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/GemmaStock1-27B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/GemmaStock1-27B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__GemmaStock1-27B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/GemmaStock1-27B | 8563301fe323c4d1060ae6f56d5737ad62a63fef | 37.43813 | apache-2.0 | 0 | 27 | true | false | false | false | 4.152741 | 0.750906 | 75.090648 | 0.656561 | 50.990136 | 0.259063 | 25.906344 | 0.364094 | 15.212528 | 0.452687 | 15.985937 | 0.472989 | 41.443189 | true | false | 2024-12-03 | 2024-12-06 | 1 | allknowingroger/GemmaStock1-27B (Merge) |
allknowingroger_HomerSlerp1-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/HomerSlerp1-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/HomerSlerp1-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__HomerSlerp1-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/HomerSlerp1-7B | 42e3df3d9a25d8ff0d470582395f165b2ddb83d8 | 28.622212 | apache-2.0 | 1 | 7 | true | false | false | false | 0.683282 | 0.462121 | 46.212051 | 0.551818 | 36.259863 | 0.280211 | 28.021148 | 0.317953 | 9.060403 | 0.435854 | 13.248438 | 0.450382 | 38.931368 | true | false | 2024-11-20 | 2024-11-22 | 1 | allknowingroger/HomerSlerp1-7B (Merge) |
allknowingroger_HomerSlerp2-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/HomerSlerp2-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/HomerSlerp2-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__HomerSlerp2-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/HomerSlerp2-7B | 210acef73da0488ea270332f5831b609298a98f0 | 28.760078 | apache-2.0 | 1 | 7 | true | false | false | false | 0.624987 | 0.448682 | 44.868172 | 0.564894 | 37.9603 | 0.285498 | 28.549849 | 0.319631 | 9.284116 | 0.435573 | 12.846615 | 0.451463 | 39.051418 | true | false | 2024-11-20 | 2024-11-22 | 1 | allknowingroger/HomerSlerp2-7B (Merge) |
allknowingroger_HomerSlerp3-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/HomerSlerp3-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/HomerSlerp3-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__HomerSlerp3-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/HomerSlerp3-7B | 4f41686caa5bc39e3b0f075360974057486ece95 | 28.601186 | apache-2.0 | 1 | 7 | true | false | false | false | 0.617234 | 0.436267 | 43.626688 | 0.559806 | 37.290018 | 0.280967 | 28.096677 | 0.317114 | 8.948546 | 0.446177 | 14.372135 | 0.453457 | 39.27305 | true | false | 2024-11-21 | 2024-11-22 | 1 | allknowingroger/HomerSlerp3-7B (Merge) |
allknowingroger_HomerSlerp4-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/HomerSlerp4-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/HomerSlerp4-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__HomerSlerp4-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/HomerSlerp4-7B | f2ce1f2afa3c645e26ca61ee30f24736873bafa1 | 28.616144 | apache-2.0 | 0 | 7 | true | false | false | false | 0.67907 | 0.437416 | 43.741606 | 0.557077 | 36.786834 | 0.295317 | 29.531722 | 0.319631 | 9.284116 | 0.440844 | 13.772135 | 0.447224 | 38.580452 | true | false | 2024-11-21 | 2024-11-22 | 1 | allknowingroger/HomerSlerp4-7B (Merge) |
allknowingroger_LimyQstar-7B-slerp_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/LimyQstar-7B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/LimyQstar-7B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__LimyQstar-7B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/LimyQstar-7B-slerp | 6dc557c7bfd6a6f9bc8190bc8a31c3b732deca40 | 18.672525 | apache-2.0 | 0 | 7 | true | false | false | false | 0.630324 | 0.349114 | 34.911369 | 0.502356 | 30.194567 | 0.068731 | 6.873112 | 0.298658 | 6.487696 | 0.414646 | 10.197396 | 0.310339 | 23.371011 | true | false | 2024-03-23 | 2024-06-26 | 1 | allknowingroger/LimyQstar-7B-slerp (Merge) |
allknowingroger_Llama3.1-60B_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Llama3.1-60B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Llama3.1-60B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Llama3.1-60B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Llama3.1-60B | 5fb1ddcce0bddc60949a9d0c2fc9f8326be5bc4e | 9.951594 | 0 | 61 | false | false | false | false | 13.491859 | 0.181452 | 18.145188 | 0.324176 | 7.784283 | 0 | 0 | 0.294463 | 5.928412 | 0.359583 | 2.18125 | 0.331034 | 25.670434 | false | false | 2024-10-08 | 0 | Removed |
||
allknowingroger_Marco-01-slerp1-7B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Marco-01-slerp1-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Marco-01-slerp1-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Marco-01-slerp1-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Marco-01-slerp1-7B | 12070d5f5bbd891024cb02c363759430ffd3dfba | 29.485317 | apache-2.0 | 0 | 7 | true | false | false | false | 0.637324 | 0.468116 | 46.811571 | 0.554094 | 36.231847 | 0.31571 | 31.570997 | 0.317114 | 8.948546 | 0.445188 | 14.648438 | 0.448305 | 38.700502 | true | false | 2024-11-22 | 2024-11-22 | 1 | allknowingroger/Marco-01-slerp1-7B (Merge) |
allknowingroger_Meme-7B-slerp_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Meme-7B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Meme-7B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Meme-7B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Meme-7B-slerp | 7836c0f4fce70286382e61003e9a05d7559365d9 | 19.326433 | apache-2.0 | 0 | 7 | true | false | false | false | 0.48245 | 0.516375 | 51.637544 | 0.466094 | 24.529486 | 0.046828 | 4.682779 | 0.286074 | 4.809843 | 0.422302 | 10.18776 | 0.281001 | 20.111185 | true | false | 2024-05-22 | 2024-06-26 | 1 | allknowingroger/Meme-7B-slerp (Merge) |
allknowingroger_Ministral-8B-slerp_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Ministral-8B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Ministral-8B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Ministral-8B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Ministral-8B-slerp | 51c40046c0f9fead83485ae83b6c0d03f4ae47f2 | 14.838025 | 0 | 7 | false | false | false | false | 1.125098 | 0.19609 | 19.608971 | 0.468602 | 25.195565 | 0 | 0 | 0.312081 | 8.277405 | 0.428531 | 12.39974 | 0.311918 | 23.546469 | false | false | 2024-10-18 | 2024-10-21 | 1 | allknowingroger/Ministral-8B-slerp (Merge) |
|
allknowingroger_MistralPhi3-11B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/MistralPhi3-11B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/MistralPhi3-11B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__MistralPhi3-11B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/MistralPhi3-11B | 3afeaf24c6306c4752c320c4fd32fa2e7694e12e | 21.627095 | apache-2.0 | 0 | 11 | true | false | false | false | 0.707038 | 0.194291 | 19.429115 | 0.623431 | 46.164629 | 0 | 0 | 0.332215 | 10.961969 | 0.426677 | 12.234635 | 0.46875 | 40.972222 | true | false | 2024-08-26 | 2024-09-02 | 1 | allknowingroger/MistralPhi3-11B (Merge) |
allknowingroger_Mistralmash1-7B-s_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Mistralmash1-7B-s" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Mistralmash1-7B-s</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Mistralmash1-7B-s-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Mistralmash1-7B-s | 730b7b2867deef63961f002b6e1e70e7d416c599 | 20.901278 | apache-2.0 | 0 | 7 | true | false | false | false | 0.661188 | 0.3961 | 39.610013 | 0.527749 | 33.448554 | 0.09139 | 9.138973 | 0.294463 | 5.928412 | 0.426708 | 11.805208 | 0.329289 | 25.476507 | true | false | 2024-08-27 | 2024-09-02 | 1 | allknowingroger/Mistralmash1-7B-s (Merge) |
allknowingroger_Mistralmash2-7B-s_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/Mistralmash2-7B-s" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/Mistralmash2-7B-s</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__Mistralmash2-7B-s-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/Mistralmash2-7B-s | 3b2aafa0f931f3d3103fbc96a6da4ac36f376d78 | 21.402269 | apache-2.0 | 0 | 7 | true | false | false | false | 0.665537 | 0.410188 | 41.01883 | 0.530486 | 33.298364 | 0.08006 | 8.006042 | 0.297819 | 6.375839 | 0.43725 | 13.65625 | 0.334525 | 26.058289 | true | false | 2024-08-27 | 2024-09-02 | 1 | allknowingroger/Mistralmash2-7B-s (Merge) |
allknowingroger_MixTAO-19B-pass_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/MixTAO-19B-pass" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/MixTAO-19B-pass</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__MixTAO-19B-pass-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/MixTAO-19B-pass | a41369cfcfbada9d5387051ba616bf1432b31d31 | 20.615004 | apache-2.0 | 1 | 19 | true | false | false | false | 1.255132 | 0.381437 | 38.143681 | 0.512825 | 31.577918 | 0.060423 | 6.042296 | 0.284396 | 4.58613 | 0.478271 | 19.950521 | 0.310505 | 23.38948 | true | false | 2024-06-02 | 2024-06-26 | 1 | allknowingroger/MixTAO-19B-pass (Merge) |
allknowingroger_MixTaoTruthful-13B-slerp_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/MixTaoTruthful-13B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/MixTaoTruthful-13B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__MixTaoTruthful-13B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/MixTaoTruthful-13B-slerp | 3324d37e138c6bf0d6891e54b6dd839c8d2f35ec | 20.265564 | apache-2.0 | 0 | 12 | true | false | false | false | 0.808109 | 0.413885 | 41.388516 | 0.520734 | 32.706362 | 0.067221 | 6.722054 | 0.284396 | 4.58613 | 0.42925 | 12.85625 | 0.310007 | 23.334072 | true | false | 2024-05-25 | 2024-06-26 | 1 | allknowingroger/MixTaoTruthful-13B-slerp (Merge) |
allknowingroger_MultiCalm-7B-slerp_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/MultiCalm-7B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/MultiCalm-7B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__MultiCalm-7B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/MultiCalm-7B-slerp | 1c23540e907fab4dfe0ef66edd0003e764bfe568 | 19.459701 | apache-2.0 | 0 | 7 | true | false | false | false | 0.616573 | 0.392653 | 39.265261 | 0.512189 | 31.466483 | 0.061178 | 6.117825 | 0.282718 | 4.362416 | 0.431948 | 12.960156 | 0.303275 | 22.586067 | true | false | 2024-05-19 | 2024-06-26 | 1 | allknowingroger/MultiCalm-7B-slerp (Merge) |
allknowingroger_MultiMash-12B-slerp_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/MultiMash-12B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/MultiMash-12B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__MultiMash-12B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/MultiMash-12B-slerp | 91a6d0fe6b9271000ca713ee9ab414c782ba4c50 | 20.192492 | apache-2.0 | 0 | 12 | true | false | false | false | 0.84198 | 0.397449 | 39.744877 | 0.514183 | 31.925677 | 0.081571 | 8.1571 | 0.276846 | 3.579418 | 0.443792 | 14.773958 | 0.306765 | 22.973921 | true | false | 2024-05-20 | 2024-06-26 | 1 | allknowingroger/MultiMash-12B-slerp (Merge) |
allknowingroger_MultiMash10-13B-slerp_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/allknowingroger/MultiMash10-13B-slerp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">allknowingroger/MultiMash10-13B-slerp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/allknowingroger__MultiMash10-13B-slerp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | allknowingroger/MultiMash10-13B-slerp | 6def2fd1a11d4c380a19b7a3bdf263a6b80cd8f3 | 20.376084 | apache-2.0 | 0 | 12 | true | false | false | false | 0.879359 | 0.416283 | 41.628324 | 0.518634 | 32.452502 | 0.068731 | 6.873112 | 0.286074 | 4.809843 | 0.431792 | 12.973958 | 0.311669 | 23.518765 | true | false | 2024-05-27 | 2024-06-26 | 1 | allknowingroger/MultiMash10-13B-slerp (Merge) |
Subsets and Splits