eval_name
stringlengths 12
111
| Precision
stringclasses 3
values | Type
stringclasses 6
values | T
stringclasses 6
values | Weight type
stringclasses 2
values | Architecture
stringclasses 52
values | Model
stringlengths 355
689
| fullname
stringlengths 4
102
| Model sha
stringlengths 0
40
| Average ⬆️
float64 1.03
52
| Hub License
stringclasses 26
values | Hub ❤️
int64 0
5.9k
| #Params (B)
int64 -1
140
| Available on the hub
bool 2
classes | MoE
bool 2
classes | Flagged
bool 2
classes | Chat Template
bool 2
classes | CO₂ cost (kg)
float64 0.03
107
| IFEval Raw
float64 0
0.9
| IFEval
float64 0
90
| BBH Raw
float64 0.27
0.75
| BBH
float64 0.81
63.5
| MATH Lvl 5 Raw
float64 0
0.51
| MATH Lvl 5
float64 0
50.7
| GPQA Raw
float64 0.22
0.44
| GPQA
float64 0
24.9
| MUSR Raw
float64 0.29
0.6
| MUSR
float64 0
38.5
| MMLU-PRO Raw
float64 0.1
0.73
| MMLU-PRO
float64 0
70
| Merged
bool 2
classes | Official Providers
bool 2
classes | Upload To Hub Date
stringclasses 424
values | Submission Date
stringclasses 169
values | Generation
int64 0
10
| Base Model
stringlengths 4
102
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
meta-llama_Meta-Llama-3.1-8B-Instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">meta-llama/Meta-Llama-3.1-8B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/meta-llama__Meta-Llama-3.1-8B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | meta-llama/Meta-Llama-3.1-8B-Instruct | df34336b42332c6d360959e259cd6271c6a09fd4 | 28.204458 | llama3.1 | 3,233 | 8 | true | false | false | true | 2.487012 | 0.785578 | 78.557782 | 0.507327 | 29.892756 | 0.193353 | 19.335347 | 0.267617 | 2.348993 | 0.38699 | 8.407031 | 0.376164 | 30.68484 | false | true | 2024-07-18 | 2024-08-15 | 1 | meta-llama/Meta-Llama-3.1-8B |
microsoft_DialoGPT-medium_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | GPT2LMHeadModel | <a target="_blank" href="https://huggingface.co/microsoft/DialoGPT-medium" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/DialoGPT-medium</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__DialoGPT-medium-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/DialoGPT-medium | 7b40bb0f92c45fefa957d088000d8648e5c7fa33 | 5.251434 | mit | 331 | 0 | true | false | false | true | 0.129464 | 0.147904 | 14.790423 | 0.301416 | 2.556856 | 0 | 0 | 0.254195 | 0.559284 | 0.428667 | 12.283333 | 0.111868 | 1.318706 | false | true | 2022-03-02 | 2024-06-13 | 0 | microsoft/DialoGPT-medium |
microsoft_Orca-2-13b_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Orca-2-13b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Orca-2-13b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Orca-2-13b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Orca-2-13b | 2539ff53e6baa4cc603774ad5a2d646f4041ea4e | 18.149404 | other | 665 | 13 | true | false | false | false | 1.008582 | 0.312793 | 31.279339 | 0.488449 | 27.308019 | 0.010574 | 1.057402 | 0.280201 | 4.026846 | 0.512969 | 25.78776 | 0.274934 | 19.437057 | false | true | 2023-11-14 | 2024-06-12 | 0 | microsoft/Orca-2-13b |
microsoft_Orca-2-7b_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Orca-2-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Orca-2-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Orca-2-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Orca-2-7b | 60e31e6bdcf582ad103b807cb74b73ee1d2c4b17 | 14.216008 | other | 216 | 7 | true | false | false | false | 1.209312 | 0.218346 | 21.834621 | 0.445213 | 22.429468 | 0.008308 | 0.830816 | 0.260906 | 1.454139 | 0.502615 | 24.09349 | 0.231882 | 14.653517 | false | true | 2023-11-14 | 2024-06-12 | 0 | microsoft/Orca-2-7b |
microsoft_Phi-3-medium-128k-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-medium-128k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-medium-128k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-medium-128k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-medium-128k-instruct | fa7d2aa4f5ea69b2e36b20d050cdae79c9bfbb3f | 31.711653 | mit | 370 | 13 | true | false | false | true | 1.947559 | 0.604003 | 60.400293 | 0.638232 | 48.460451 | 0.172961 | 17.296073 | 0.336409 | 11.521253 | 0.412948 | 11.351823 | 0.47116 | 41.240027 | false | true | 2024-05-07 | 2024-08-21 | 0 | microsoft/Phi-3-medium-128k-instruct |
microsoft_Phi-3-medium-4k-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-medium-4k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-medium-4k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-medium-4k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-medium-4k-instruct | d194e4e74ffad5a5e193e26af25bcfc80c7f1ffc | 32.89625 | mit | 211 | 13 | true | false | false | true | 1.455263 | 0.642271 | 64.22714 | 0.641246 | 49.38061 | 0.183535 | 18.353474 | 0.336409 | 11.521253 | 0.42575 | 13.052083 | 0.467586 | 40.842937 | false | true | 2024-05-07 | 2024-06-12 | 0 | microsoft/Phi-3-medium-4k-instruct |
microsoft_Phi-3-mini-128k-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-mini-128k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-mini-128k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-mini-128k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-mini-128k-instruct | 5be6479b4bc06a081e8f4c6ece294241ccd32dec | 25.626287 | mit | 1,608 | 3 | true | false | false | true | 24.222252 | 0.597633 | 59.763317 | 0.557453 | 37.099767 | 0.097432 | 9.743202 | 0.317953 | 9.060403 | 0.393688 | 7.710938 | 0.373421 | 30.380098 | false | true | 2024-04-22 | 2024-08-21 | 0 | microsoft/Phi-3-mini-128k-instruct |
microsoft_Phi-3-mini-4k-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-mini-4k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-mini-4k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-mini-4k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-mini-4k-instruct | ff07dc01615f8113924aed013115ab2abd32115b | 25.967733 | mit | 1,086 | 3 | true | false | false | true | 0.804075 | 0.561288 | 56.128849 | 0.567597 | 39.269335 | 0.116314 | 11.63142 | 0.319631 | 9.284116 | 0.395021 | 7.644271 | 0.386636 | 31.848404 | false | true | 2024-04-22 | 2024-06-12 | 0 | microsoft/Phi-3-mini-4k-instruct |
microsoft_Phi-3-mini-4k-instruct_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-mini-4k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-mini-4k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-mini-4k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-mini-4k-instruct | c1358f8a35e6d2af81890deffbbfa575b978c62f | 27.411117 | mit | 1,086 | 3 | true | false | false | true | 0.786699 | 0.547675 | 54.767461 | 0.549072 | 36.559855 | 0.154834 | 15.483384 | 0.332215 | 10.961969 | 0.428417 | 13.11875 | 0.402178 | 33.575281 | false | true | 2024-04-22 | 2024-07-02 | 0 | microsoft/Phi-3-mini-4k-instruct |
microsoft_Phi-3-small-128k-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3SmallForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-small-128k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-small-128k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-small-128k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-small-128k-instruct | f80aaa30bfc64c2b8ab214b541d9050e97163bc4 | 28.590992 | mit | 172 | 7 | true | false | false | true | 2.508468 | 0.636826 | 63.682584 | 0.620218 | 45.63407 | 0 | 0 | 0.317114 | 8.948546 | 0.437844 | 14.497135 | 0.449053 | 38.783614 | false | true | 2024-05-07 | 2024-06-13 | 0 | microsoft/Phi-3-small-128k-instruct |
microsoft_Phi-3-small-8k-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3SmallForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3-small-8k-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3-small-8k-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3-small-8k-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3-small-8k-instruct | 1535ae26fb4faada95c6950e8bc6e867cdad6b00 | 29.670922 | mit | 159 | 7 | true | false | false | true | 1.025454 | 0.649665 | 64.966511 | 0.620836 | 46.20557 | 0.02843 | 2.843014 | 0.312081 | 8.277405 | 0.455792 | 16.773958 | 0.450632 | 38.959072 | false | true | 2024-05-07 | 2024-06-13 | 0 | microsoft/Phi-3-small-8k-instruct |
microsoft_Phi-3.5-MoE-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3.5-MoE-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3.5-MoE-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3.5-MoE-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3.5-MoE-instruct | 482a9ba0eb0e1fa1671e3560e009d7cec2e5147c | 35.456508 | mit | 531 | 42 | true | true | false | true | 4.632279 | 0.692455 | 69.245491 | 0.640763 | 48.774646 | 0.226586 | 22.65861 | 0.355705 | 14.09396 | 0.456479 | 17.326562 | 0.465758 | 40.639775 | false | true | 2024-08-17 | 2024-08-21 | 0 | microsoft/Phi-3.5-MoE-instruct |
microsoft_Phi-3.5-mini-instruct_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/Phi-3.5-mini-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/Phi-3.5-mini-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__Phi-3.5-mini-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/Phi-3.5-mini-instruct | 64963004ad95869fa73a30279371c8778509ac84 | 27.567573 | mit | 675 | 3 | true | false | false | true | 3.696004 | 0.57745 | 57.745005 | 0.551779 | 36.745854 | 0.159366 | 15.936556 | 0.339765 | 11.96868 | 0.402125 | 10.098958 | 0.396193 | 32.910387 | false | true | 2024-08-16 | 2024-08-21 | 0 | microsoft/Phi-3.5-mini-instruct |
microsoft_phi-1_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | PhiForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/phi-1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/phi-1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__phi-1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/phi-1 | b9ac0e6d78d43970ecf88e9e0154b3a7da20ed89 | 5.523966 | mit | 206 | 1 | true | false | false | false | 0.286229 | 0.206806 | 20.680572 | 0.313948 | 4.273999 | 0.006798 | 0.679758 | 0.265101 | 2.013423 | 0.35251 | 3.697135 | 0.11619 | 1.798907 | false | true | 2023-09-10 | 2024-06-13 | 0 | microsoft/phi-1 |
microsoft_phi-1_5_float16 | float16 | 🟢 pretrained | 🟢 | Original | PhiForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/phi-1_5" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/phi-1_5</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__phi-1_5-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/phi-1_5 | 675aa382d814580b22651a30acb1a585d7c25963 | 7.057674 | mit | 1,317 | 1 | true | false | false | false | 0.340862 | 0.203284 | 20.328395 | 0.335976 | 7.468939 | 0.011329 | 1.132931 | 0.267617 | 2.348993 | 0.340417 | 3.385417 | 0.169132 | 7.681368 | false | true | 2023-09-10 | 2024-06-09 | 0 | microsoft/phi-1_5 |
microsoft_phi-2_float16 | float16 | 🟢 pretrained | 🟢 | Original | PhiForCausalLM | <a target="_blank" href="https://huggingface.co/microsoft/phi-2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">microsoft/phi-2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/microsoft__phi-2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | microsoft/phi-2 | ef382358ec9e382308935a992d908de099b64c23 | 15.471351 | mit | 3,249 | 2 | true | false | false | false | 0.423521 | 0.273876 | 27.387554 | 0.488121 | 28.038519 | 0.02568 | 2.567976 | 0.271812 | 2.908277 | 0.409896 | 13.836979 | 0.262799 | 18.0888 | false | true | 2023-12-13 | 2024-06-09 | 0 | microsoft/phi-2 |
migtissera_Llama-3-70B-Synthia-v3.5_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Llama-3-70B-Synthia-v3.5" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Llama-3-70B-Synthia-v3.5</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Llama-3-70B-Synthia-v3.5-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Llama-3-70B-Synthia-v3.5 | 8744db0bccfc18f1847633da9d29fc89b35b4190 | 35.204299 | llama3 | 5 | 70 | true | false | false | true | 8.769698 | 0.60765 | 60.764992 | 0.648864 | 49.11816 | 0.189577 | 18.957704 | 0.387584 | 18.344519 | 0.492198 | 23.391406 | 0.465841 | 40.64901 | false | false | 2024-05-26 | 2024-08-28 | 0 | migtissera/Llama-3-70B-Synthia-v3.5 |
migtissera_Llama-3-8B-Synthia-v3.5_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Llama-3-8B-Synthia-v3.5" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Llama-3-8B-Synthia-v3.5</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Llama-3-8B-Synthia-v3.5-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Llama-3-8B-Synthia-v3.5 | af4990801a24fee7acf16370cb5aa5643b5e9d6c | 19.696678 | llama3 | 15 | 8 | true | false | false | true | 0.828698 | 0.506958 | 50.69582 | 0.488794 | 27.542339 | 0.050604 | 5.060423 | 0.271812 | 2.908277 | 0.404385 | 9.414844 | 0.303025 | 22.558363 | false | false | 2024-05-17 | 2024-08-28 | 0 | migtissera/Llama-3-8B-Synthia-v3.5 |
migtissera_Tess-3-7B-SFT_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Tess-3-7B-SFT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Tess-3-7B-SFT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Tess-3-7B-SFT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Tess-3-7B-SFT | 404de3b56564dbd43cd64d97f8574b43189462f3 | 17.096163 | apache-2.0 | 3 | 7 | true | false | false | true | 0.64717 | 0.394626 | 39.462626 | 0.460735 | 24.123847 | 0.033233 | 3.323263 | 0.270973 | 2.796421 | 0.411271 | 10.275521 | 0.303358 | 22.595301 | false | false | 2024-07-09 | 2024-07-20 | 1 | mistralai/Mistral-7B-v0.3 |
migtissera_Tess-3-Mistral-Nemo-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Tess-3-Mistral-Nemo-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Tess-3-Mistral-Nemo-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Tess-3-Mistral-Nemo-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Tess-3-Mistral-Nemo-12B | 0b82dea6e8f4aed4a1c2e10198d68991c30d171b | 16.543939 | apache-2.0 | 12 | 12 | true | false | false | true | 1.889992 | 0.3355 | 33.549981 | 0.489942 | 28.042728 | 0.046828 | 4.682779 | 0.250839 | 0.111857 | 0.445781 | 15.489323 | 0.256483 | 17.386968 | false | false | 2024-08-13 | 2024-09-16 | 0 | migtissera/Tess-3-Mistral-Nemo-12B |
migtissera_Tess-v2.5-Phi-3-medium-128k-14B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Tess-v2.5-Phi-3-medium-128k-14B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Tess-v2.5-Phi-3-medium-128k-14B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Tess-v2.5-Phi-3-medium-128k-14B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Tess-v2.5-Phi-3-medium-128k-14B | 3a4dbce32e765f659d418c57f0040d290b8b480d | 23.738382 | mit | 3 | 13 | true | false | false | true | 2.23817 | 0.453877 | 45.387682 | 0.620661 | 46.215828 | 0.026435 | 2.643505 | 0.307886 | 7.718121 | 0.411302 | 10.11276 | 0.373172 | 30.352394 | false | false | 2024-06-05 | 2024-08-30 | 1 | microsoft/Phi-3-medium-128k-instruct |
migtissera_Tess-v2.5.2-Qwen2-72B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Tess-v2.5.2-Qwen2-72B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Tess-v2.5.2-Qwen2-72B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Tess-v2.5.2-Qwen2-72B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Tess-v2.5.2-Qwen2-72B | 0435e634ad9bc8b1172395a535b78e6f25f3594f | 33.276047 | other | 11 | 72 | true | false | false | true | 14.613087 | 0.449431 | 44.943084 | 0.664679 | 52.308136 | 0.274169 | 27.416918 | 0.350671 | 13.422819 | 0.418833 | 10.8875 | 0.5561 | 50.677822 | false | false | 2024-06-13 | 2024-08-10 | 0 | migtissera/Tess-v2.5.2-Qwen2-72B |
migtissera_Trinity-2-Codestral-22B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Trinity-2-Codestral-22B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Trinity-2-Codestral-22B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Trinity-2-Codestral-22B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Trinity-2-Codestral-22B | 5f20b9d8af1a75c135c70bd7295e58301cce63fc | 21.819011 | other | 11 | 22 | true | false | false | true | 1.502157 | 0.420205 | 42.020506 | 0.559324 | 36.412738 | 0.086103 | 8.610272 | 0.314597 | 8.612975 | 0.411052 | 9.614844 | 0.330785 | 25.64273 | false | false | 2024-08-07 | 2024-09-16 | 1 | mistralai/Codestral-22B-v0.1 |
migtissera_Trinity-2-Codestral-22B-v0.2_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Trinity-2-Codestral-22B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Trinity-2-Codestral-22B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Trinity-2-Codestral-22B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Trinity-2-Codestral-22B-v0.2 | 63513c3eb9b7c552fc163f58a2e7dc1fa09573b5 | 21.869825 | other | 6 | 22 | true | false | false | true | 1.553522 | 0.434468 | 43.446832 | 0.568636 | 37.614246 | 0.083837 | 8.383686 | 0.300336 | 6.711409 | 0.404479 | 9.059896 | 0.334026 | 26.002881 | false | false | 2024-08-13 | 2024-08-28 | 1 | mistralai/Codestral-22B-v0.1 |
migtissera_Trinity-2-Codestral-22B-v0.2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/migtissera/Trinity-2-Codestral-22B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">migtissera/Trinity-2-Codestral-22B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/migtissera__Trinity-2-Codestral-22B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | migtissera/Trinity-2-Codestral-22B-v0.2 | 9452a82ac7bfa9092a061ec913e9078ef3525a03 | 22.1118 | other | 6 | 22 | true | false | false | true | 1.561208 | 0.443011 | 44.301121 | 0.570647 | 37.786041 | 0.07855 | 7.854985 | 0.307886 | 7.718121 | 0.403146 | 8.859896 | 0.335356 | 26.150635 | false | false | 2024-08-13 | 2024-09-16 | 1 | mistralai/Codestral-22B-v0.1 |
minghaowu_Qwen1.5-1.8B-OpenHermes-2.5_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/minghaowu/Qwen1.5-1.8B-OpenHermes-2.5" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">minghaowu/Qwen1.5-1.8B-OpenHermes-2.5</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/minghaowu__Qwen1.5-1.8B-OpenHermes-2.5-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | minghaowu/Qwen1.5-1.8B-OpenHermes-2.5 | 40700de82968350c192318877fe522630d0ef76d | 8.319696 | 0 | 1 | false | false | false | true | 1.0949 | 0.277797 | 27.779736 | 0.337464 | 7.561478 | 0.002266 | 0.226586 | 0.283557 | 4.474273 | 0.352885 | 1.077344 | 0.179189 | 8.798759 | false | false | 2024-09-12 | 0 | Removed |
||
ministral_Ministral-3b-instruct_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/ministral/Ministral-3b-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ministral/Ministral-3b-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ministral__Ministral-3b-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | ministral/Ministral-3b-instruct | 2c95908929198d6e69af8638f0dbbd9bc6b93f9e | 3.381613 | apache-2.0 | 33 | 3 | true | false | false | false | 0.264487 | 0.135764 | 13.576422 | 0.319186 | 4.675864 | 0 | 0 | 0.251678 | 0.223714 | 0.33825 | 0.78125 | 0.109292 | 1.032432 | false | false | 2024-03-14 | 2024-10-25 | 0 | ministral/Ministral-3b-instruct |
mistral-community_Mistral-7B-v0.2_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistral-community/Mistral-7B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistral-community/Mistral-7B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistral-community__Mistral-7B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistral-community/Mistral-7B-v0.2 | 2c3e624962b1a3f3fbf52e15969565caa7bc064a | 14.215362 | apache-2.0 | 232 | 7 | true | false | false | false | 0.553213 | 0.22664 | 22.663976 | 0.451019 | 23.950865 | 0.030211 | 3.021148 | 0.291946 | 5.592841 | 0.403177 | 8.363802 | 0.295296 | 21.699542 | false | true | 2024-03-23 | 2024-06-12 | 0 | mistral-community/Mistral-7B-v0.2 |
mistral-community_mixtral-8x22B-v0.3_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mistral-community/mixtral-8x22B-v0.3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistral-community/mixtral-8x22B-v0.3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistral-community__mixtral-8x22B-v0.3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistral-community/mixtral-8x22B-v0.3 | 211b177b79ab5ef245ee334d106c27623e786882 | 25.789407 | apache-2.0 | 3 | 140 | true | true | false | false | 52.494485 | 0.258264 | 25.826363 | 0.625 | 45.731041 | 0.182779 | 18.277946 | 0.377517 | 17.002237 | 0.403698 | 7.46224 | 0.46393 | 40.436613 | false | true | 2024-05-25 | 2024-06-13 | 0 | mistral-community/mixtral-8x22B-v0.3 |
mistralai_Codestral-22B-v0.1_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Codestral-22B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Codestral-22B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Codestral-22B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Codestral-22B-v0.1 | 8f5fe23af91885222a1563283c87416745a5e212 | 23.279917 | other | 1,156 | 22 | true | false | false | true | 1.30667 | 0.577175 | 57.717523 | 0.513914 | 30.737634 | 0.100453 | 10.045317 | 0.298658 | 6.487696 | 0.418708 | 10.738542 | 0.315575 | 23.952793 | false | true | 2024-05-29 | 2024-09-28 | 0 | mistralai/Codestral-22B-v0.1 |
mistralai_Ministral-8B-Instruct-2410_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Ministral-8B-Instruct-2410" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Ministral-8B-Instruct-2410</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Ministral-8B-Instruct-2410-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Ministral-8B-Instruct-2410 | 199e57c1d66379760f6413f79d27008d1d1dbd6e | 22.007859 | other | 358 | 8 | true | false | false | true | 0.797086 | 0.58964 | 58.963993 | 0.476164 | 25.824774 | 0.064955 | 6.495468 | 0.284396 | 4.58613 | 0.41375 | 10.71875 | 0.329122 | 25.458038 | false | true | 2024-10-15 | 2024-12-01 | 0 | mistralai/Ministral-8B-Instruct-2410 |
mistralai_Mistral-7B-Instruct-v0.1_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-7B-Instruct-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-7B-Instruct-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-7B-Instruct-v0.1 | 73068f3702d050a2fd5aa2ca1e612e5036429398 | 12.695701 | apache-2.0 | 1,534 | 7 | true | false | false | true | 1.216045 | 0.448706 | 44.87061 | 0.335481 | 7.647021 | 0.018127 | 1.812689 | 0.25 | 0 | 0.38476 | 6.128385 | 0.241439 | 15.715499 | false | true | 2023-09-27 | 2024-06-27 | 1 | mistralai/Mistral-7B-v0.1 |
mistralai_Mistral-7B-Instruct-v0.2_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-7B-Instruct-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-7B-Instruct-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-7B-Instruct-v0.2 | 41b61a33a2483885c981aa79e0df6b32407ed873 | 18.457539 | apache-2.0 | 2,586 | 7 | true | false | false | true | 0.534407 | 0.549623 | 54.962278 | 0.445974 | 22.910602 | 0.02719 | 2.719033 | 0.276007 | 3.467562 | 0.396604 | 7.608854 | 0.271692 | 19.076906 | false | true | 2023-12-11 | 2024-06-12 | 0 | mistralai/Mistral-7B-Instruct-v0.2 |
mistralai_Mistral-7B-Instruct-v0.3_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-7B-Instruct-v0.3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-7B-Instruct-v0.3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-7B-Instruct-v0.3 | 83e9aa141f2e28c82232fea5325f54edf17c43de | 19.174746 | apache-2.0 | 1,175 | 7 | true | false | false | true | 0.537783 | 0.546525 | 54.652544 | 0.472196 | 25.569115 | 0.035498 | 3.549849 | 0.279362 | 3.914989 | 0.373906 | 4.304948 | 0.307513 | 23.057033 | false | true | 2024-05-22 | 2024-06-12 | 1 | mistralai/Mistral-7B-v0.3 |
mistralai_Mistral-7B-v0.1_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-7B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-7B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-7B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-7B-v0.1 | 26bca36bde8333b5d7f72e9ed20ccda6a618af24 | 14.562619 | apache-2.0 | 3,480 | 7 | true | false | false | false | 0.675534 | 0.238555 | 23.855481 | 0.443107 | 22.168402 | 0.02719 | 2.719033 | 0.291946 | 5.592841 | 0.413938 | 10.675521 | 0.30128 | 22.364436 | false | true | 2023-09-20 | 2024-06-12 | 0 | mistralai/Mistral-7B-v0.1 |
mistralai_Mistral-7B-v0.3_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-7B-v0.3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-7B-v0.3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-7B-v0.3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-7B-v0.3 | b67d6a03ca097c5122fa65904fce0413500bf8c8 | 14.215362 | apache-2.0 | 400 | 7 | true | false | false | false | 0.660476 | 0.22664 | 22.663976 | 0.451019 | 23.950865 | 0.030211 | 3.021148 | 0.291946 | 5.592841 | 0.403177 | 8.363802 | 0.295296 | 21.699542 | false | true | 2024-05-22 | 2024-06-12 | 0 | mistralai/Mistral-7B-v0.3 |
mistralai_Mistral-Large-Instruct-2411_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-Large-Instruct-2411" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-Large-Instruct-2411</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-Large-Instruct-2411-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-Large-Instruct-2411 | 3a5cb136f6106edf5c1210369068eb5a4f787cab | 38.455231 | other | 163 | 122 | true | false | false | true | 26.272305 | 0.840058 | 84.005771 | 0.674665 | 52.744892 | 0.011329 | 1.132931 | 0.437081 | 24.944072 | 0.454 | 17.216667 | 0.556184 | 50.687057 | false | true | 2024-11-14 | 2024-11-19 | 0 | mistralai/Mistral-Large-Instruct-2411 |
mistralai_Mistral-Nemo-Base-2407_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-Nemo-Base-2407" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-Nemo-Base-2407</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-Nemo-Base-2407-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-Nemo-Base-2407 | d2efb15544d5401f761235bef327babb850887d0 | 15.138651 | apache-2.0 | 263 | 11 | true | false | false | false | 1.702995 | 0.162992 | 16.299197 | 0.503506 | 29.374736 | 0.053625 | 5.362538 | 0.293624 | 5.816555 | 0.392135 | 6.516927 | 0.347158 | 27.461953 | false | true | 2024-07-18 | 2024-07-19 | 0 | mistralai/Mistral-Nemo-Base-2407 |
mistralai_Mistral-Nemo-Instruct-2407_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-Nemo-Instruct-2407</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-Nemo-Instruct-2407-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-Nemo-Instruct-2407 | 4d14c1db68fe20dbf80b8eca85d39b909c5fe1d5 | 23.633374 | apache-2.0 | 1,285 | 12 | true | false | false | true | 2.997601 | 0.638025 | 63.802489 | 0.503652 | 29.67997 | 0.064955 | 6.495468 | 0.290268 | 5.369128 | 0.39 | 8.483333 | 0.351729 | 27.969858 | false | true | 2024-07-17 | 2024-08-29 | 1 | mistralai/Mistral-Nemo-Base-2407 |
mistralai_Mistral-Small-Instruct-2409_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-Small-Instruct-2409" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-Small-Instruct-2409</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-Small-Instruct-2409-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-Small-Instruct-2409 | 63e53df6575e7085d62113f4383835ff979b3795 | 26.262749 | other | 356 | 22 | true | false | false | true | 1.379338 | 0.666976 | 66.697585 | 0.521308 | 30.792096 | 0.143505 | 14.350453 | 0.323826 | 9.8434 | 0.363208 | 3.001042 | 0.396027 | 32.891918 | false | true | 2024-09-17 | 2024-09-19 | 0 | mistralai/Mistral-Small-Instruct-2409 |
mistralai_Mistral-Small-Instruct-2409_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mistral-Small-Instruct-2409" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mistral-Small-Instruct-2409</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mistral-Small-Instruct-2409-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mistral-Small-Instruct-2409 | 63e53df6575e7085d62113f4383835ff979b3795 | 29.818243 | other | 356 | 22 | true | false | false | false | 1.610007 | 0.628283 | 62.828296 | 0.583028 | 40.559713 | 0.197885 | 19.78852 | 0.333054 | 11.073826 | 0.406333 | 10.225 | 0.409907 | 34.434102 | false | true | 2024-09-17 | 2024-09-25 | 0 | mistralai/Mistral-Small-Instruct-2409 |
mistralai_Mixtral-8x22B-Instruct-v0.1_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mixtral-8x22B-Instruct-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mixtral-8x22B-Instruct-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mixtral-8x22B-Instruct-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mixtral-8x22B-Instruct-v0.1 | b0c3516041d014f640267b14feb4e9a84c8e8c71 | 33.88568 | apache-2.0 | 693 | 140 | true | true | false | true | 47.147579 | 0.718358 | 71.83584 | 0.612492 | 44.114346 | 0.187311 | 18.731118 | 0.373322 | 16.442953 | 0.431115 | 13.489323 | 0.448305 | 38.700502 | false | true | 2024-04-16 | 2024-06-12 | 1 | mistralai/Mixtral-8x22B-v0.1 |
mistralai_Mixtral-8x22B-v0.1_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mixtral-8x22B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mixtral-8x22B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mixtral-8x22B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mixtral-8x22B-v0.1 | b03e260818710044a2f088d88fab12bb220884fb | 25.728348 | apache-2.0 | 201 | 140 | true | true | false | false | 104.697316 | 0.258264 | 25.826363 | 0.623981 | 45.588404 | 0.182779 | 18.277946 | 0.375839 | 16.778523 | 0.403698 | 7.46224 | 0.46393 | 40.436613 | false | true | 2024-04-16 | 2024-06-12 | 0 | mistralai/Mixtral-8x22B-v0.1 |
mistralai_Mixtral-8x7B-Instruct-v0.1_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mixtral-8x7B-Instruct-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mixtral-8x7B-Instruct-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mixtral-8x7B-Instruct-v0.1 | 1e637f2d7cb0a9d6fb1922f305cb784995190a83 | 23.842279 | apache-2.0 | 4,218 | 46 | true | true | false | true | 13.764939 | 0.559914 | 55.991436 | 0.496237 | 29.742398 | 0.0929 | 9.29003 | 0.302852 | 7.04698 | 0.420323 | 11.073698 | 0.369182 | 29.909131 | false | true | 2023-12-10 | 2024-06-12 | 1 | mistralai/Mixtral-8x7B-v0.1 |
mistralai_Mixtral-8x7B-v0.1_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mixtral-8x7B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mixtral-8x7B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mixtral-8x7B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mixtral-8x7B-v0.1 | 985aa055896a8f943d4a9f2572e6ea1341823841 | 19.451988 | apache-2.0 | 1,651 | 46 | true | true | false | false | 18.387865 | 0.241527 | 24.152693 | 0.508667 | 30.294195 | 0.095166 | 9.516616 | 0.313758 | 8.501119 | 0.432135 | 12.583594 | 0.384973 | 31.663712 | false | true | 2023-12-01 | 2024-08-20 | 0 | mistralai/Mixtral-8x7B-v0.1 |
mistralai_Mixtral-8x7B-v0.1_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mistralai/Mixtral-8x7B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mistralai/Mixtral-8x7B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mistralai__Mixtral-8x7B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mistralai/Mixtral-8x7B-v0.1 | 985aa055896a8f943d4a9f2572e6ea1341823841 | 19.665109 | apache-2.0 | 1,651 | 46 | true | true | false | false | 5.1351 | 0.232609 | 23.260948 | 0.509771 | 30.400299 | 0.093656 | 9.365559 | 0.32047 | 9.395973 | 0.441313 | 13.664063 | 0.387134 | 31.903812 | false | true | 2023-12-01 | 2024-06-27 | 0 | mistralai/Mixtral-8x7B-v0.1 |
mixtao_MixTAO-7Bx2-MoE-v8.1_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mixtao/MixTAO-7Bx2-MoE-v8.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mixtao/MixTAO-7Bx2-MoE-v8.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mixtao__MixTAO-7Bx2-MoE-v8.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mixtao/MixTAO-7Bx2-MoE-v8.1 | 339130b87b6ef2484fea9fbfacba8a714ac03347 | 21.077927 | apache-2.0 | 55 | 12 | true | true | false | false | 0.924035 | 0.416233 | 41.623337 | 0.518906 | 32.310342 | 0.090634 | 9.063444 | 0.284396 | 4.58613 | 0.446333 | 15.291667 | 0.312334 | 23.592642 | false | false | 2024-02-26 | 2024-10-04 | 0 | mixtao/MixTAO-7Bx2-MoE-v8.1 |
mkxu_llama-3-8b-po1_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mkxu/llama-3-8b-po1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mkxu/llama-3-8b-po1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mkxu__llama-3-8b-po1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mkxu/llama-3-8b-po1 | 1b16e10de696c43cd2b49fac9f6195dc551438ee | 19.767002 | 0 | 8 | false | false | false | false | 0.512188 | 0.408115 | 40.811491 | 0.497609 | 29.181759 | 0.070242 | 7.024169 | 0.29698 | 6.263982 | 0.380417 | 6.852083 | 0.356217 | 28.468528 | false | false | 2024-11-29 | 2024-11-29 | 0 | mkxu/llama-3-8b-po1 |
|
mlabonne_AlphaMonarch-7B_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/AlphaMonarch-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/AlphaMonarch-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__AlphaMonarch-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/AlphaMonarch-7B | 3de065d84411d74e5b3590f67f52b0b71faf6161 | 17.655797 | cc-by-nc-4.0 | 148 | 7 | true | false | false | true | 0.572572 | 0.493944 | 49.394385 | 0.462552 | 23.947378 | 0.042296 | 4.229607 | 0.270134 | 2.684564 | 0.412135 | 9.316927 | 0.247257 | 16.361924 | true | true | 2024-02-14 | 2024-06-12 | 1 | mlabonne/AlphaMonarch-7B (Merge) |
mlabonne_Beyonder-4x7B-v3_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/Beyonder-4x7B-v3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/Beyonder-4x7B-v3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__Beyonder-4x7B-v3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/Beyonder-4x7B-v3 | 8e923fa480f511ab54d79b44b0487768bdd3de4e | 19.381682 | cc-by-nc-4.0 | 58 | 24 | true | true | false | true | 1.386303 | 0.560839 | 56.083857 | 0.467052 | 24.557209 | 0.052115 | 5.21148 | 0.285235 | 4.697987 | 0.404542 | 8.934375 | 0.251247 | 16.805186 | true | true | 2024-03-21 | 2024-06-12 | 1 | mlabonne/Beyonder-4x7B-v3 (Merge) |
mlabonne_BigQwen2.5-52B-Instruct_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/BigQwen2.5-52B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/BigQwen2.5-52B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__BigQwen2.5-52B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/BigQwen2.5-52B-Instruct | 425b9bffc9871085cc0d42c34138ce776f96ba02 | 37.558281 | apache-2.0 | 2 | 52 | true | false | false | true | 27.563508 | 0.792872 | 79.28718 | 0.7121 | 59.809607 | 0.186556 | 18.655589 | 0.302013 | 6.935123 | 0.411302 | 10.446094 | 0.551945 | 50.21609 | true | true | 2024-09-23 | 2024-09-25 | 1 | mlabonne/BigQwen2.5-52B-Instruct (Merge) |
mlabonne_BigQwen2.5-Echo-47B-Instruct_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/BigQwen2.5-Echo-47B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/BigQwen2.5-Echo-47B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__BigQwen2.5-Echo-47B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/BigQwen2.5-Echo-47B-Instruct | f95fcf22f8ab87c2dbb1893b87c8a132820acb5e | 30.322429 | apache-2.0 | 3 | 47 | true | false | false | true | 8.523077 | 0.735669 | 73.566914 | 0.612511 | 44.522244 | 0.035498 | 3.549849 | 0.314597 | 8.612975 | 0.412479 | 10.193229 | 0.473404 | 41.489362 | true | true | 2024-09-23 | 2024-09-24 | 1 | mlabonne/BigQwen2.5-Echo-47B-Instruct (Merge) |
mlabonne_ChimeraLlama-3-8B-v2_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/ChimeraLlama-3-8B-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/ChimeraLlama-3-8B-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__ChimeraLlama-3-8B-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/ChimeraLlama-3-8B-v2 | d90a12b1574d7be084e53e0ad610282638ab29cf | 20.133093 | other | 14 | 8 | true | false | false | false | 0.83741 | 0.446883 | 44.688316 | 0.50456 | 28.478796 | 0.09139 | 9.138973 | 0.285235 | 4.697987 | 0.379083 | 5.252083 | 0.356882 | 28.542405 | true | true | 2024-04-22 | 2024-08-25 | 1 | mlabonne/ChimeraLlama-3-8B-v2 (Merge) |
mlabonne_ChimeraLlama-3-8B-v3_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/ChimeraLlama-3-8B-v3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/ChimeraLlama-3-8B-v3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__ChimeraLlama-3-8B-v3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/ChimeraLlama-3-8B-v3 | c8c1787e1426e3979ae82134f4eb7fa332f58ae0 | 20.684542 | other | 15 | 8 | true | false | false | false | 0.82374 | 0.440788 | 44.078822 | 0.497819 | 27.646094 | 0.087613 | 8.761329 | 0.291946 | 5.592841 | 0.400354 | 8.377604 | 0.366855 | 29.650561 | true | true | 2024-05-01 | 2024-08-25 | 1 | mlabonne/ChimeraLlama-3-8B-v3 (Merge) |
mlabonne_Daredevil-8B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/Daredevil-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/Daredevil-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__Daredevil-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/Daredevil-8B | 717953c83631cc9adf2dddccfff06739308f10f7 | 22.308941 | other | 36 | 8 | true | false | false | true | 1.511915 | 0.454777 | 45.477666 | 0.519441 | 31.626855 | 0.100453 | 10.045317 | 0.307886 | 7.718121 | 0.393875 | 7.534375 | 0.383062 | 31.451315 | true | true | 2024-05-25 | 2024-07-02 | 1 | mlabonne/Daredevil-8B (Merge) |
mlabonne_Daredevil-8B-abliterated_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/Daredevil-8B-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/Daredevil-8B-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__Daredevil-8B-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/Daredevil-8B-abliterated | 034c0ce8ceeba075d1dff2bac1b113a017c79390 | 19.72576 | other | 33 | 8 | true | false | false | true | 1.198362 | 0.442637 | 44.263665 | 0.425427 | 19.865777 | 0.096677 | 9.667674 | 0.290268 | 5.369128 | 0.407021 | 9.177604 | 0.370096 | 30.010712 | false | true | 2024-05-26 | 2024-07-02 | 0 | mlabonne/Daredevil-8B-abliterated |
mlabonne_Hermes-3-Llama-3.1-70B-lorablated_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/Hermes-3-Llama-3.1-70B-lorablated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/Hermes-3-Llama-3.1-70B-lorablated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__Hermes-3-Llama-3.1-70B-lorablated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/Hermes-3-Llama-3.1-70B-lorablated | 4303ff3b524418e9aa5e787d60595a44a6173b02 | 31.544445 | 25 | 70 | true | false | false | false | 30.879892 | 0.342444 | 34.244361 | 0.669317 | 52.750073 | 0.212236 | 21.223565 | 0.365772 | 15.436242 | 0.502927 | 24.732552 | 0.467919 | 40.879876 | true | true | 2024-08-16 | 2024-11-27 | 1 | mlabonne/Hermes-3-Llama-3.1-70B-lorablated (Merge) |
|
mlabonne_Meta-Llama-3.1-8B-Instruct-abliterated_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__Meta-Llama-3.1-8B-Instruct-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated | aef878bdf42c119d007322967006fcdef5ae6ee1 | 23.290888 | llama3.1 | 139 | 8 | true | false | false | true | 2.45405 | 0.73447 | 73.447009 | 0.487406 | 27.129165 | 0.072508 | 7.250755 | 0.256711 | 0.894855 | 0.364885 | 3.210677 | 0.350316 | 27.812869 | false | true | 2024-07-24 | 2024-10-13 | 2 | meta-llama/Meta-Llama-3.1-8B |
mlabonne_NeuralBeagle14-7B_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/NeuralBeagle14-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/NeuralBeagle14-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__NeuralBeagle14-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/NeuralBeagle14-7B | 1567ad618a0998139654cb355738bb9bc018ca64 | 18.947841 | cc-by-nc-4.0 | 158 | 7 | true | false | false | true | 0.671707 | 0.493519 | 49.351942 | 0.462787 | 23.959695 | 0.054381 | 5.438066 | 0.281879 | 4.250559 | 0.431948 | 12.89349 | 0.26014 | 17.793292 | true | true | 2024-01-15 | 2024-06-27 | 2 | mlabonne/Beagle14-7B (Merge) |
mlabonne_NeuralDaredevil-8B-abliterated_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/NeuralDaredevil-8B-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__NeuralDaredevil-8B-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/NeuralDaredevil-8B-abliterated | 2f4a5e8a8522f19dff345c7189b7891468763061 | 27.148976 | llama3 | 157 | 8 | true | false | false | true | 2.707744 | 0.756077 | 75.607721 | 0.511057 | 30.307986 | 0.088369 | 8.836858 | 0.306208 | 7.494407 | 0.401938 | 9.075521 | 0.384142 | 31.571365 | false | true | 2024-05-27 | 2024-07-25 | 0 | mlabonne/NeuralDaredevil-8B-abliterated |
mlabonne_NeuralDaredevil-8B-abliterated_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/NeuralDaredevil-8B-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__NeuralDaredevil-8B-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/NeuralDaredevil-8B-abliterated | 89b01e3292e031ed85ad21545849182f5627021e | 21.499914 | llama3 | 157 | 8 | true | false | false | false | 0.985007 | 0.416233 | 41.623337 | 0.512396 | 29.763198 | 0.085347 | 8.534743 | 0.302852 | 7.04698 | 0.414958 | 10.903125 | 0.380153 | 31.128103 | false | true | 2024-05-27 | 2024-06-27 | 0 | mlabonne/NeuralDaredevil-8B-abliterated |
mlabonne_OrpoLlama-3-8B_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/OrpoLlama-3-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/OrpoLlama-3-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__OrpoLlama-3-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/OrpoLlama-3-8B | 7f200e4c84ad0daa3ff6bc414012d8d0bacbf90e | 14.980803 | other | 54 | 8 | true | false | false | true | 0.8903 | 0.365275 | 36.527525 | 0.442408 | 21.954108 | 0.045317 | 4.531722 | 0.279362 | 3.914989 | 0.357938 | 4.008854 | 0.270529 | 18.947621 | false | true | 2024-04-18 | 2024-06-12 | 1 | meta-llama/Meta-Llama-3-8B |
mlabonne_phixtral-2x2_8_float16 | float16 | 🤝 base merges and moerges | 🤝 | Original | PhiForCausalLM | <a target="_blank" href="https://huggingface.co/mlabonne/phixtral-2x2_8" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mlabonne/phixtral-2x2_8</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mlabonne__phixtral-2x2_8-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mlabonne/phixtral-2x2_8 | 7744a977d83f132ae5808d8c3b70157031f7de44 | 15.464997 | mit | 146 | 4 | true | true | false | true | 0.960951 | 0.343118 | 34.311848 | 0.488859 | 28.502645 | 0.030211 | 3.021148 | 0.265101 | 2.013423 | 0.364354 | 7.710938 | 0.25507 | 17.229979 | false | true | 2024-01-07 | 2024-06-12 | 0 | mlabonne/phixtral-2x2_8 |
mmnga_Llama-3-70B-japanese-suzume-vector-v0.1_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/mmnga/Llama-3-70B-japanese-suzume-vector-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mmnga/Llama-3-70B-japanese-suzume-vector-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mmnga__Llama-3-70B-japanese-suzume-vector-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mmnga/Llama-3-70B-japanese-suzume-vector-v0.1 | 16f98b2d45684af2c4a9ff5da75b00ef13cca808 | 30.883569 | llama3 | 4 | 70 | true | false | false | true | 16.097125 | 0.464893 | 46.489315 | 0.654176 | 50.022661 | 0.26284 | 26.283988 | 0.286074 | 4.809843 | 0.414063 | 10.757813 | 0.52244 | 46.937796 | false | false | 2024-04-28 | 2024-09-19 | 0 | mmnga/Llama-3-70B-japanese-suzume-vector-v0.1 |
moeru-ai_L3.1-Moe-2x8B-v0.2_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/moeru-ai/L3.1-Moe-2x8B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">moeru-ai/L3.1-Moe-2x8B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/moeru-ai__L3.1-Moe-2x8B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | moeru-ai/L3.1-Moe-2x8B-v0.2 | 1a0b4d4d1e839e332c67c9c16a2fc1f7ccc7f81e | 28.588568 | llama3.1 | 6 | 13 | true | true | false | true | 1.926568 | 0.734795 | 73.479479 | 0.525569 | 32.945891 | 0.152568 | 15.256798 | 0.300336 | 6.711409 | 0.419854 | 11.381771 | 0.385805 | 31.756058 | true | false | 2024-10-25 | 2024-10-25 | 1 | moeru-ai/L3.1-Moe-2x8B-v0.2 (Merge) |
moeru-ai_L3.1-Moe-4x8B-v0.1_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/moeru-ai/L3.1-Moe-4x8B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">moeru-ai/L3.1-Moe-4x8B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/moeru-ai__L3.1-Moe-4x8B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | moeru-ai/L3.1-Moe-4x8B-v0.1 | f8d477fad4c02c099c80ef38865c01e2c882e996 | 19.126854 | llama3.1 | 3 | 24 | true | true | false | true | 5.805487 | 0.433219 | 43.321941 | 0.493928 | 27.856765 | 0.111027 | 11.102719 | 0.259228 | 1.230425 | 0.360917 | 3.98125 | 0.345412 | 27.268026 | true | false | 2024-10-23 | 2024-10-23 | 1 | moeru-ai/L3.1-Moe-4x8B-v0.1 (Merge) |
moeru-ai_L3.1-Moe-4x8B-v0.2_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MixtralForCausalLM | <a target="_blank" href="https://huggingface.co/moeru-ai/L3.1-Moe-4x8B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">moeru-ai/L3.1-Moe-4x8B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/moeru-ai__L3.1-Moe-4x8B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | moeru-ai/L3.1-Moe-4x8B-v0.2 | fab49d865eb51f00e955c5624712184c39d207c9 | 17.467109 | llama3.1 | 2 | 24 | true | true | false | true | 3.422218 | 0.540655 | 54.065546 | 0.446626 | 21.337007 | 0.05287 | 5.287009 | 0.266779 | 2.237136 | 0.323396 | 2.291146 | 0.276263 | 19.584811 | true | false | 2024-10-30 | 2024-10-30 | 1 | moeru-ai/L3.1-Moe-4x8B-v0.2 (Merge) |
monsterapi_Llama-3_1-8B-Instruct-orca-ORPO_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Adapter | ? | <a target="_blank" href="https://huggingface.co/monsterapi/Llama-3_1-8B-Instruct-orca-ORPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">monsterapi/Llama-3_1-8B-Instruct-orca-ORPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/monsterapi__Llama-3_1-8B-Instruct-orca-ORPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | monsterapi/Llama-3_1-8B-Instruct-orca-ORPO | 5206a32e0bd3067aef1ce90f5528ade7d866253f | 4.832138 | apache-2.0 | 2 | 16 | true | false | false | true | 1.53268 | 0.227289 | 22.728915 | 0.286536 | 1.340469 | 0 | 0 | 0.249161 | 0 | 0.344479 | 3.059896 | 0.116772 | 1.863549 | false | false | 2024-08-01 | 2024-08-30 | 2 | meta-llama/Meta-Llama-3.1-8B |
monsterapi_gemma-2-2b-LoRA-MonsterInstruct_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/monsterapi/gemma-2-2b-LoRA-MonsterInstruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">monsterapi/gemma-2-2b-LoRA-MonsterInstruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/monsterapi__gemma-2-2b-LoRA-MonsterInstruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | monsterapi/gemma-2-2b-LoRA-MonsterInstruct | 6422e27e96e15cf93b966c973aacc15f8a27a458 | 11.865291 | gemma | 0 | 2 | true | false | false | true | 1.338688 | 0.390255 | 39.025452 | 0.364969 | 11.965057 | 0.011329 | 1.132931 | 0.270134 | 2.684564 | 0.364385 | 5.414844 | 0.19872 | 10.968898 | false | false | 2024-08-03 | 2024-08-05 | 0 | monsterapi/gemma-2-2b-LoRA-MonsterInstruct |
mosaicml_mpt-7b_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | MPTForCausalLM | <a target="_blank" href="https://huggingface.co/mosaicml/mpt-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">mosaicml/mpt-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/mosaicml__mpt-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | mosaicml/mpt-7b | 039e37745f00858f0e01e988383a8c4393b1a4f5 | 5.994265 | apache-2.0 | 1,163 | 7 | true | false | false | false | 0.643503 | 0.215199 | 21.519901 | 0.329974 | 6.550601 | 0.013595 | 1.359517 | 0.260067 | 1.342282 | 0.36724 | 2.904948 | 0.120595 | 2.288342 | false | true | 2023-05-05 | 2024-06-08 | 0 | mosaicml/mpt-7b |
natong19_Mistral-Nemo-Instruct-2407-abliterated_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/natong19/Mistral-Nemo-Instruct-2407-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">natong19/Mistral-Nemo-Instruct-2407-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/natong19__Mistral-Nemo-Instruct-2407-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | natong19/Mistral-Nemo-Instruct-2407-abliterated | 9c7087f62e6ab10ec4aeeb268e25cb3d4000696b | 23.872107 | apache-2.0 | 9 | 12 | true | false | false | true | 1.237857 | 0.639224 | 63.922393 | 0.504845 | 29.915044 | 0.063444 | 6.344411 | 0.286913 | 4.9217 | 0.403333 | 10.15 | 0.351812 | 27.979093 | false | false | 2024-08-15 | 2024-09-21 | 0 | natong19/Mistral-Nemo-Instruct-2407-abliterated |
natong19_Qwen2-7B-Instruct-abliterated_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/natong19/Qwen2-7B-Instruct-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">natong19/Qwen2-7B-Instruct-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/natong19__Qwen2-7B-Instruct-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | natong19/Qwen2-7B-Instruct-abliterated | 127962453ae87879719a82a97384ac1859787a25 | 25.758544 | apache-2.0 | 5 | 7 | true | false | false | true | 1.075836 | 0.583695 | 58.36946 | 0.555304 | 37.746834 | 0.111027 | 11.102719 | 0.301174 | 6.823266 | 0.403427 | 8.928385 | 0.384225 | 31.5806 | false | false | 2024-06-14 | 2024-07-29 | 0 | natong19/Qwen2-7B-Instruct-abliterated |
nazimali_Mistral-Nemo-Kurdish_bfloat16 | bfloat16 | 🟩 continuously pretrained | 🟩 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nazimali/Mistral-Nemo-Kurdish" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nazimali/Mistral-Nemo-Kurdish</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nazimali__Mistral-Nemo-Kurdish-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nazimali/Mistral-Nemo-Kurdish | 1eb544577a2874d8df0b77ca83ff1c88dd20f481 | 19.368945 | apache-2.0 | 2 | 12 | true | false | false | false | 1.849727 | 0.340121 | 34.012088 | 0.513332 | 29.855897 | 0.089124 | 8.912387 | 0.301174 | 6.823266 | 0.411573 | 11.779948 | 0.323471 | 24.830083 | false | false | 2024-10-09 | 2024-10-14 | 1 | nazimali/Mistral-Nemo-Kurdish (Merge) |
nazimali_Mistral-Nemo-Kurdish-Instruct_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nazimali/Mistral-Nemo-Kurdish-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nazimali/Mistral-Nemo-Kurdish-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nazimali__Mistral-Nemo-Kurdish-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nazimali/Mistral-Nemo-Kurdish-Instruct | 512140572f11203441e60ca26b5ede2b9979cb1d | 18.555958 | apache-2.0 | 2 | 12 | true | false | false | true | 1.702117 | 0.496392 | 49.63918 | 0.469942 | 25.561423 | 0.004532 | 0.453172 | 0.282718 | 4.362416 | 0.397875 | 8.401042 | 0.306267 | 22.918514 | false | false | 2024-10-09 | 2024-10-14 | 1 | nazimali/Mistral-Nemo-Kurdish-Instruct (Merge) |
nazimali_Mistral-Nemo-Kurdish-Instruct_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nazimali/Mistral-Nemo-Kurdish-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nazimali/Mistral-Nemo-Kurdish-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nazimali__Mistral-Nemo-Kurdish-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nazimali/Mistral-Nemo-Kurdish-Instruct | 512140572f11203441e60ca26b5ede2b9979cb1d | 18.589106 | apache-2.0 | 2 | 12 | true | false | false | true | 1.751721 | 0.486 | 48.600048 | 0.472144 | 26.021741 | 0.003021 | 0.302115 | 0.284396 | 4.58613 | 0.400573 | 8.838281 | 0.308677 | 23.186318 | false | false | 2024-10-09 | 2024-10-14 | 1 | nazimali/Mistral-Nemo-Kurdish-Instruct (Merge) |
nbeerbower_Flammades-Mistral-Nemo-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Flammades-Mistral-Nemo-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Flammades-Mistral-Nemo-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Flammades-Mistral-Nemo-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Flammades-Mistral-Nemo-12B | ddc76d1976af06aedc7f06bbffcaa34166c1cbdd | 22.466019 | apache-2.0 | 2 | 12 | true | false | false | false | 1.62742 | 0.38416 | 38.415959 | 0.529961 | 32.393772 | 0.069486 | 6.94864 | 0.303691 | 7.158837 | 0.480625 | 20.311458 | 0.366107 | 29.56745 | false | false | 2024-10-05 | 2024-10-06 | 1 | nbeerbower/Flammades-Mistral-Nemo-12B (Merge) |
nbeerbower_Gemma2-Gutenberg-Doppel-9B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Gemma2-Gutenberg-Doppel-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Gemma2-Gutenberg-Doppel-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Gemma2-Gutenberg-Doppel-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Gemma2-Gutenberg-Doppel-9B | f425bc69783891088e89e0afe44ec62b730567ba | 29.835999 | gemma | 4 | 9 | true | false | false | false | 1.947316 | 0.717109 | 71.710949 | 0.587011 | 41.083063 | 0.035498 | 3.549849 | 0.329698 | 10.626398 | 0.460781 | 17.297656 | 0.412733 | 34.748079 | false | false | 2024-09-29 | 2024-10-01 | 1 | nbeerbower/Gemma2-Gutenberg-Doppel-9B (Merge) |
nbeerbower_Gutensuppe-mistral-nemo-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Gutensuppe-mistral-nemo-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Gutensuppe-mistral-nemo-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Gutensuppe-mistral-nemo-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Gutensuppe-mistral-nemo-12B | 6ee13f347071bc3c4ee95c9dc3488a4093927143 | 22.080224 | 5 | 12 | false | false | false | false | 1.556249 | 0.291611 | 29.16107 | 0.548683 | 35.569348 | 0.120091 | 12.009063 | 0.337248 | 11.63311 | 0.429031 | 14.328906 | 0.368019 | 29.779846 | false | false | 2024-08-23 | 2024-09-03 | 1 | nbeerbower/Gutensuppe-mistral-nemo-12B (Merge) |
|
nbeerbower_Hermes2-Gutenberg2-Mistral-7B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Hermes2-Gutenberg2-Mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Hermes2-Gutenberg2-Mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Hermes2-Gutenberg2-Mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Hermes2-Gutenberg2-Mistral-7B | 5eec0dfd29999ef1d7775010b7e9c7be9ed89bfd | 19.376449 | apache-2.0 | 2 | 7 | true | false | false | false | 0.581121 | 0.372145 | 37.21448 | 0.498145 | 28.907335 | 0.058157 | 5.81571 | 0.28943 | 5.257271 | 0.462302 | 16.921094 | 0.299285 | 22.142804 | false | false | 2024-09-30 | 2024-10-01 | 1 | nbeerbower/Hermes2-Gutenberg2-Mistral-7B (Merge) |
nbeerbower_Llama-3.1-Nemotron-lorablated-70B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Llama-3.1-Nemotron-lorablated-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Llama-3.1-Nemotron-lorablated-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Llama-3.1-Nemotron-lorablated-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Llama-3.1-Nemotron-lorablated-70B | f335a582cdb7fb0e63a7343a908766ebd0ed9882 | 40.725393 | llama3.1 | 12 | 70 | true | false | false | false | 41.991923 | 0.72288 | 72.287974 | 0.682505 | 54.182581 | 0.324773 | 32.477341 | 0.39094 | 18.791946 | 0.468167 | 18.354167 | 0.534325 | 48.258348 | true | false | 2024-10-17 | 2024-11-27 | 1 | nbeerbower/Llama-3.1-Nemotron-lorablated-70B (Merge) |
nbeerbower_Llama3.1-Gutenberg-Doppel-70B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Llama3.1-Gutenberg-Doppel-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Llama3.1-Gutenberg-Doppel-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Llama3.1-Gutenberg-Doppel-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Llama3.1-Gutenberg-Doppel-70B | 5de156e97f776ce1b88ce5b2e2dc1e7709205a82 | 35.803048 | llama3.1 | 5 | 70 | true | false | false | true | 9.993593 | 0.709216 | 70.921599 | 0.666089 | 52.556779 | 0.145015 | 14.501511 | 0.344799 | 12.639821 | 0.489719 | 22.68151 | 0.473654 | 41.517066 | false | false | 2024-10-11 | 2024-10-12 | 1 | nbeerbower/Llama3.1-Gutenberg-Doppel-70B (Merge) |
nbeerbower_Lyra-Gutenberg-mistral-nemo-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Lyra-Gutenberg-mistral-nemo-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Lyra-Gutenberg-mistral-nemo-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Lyra-Gutenberg-mistral-nemo-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Lyra-Gutenberg-mistral-nemo-12B | 5c506391eb02075e02f4cf5953b443505d646bce | 22.703718 | cc-by-nc-4.0 | 17 | 12 | true | false | false | true | 1.918602 | 0.349488 | 34.948825 | 0.558625 | 36.992432 | 0.09139 | 9.138973 | 0.333893 | 11.185682 | 0.435667 | 14.758333 | 0.362783 | 29.198064 | false | false | 2024-08-23 | 2024-09-03 | 1 | nbeerbower/Lyra-Gutenberg-mistral-nemo-12B (Merge) |
nbeerbower_Lyra4-Gutenberg-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Lyra4-Gutenberg-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Lyra4-Gutenberg-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Lyra4-Gutenberg-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Lyra4-Gutenberg-12B | cb6911be3475da99a810071c04803d6edfb5965b | 19.818943 | cc-by-nc-4.0 | 19 | 12 | true | false | false | false | 1.690534 | 0.221219 | 22.121859 | 0.538669 | 34.235593 | 0.128399 | 12.839879 | 0.318792 | 9.17226 | 0.403792 | 11.973958 | 0.357131 | 28.570109 | false | false | 2024-09-09 | 2024-09-12 | 1 | nbeerbower/Lyra4-Gutenberg-12B (Merge) |
nbeerbower_Lyra4-Gutenberg2-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Lyra4-Gutenberg2-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Lyra4-Gutenberg2-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Lyra4-Gutenberg2-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Lyra4-Gutenberg2-12B | 6a5f117695cc729de16da87654b979e6df72ed2f | 19.932294 | cc-by-nc-4.0 | 8 | 12 | true | false | false | false | 1.80934 | 0.258513 | 25.851297 | 0.534453 | 33.73064 | 0.116314 | 11.63142 | 0.312919 | 8.389262 | 0.397219 | 11.485677 | 0.356549 | 28.505467 | false | false | 2024-09-29 | 2024-10-01 | 1 | nbeerbower/Lyra4-Gutenberg2-12B (Merge) |
nbeerbower_Mahou-1.5-mistral-nemo-12B-lorablated_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mahou-1.5-mistral-nemo-12B-lorablated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated | 8c9eecaace50659647c7d8b569237ad068a6c837 | 26.53481 | apache-2.0 | 2 | 12 | true | false | false | true | 1.405424 | 0.682488 | 68.248802 | 0.549604 | 36.077381 | 0.058157 | 5.81571 | 0.279362 | 3.914989 | 0.452167 | 16.554167 | 0.35738 | 28.597813 | true | false | 2024-10-19 | 2024-10-19 | 1 | nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated (Merge) |
nbeerbower_Mistral-Gutenberg-Doppel-7B-FFT_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Gutenberg-Doppel-7B-FFT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Gutenberg-Doppel-7B-FFT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Gutenberg-Doppel-7B-FFT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Gutenberg-Doppel-7B-FFT | 5735876465b6f2523fdedb73120c3f97d04556d3 | 18.338276 | apache-2.0 | 1 | 7 | true | false | false | true | 0.436816 | 0.57168 | 57.167981 | 0.407625 | 17.346575 | 0.024924 | 2.492447 | 0.283557 | 4.474273 | 0.405938 | 9.342188 | 0.272856 | 19.206191 | false | false | 2024-11-18 | 2024-11-18 | 1 | nbeerbower/Mistral-Gutenberg-Doppel-7B-FFT (Merge) |
nbeerbower_Mistral-Nemo-Gutenberg-Doppel-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Nemo-Gutenberg-Doppel-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B | 0eaaac89d4b53e94d5b78220b24439a026ee29e6 | 21.475046 | apache-2.0 | 3 | 12 | true | false | false | false | 1.776771 | 0.356707 | 35.670687 | 0.527461 | 32.421527 | 0.117825 | 11.782477 | 0.316275 | 8.836689 | 0.413219 | 11.485677 | 0.357879 | 28.653221 | false | false | 2024-09-26 | 2024-09-26 | 1 | nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B (Merge) |
nbeerbower_Mistral-Nemo-Gutenberg-Doppel-12B-v2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Nemo-Gutenberg-Doppel-12B-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2 | adc1ccd9d83d24e41bed895f989803af87ea2d2c | 24.717981 | apache-2.0 | 7 | 12 | true | false | false | true | 1.404857 | 0.653587 | 65.358693 | 0.53745 | 34.357413 | 0.044562 | 4.456193 | 0.270973 | 2.796421 | 0.423302 | 13.046094 | 0.354638 | 28.29307 | false | false | 2024-10-04 | 2024-10-09 | 1 | nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2 (Merge) |
nbeerbower_Mistral-Nemo-Moderne-12B-FFT-experimental_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Nemo-Moderne-12B-FFT-experimental" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Nemo-Moderne-12B-FFT-experimental</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Nemo-Moderne-12B-FFT-experimental-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Nemo-Moderne-12B-FFT-experimental | e52f7b7c3ade2a6212f29dd1054332cee21ab85d | 17.175357 | apache-2.0 | 1 | 12 | true | false | false | true | 1.217157 | 0.335225 | 33.522498 | 0.523409 | 32.07154 | 0.020393 | 2.039275 | 0.28104 | 4.138702 | 0.37149 | 4.002865 | 0.345495 | 27.277261 | false | false | 2024-11-19 | 2024-11-26 | 1 | nbeerbower/Mistral-Nemo-Moderne-12B-FFT-experimental (Merge) |
nbeerbower_Mistral-Nemo-Prism-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Nemo-Prism-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Nemo-Prism-12B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Nemo-Prism-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Nemo-Prism-12B | a39e1c8c083c172aaa3ca81faf8ba3b4799a888f | 27.344954 | apache-2.0 | 3 | 12 | true | false | false | true | 0.959288 | 0.68581 | 68.581032 | 0.547519 | 35.918008 | 0.052115 | 5.21148 | 0.307886 | 7.718121 | 0.462615 | 17.960156 | 0.358128 | 28.680925 | false | false | 2024-11-12 | 2024-11-12 | 1 | nbeerbower/Mistral-Nemo-Prism-12B (Merge) |
nbeerbower_Mistral-Nemo-Prism-12B-v2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Nemo-Prism-12B-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Nemo-Prism-12B-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Nemo-Prism-12B-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Nemo-Prism-12B-v2 | d7545999274cb56b5f961580b5234e8a647e023a | 27.541764 | apache-2.0 | 2 | 12 | true | false | false | true | 0.935552 | 0.697401 | 69.740067 | 0.549188 | 36.199788 | 0.057402 | 5.740181 | 0.305369 | 7.38255 | 0.459979 | 17.664063 | 0.356715 | 28.523936 | false | false | 2024-11-12 | 2024-11-26 | 1 | nbeerbower/Mistral-Nemo-Prism-12B-v2 (Merge) |
nbeerbower_Mistral-Nemo-Prism-12B-v7_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Nemo-Prism-12B-v7" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Nemo-Prism-12B-v7</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Nemo-Prism-12B-v7-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Nemo-Prism-12B-v7 | 0c9da9f3903be14fda1fcae245c22f873442b86f | 27.342442 | apache-2.0 | 5 | 12 | true | false | false | true | 0.975511 | 0.696152 | 69.615177 | 0.55211 | 36.440017 | 0.045317 | 4.531722 | 0.299497 | 6.599553 | 0.463885 | 18.085677 | 0.359043 | 28.782506 | false | false | 2024-11-13 | 2024-11-26 | 1 | nbeerbower/Mistral-Nemo-Prism-12B-v7 (Merge) |
nbeerbower_Mistral-Small-Drummer-22B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Small-Drummer-22B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Small-Drummer-22B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Small-Drummer-22B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Small-Drummer-22B | 53b21ece0c64ffc8aba81f294ad19e2c06e9852c | 29.74388 | other | 11 | 22 | true | false | false | false | 1.612722 | 0.633129 | 63.312899 | 0.57932 | 40.12177 | 0.18429 | 18.429003 | 0.343121 | 12.416107 | 0.406365 | 9.795573 | 0.409491 | 34.387928 | false | false | 2024-09-26 | 2024-10-01 | 1 | nbeerbower/Mistral-Small-Drummer-22B (Merge) |
nbeerbower_Mistral-Small-Gutenberg-Doppel-22B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Mistral-Small-Gutenberg-Doppel-22B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Mistral-Small-Gutenberg-Doppel-22B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Mistral-Small-Gutenberg-Doppel-22B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Mistral-Small-Gutenberg-Doppel-22B | d8091aad5f882b714321e4d51f504cc61996ee67 | 27.858747 | other | 10 | 22 | true | false | false | false | 1.588603 | 0.489323 | 48.932277 | 0.585893 | 40.931345 | 0.21148 | 21.148036 | 0.346477 | 12.863535 | 0.397063 | 8.566146 | 0.4124 | 34.711141 | false | false | 2024-09-25 | 2024-09-25 | 1 | nbeerbower/Mistral-Small-Gutenberg-Doppel-22B (Merge) |
nbeerbower_Nemo-Loony-12B-experimental_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Nemo-Loony-12B-experimental" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Nemo-Loony-12B-experimental</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Nemo-Loony-12B-experimental-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Nemo-Loony-12B-experimental | 7b06f30502a9b58c028ac1079e1b3d2988b76866 | 10.431803 | 0 | 12 | false | false | false | true | 1.237582 | 0.373444 | 37.344357 | 0.382222 | 12.974588 | 0.01284 | 1.283988 | 0.270134 | 2.684564 | 0.334063 | 1.757812 | 0.15891 | 6.545508 | false | false | 2024-11-26 | 2024-11-26 | 1 | nbeerbower/Nemo-Loony-12B-experimental (Merge) |
|
nbeerbower_Qwen2.5-Gutenberg-Doppel-14B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Qwen2.5-Gutenberg-Doppel-14B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Qwen2.5-Gutenberg-Doppel-14B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Qwen2.5-Gutenberg-Doppel-14B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Qwen2.5-Gutenberg-Doppel-14B | 11a5060f9e7315ea07241106f086ac4694dded60 | 32.302115 | apache-2.0 | 11 | 14 | true | false | false | true | 1.690612 | 0.809083 | 80.908323 | 0.638174 | 48.238909 | 0 | 0 | 0.333054 | 11.073826 | 0.410063 | 10.024479 | 0.492104 | 43.567154 | false | false | 2024-11-11 | 2024-11-11 | 1 | nbeerbower/Qwen2.5-Gutenberg-Doppel-14B (Merge) |
nbeerbower_SmolNemo-12B-FFT-experimental_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/SmolNemo-12B-FFT-experimental" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/SmolNemo-12B-FFT-experimental</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__SmolNemo-12B-FFT-experimental-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/SmolNemo-12B-FFT-experimental | d8d7a90ae9b9cb79cdc0912a685c3cb8d7a25560 | 8.320055 | apache-2.0 | 0 | 12 | true | false | false | true | 1.225415 | 0.334801 | 33.480055 | 0.333609 | 6.542439 | 0.002266 | 0.226586 | 0.260067 | 1.342282 | 0.384698 | 5.920573 | 0.121676 | 2.408392 | false | false | 2024-11-25 | 2024-11-26 | 1 | nbeerbower/SmolNemo-12B-FFT-experimental (Merge) |
nbeerbower_Stella-mistral-nemo-12B-v2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/Stella-mistral-nemo-12B-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/Stella-mistral-nemo-12B-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__Stella-mistral-nemo-12B-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/Stella-mistral-nemo-12B-v2 | b81bab28f7dcb25a0aa0fe4dcf957f3083ee6b43 | 22.430369 | 3 | 12 | false | false | false | false | 1.740872 | 0.327431 | 32.743122 | 0.548375 | 35.364516 | 0.112538 | 11.253776 | 0.332215 | 10.961969 | 0.430396 | 14.432812 | 0.368434 | 29.82602 | false | false | 2024-09-07 | 2024-09-14 | 1 | nbeerbower/Stella-mistral-nemo-12B-v2 (Merge) |
|
nbeerbower_gemma2-gutenberg-27B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/gemma2-gutenberg-27B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/gemma2-gutenberg-27B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__gemma2-gutenberg-27B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/gemma2-gutenberg-27B | d4febe52e8b7b13a98126dbf1716ed1329f48922 | 10.108961 | gemma | 4 | 27 | true | false | false | false | 7.695458 | 0.294708 | 29.470804 | 0.379657 | 13.091525 | 0 | 0 | 0.272651 | 3.020134 | 0.372729 | 4.157813 | 0.198221 | 10.91349 | false | false | 2024-09-09 | 2024-09-23 | 1 | nbeerbower/gemma2-gutenberg-27B (Merge) |
nbeerbower_gemma2-gutenberg-9B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co/nbeerbower/gemma2-gutenberg-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">nbeerbower/gemma2-gutenberg-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/nbeerbower__gemma2-gutenberg-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | nbeerbower/gemma2-gutenberg-9B | ebdab2d41f257fc9e7c858498653644d13386ce5 | 22.649257 | gemma | 12 | 9 | true | false | false | false | 2.809609 | 0.279595 | 27.959481 | 0.59509 | 42.355611 | 0.016616 | 1.661631 | 0.338087 | 11.744966 | 0.45951 | 16.705469 | 0.419215 | 35.468381 | false | false | 2024-07-14 | 2024-08-03 | 1 | nbeerbower/gemma2-gutenberg-9B (Merge) |