eval_name
stringlengths
12
111
Precision
stringclasses
3 values
Type
stringclasses
6 values
T
stringclasses
6 values
Weight type
stringclasses
2 values
Architecture
stringclasses
52 values
Model
stringlengths
355
689
fullname
stringlengths
4
102
Model sha
stringlengths
0
40
Average ⬆️
float64
1.03
52
Hub License
stringclasses
26 values
Hub ❤️
int64
0
5.9k
#Params (B)
int64
-1
140
Available on the hub
bool
2 classes
MoE
bool
2 classes
Flagged
bool
2 classes
Chat Template
bool
2 classes
CO₂ cost (kg)
float64
0.03
107
IFEval Raw
float64
0
0.9
IFEval
float64
0
90
BBH Raw
float64
0.27
0.75
BBH
float64
0.81
63.5
MATH Lvl 5 Raw
float64
0
0.51
MATH Lvl 5
float64
0
50.7
GPQA Raw
float64
0.22
0.44
GPQA
float64
0
24.9
MUSR Raw
float64
0.29
0.6
MUSR
float64
0
38.5
MMLU-PRO Raw
float64
0.1
0.73
MMLU-PRO
float64
0
70
Merged
bool
2 classes
Official Providers
bool
2 classes
Upload To Hub Date
stringclasses
424 values
Submission Date
stringclasses
169 values
Generation
int64
0
10
Base Model
stringlengths
4
102
awnr_Mistral-7B-v0.1-signtensors-1-over-2_bfloat16
bfloat16
🟢 pretrained
🟢
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/awnr/Mistral-7B-v0.1-signtensors-1-over-2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">awnr/Mistral-7B-v0.1-signtensors-1-over-2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/awnr__Mistral-7B-v0.1-signtensors-1-over-2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
awnr/Mistral-7B-v0.1-signtensors-1-over-2
9575327242f8539eac59b6d788beccf54a6f9414
14.257194
apache-2.0
2
7
true
false
false
false
2.567058
0.217922
21.792178
0.442288
22.400153
0.02719
2.719033
0.307047
7.606264
0.400604
8.808854
0.29995
22.216681
false
false
2024-06-27
2024-07-30
0
awnr/Mistral-7B-v0.1-signtensors-1-over-2
awnr_Mistral-7B-v0.1-signtensors-1-over-4_bfloat16
bfloat16
🟢 pretrained
🟢
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/awnr/Mistral-7B-v0.1-signtensors-1-over-4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">awnr/Mistral-7B-v0.1-signtensors-1-over-4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/awnr__Mistral-7B-v0.1-signtensors-1-over-4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
awnr/Mistral-7B-v0.1-signtensors-1-over-4
b288ab9d8adfd2963a44a7935bb47649f55bcbee
8.709433
apache-2.0
1
7
true
false
false
false
1.28233
0.213301
21.330071
0.350709
9.227694
0.022659
2.265861
0.270134
2.684564
0.346031
2.18724
0.231051
14.56117
false
false
2024-07-29
2024-07-29
0
awnr/Mistral-7B-v0.1-signtensors-1-over-4
awnr_Mistral-7B-v0.1-signtensors-3-over-8_bfloat16
bfloat16
🟢 pretrained
🟢
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/awnr/Mistral-7B-v0.1-signtensors-3-over-8" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">awnr/Mistral-7B-v0.1-signtensors-3-over-8</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/awnr__Mistral-7B-v0.1-signtensors-3-over-8-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
awnr/Mistral-7B-v0.1-signtensors-3-over-8
fa368f705ace05da2fef25c030fe740cf1fef176
13.725352
apache-2.0
1
7
true
false
false
false
1.270991
0.239429
23.942916
0.429994
20.435231
0.027946
2.794562
0.303691
7.158837
0.38175
5.785417
0.300116
22.235151
false
false
2024-07-29
2024-07-29
0
awnr/Mistral-7B-v0.1-signtensors-3-over-8
awnr_Mistral-7B-v0.1-signtensors-5-over-16_bfloat16
bfloat16
🟢 pretrained
🟢
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/awnr/Mistral-7B-v0.1-signtensors-5-over-16" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">awnr/Mistral-7B-v0.1-signtensors-5-over-16</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/awnr__Mistral-7B-v0.1-signtensors-5-over-16-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
awnr/Mistral-7B-v0.1-signtensors-5-over-16
5ea13b3d0723237889e1512bc70dae72f71884d1
12.158648
apache-2.0
1
7
true
false
false
false
0.649735
0.211827
21.182684
0.412415
17.543031
0.021903
2.190332
0.28104
4.138702
0.368604
6.142188
0.295795
21.75495
false
false
2024-07-29
2024-07-29
0
awnr/Mistral-7B-v0.1-signtensors-5-over-16
awnr_Mistral-7B-v0.1-signtensors-7-over-16_bfloat16
bfloat16
🟢 pretrained
🟢
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/awnr/Mistral-7B-v0.1-signtensors-7-over-16" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">awnr/Mistral-7B-v0.1-signtensors-7-over-16</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/awnr__Mistral-7B-v0.1-signtensors-7-over-16-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
awnr/Mistral-7B-v0.1-signtensors-7-over-16
0e1f2cb0a81c38fc6c567d9c007883ab62fae266
14.146
apache-2.0
1
7
true
false
false
false
1.296433
0.229363
22.936254
0.431582
21.040437
0.032477
3.247734
0.303691
7.158837
0.395208
7.934375
0.303025
22.558363
false
false
2024-07-29
2024-07-29
0
awnr/Mistral-7B-v0.1-signtensors-7-over-16
aws-prototyping_MegaBeam-Mistral-7B-512k_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/aws-prototyping/MegaBeam-Mistral-7B-512k" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">aws-prototyping/MegaBeam-Mistral-7B-512k</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/aws-prototyping__MegaBeam-Mistral-7B-512k-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
aws-prototyping/MegaBeam-Mistral-7B-512k
3e3b8c4b933650eed81ede7c4395df943d2a0796
17.59507
apache-2.0
44
7
true
false
false
true
0.647188
0.597259
59.725861
0.366234
12.361178
0.029456
2.945619
0.282718
4.362416
0.399365
8.520573
0.258893
17.654772
false
false
2024-07-30
2024-10-07
0
aws-prototyping/MegaBeam-Mistral-7B-512k
axolotl-ai-co_romulus-mistral-nemo-12b-simpo_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/axolotl-ai-co/romulus-mistral-nemo-12b-simpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">axolotl-ai-co/romulus-mistral-nemo-12b-simpo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/axolotl-ai-co__romulus-mistral-nemo-12b-simpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
axolotl-ai-co/romulus-mistral-nemo-12b-simpo
15fd3ffa46c1ea51aa5d26a1da24214e324d7cf2
23.426338
apache-2.0
15
12
true
false
false
true
2.715152
0.607925
60.792475
0.539506
34.642401
0.009063
0.906344
0.278523
3.803132
0.423302
12.979427
0.346908
27.434249
false
false
2024-07-24
2024-09-21
1
Removed
beomi_gemma-mling-7b_bfloat16
bfloat16
🟩 continuously pretrained
🟩
Original
GemmaForCausalLM
<a target="_blank" href="https://huggingface.co/beomi/gemma-mling-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">beomi/gemma-mling-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/beomi__gemma-mling-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
beomi/gemma-mling-7b
3f442e28bd50db6c438ce2a15b3a003532babba0
11.253704
other
14
8
true
false
false
false
1.643506
0.202909
20.290939
0.406759
17.631391
0.046073
4.607251
0.25
0
0.375854
6.848437
0.263298
18.144208
false
false
2024-04-15
2024-07-17
0
beomi/gemma-mling-7b
beowolx_CodeNinja-1.0-OpenChat-7B_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/beowolx/CodeNinja-1.0-OpenChat-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">beowolx/CodeNinja-1.0-OpenChat-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/beowolx__CodeNinja-1.0-OpenChat-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
beowolx/CodeNinja-1.0-OpenChat-7B
9934c04c767e6ae0f792712a060f02915391d4ec
20.347389
mit
105
7
true
false
false
true
0.636073
0.544677
54.467701
0.444134
21.713423
0.060423
6.042296
0.294463
5.928412
0.424323
11.540365
0.301529
22.392139
false
false
2023-12-20
2024-07-30
0
beowolx/CodeNinja-1.0-OpenChat-7B
berkeley-nest_Starling-LM-7B-alpha_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">berkeley-nest/Starling-LM-7B-alpha</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/berkeley-nest__Starling-LM-7B-alpha-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
berkeley-nest/Starling-LM-7B-alpha
1dddf3b95bc1391f6307299eb1c162c194bde9bd
20.826773
apache-2.0
555
7
true
false
false
true
0.551629
0.548049
54.804918
0.444007
21.954028
0.083082
8.308157
0.29698
6.263982
0.41201
9.501302
0.317154
24.128251
false
true
2023-11-25
2024-06-12
0
berkeley-nest/Starling-LM-7B-alpha
bigcode_starcoder2-15b_bfloat16
bfloat16
🟢 pretrained
🟢
Original
Starcoder2ForCausalLM
<a target="_blank" href="https://huggingface.co/bigcode/starcoder2-15b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigcode/starcoder2-15b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigcode__starcoder2-15b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigcode/starcoder2-15b
46d44742909c03ac8cee08eb03fdebce02e193ec
12.551764
bigcode-openrail-m
570
15
true
false
false
false
35.044548
0.278022
27.802231
0.444796
20.373541
0.060423
6.042296
0.27349
3.131991
0.350094
2.928385
0.235289
15.032137
false
true
2024-02-20
2024-06-09
0
bigcode/starcoder2-15b
bigcode_starcoder2-3b_bfloat16
bfloat16
🟢 pretrained
🟢
Original
Starcoder2ForCausalLM
<a target="_blank" href="https://huggingface.co/bigcode/starcoder2-3b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigcode/starcoder2-3b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigcode__starcoder2-3b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigcode/starcoder2-3b
733247c55e3f73af49ce8e9c7949bf14af205928
6.53656
bigcode-openrail-m
152
3
true
false
false
false
0.446629
0.203708
20.370838
0.350871
8.909299
0.01435
1.435045
0.244128
0
0.343458
1.432292
0.163647
7.071882
false
true
2023-11-29
2024-06-09
0
bigcode/starcoder2-3b
bigcode_starcoder2-7b_bfloat16
bfloat16
🟢 pretrained
🟢
Original
Starcoder2ForCausalLM
<a target="_blank" href="https://huggingface.co/bigcode/starcoder2-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigcode/starcoder2-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigcode__starcoder2-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigcode/starcoder2-7b
a3d33687b51284b528abeb17830776ffd24892a9
8.255674
bigcode-openrail-m
161
7
true
false
false
false
0.506401
0.220919
22.091938
0.366099
11.39511
0.028701
2.870091
0.251678
0.223714
0.379333
5.816667
0.164229
7.136525
false
true
2024-02-20
2024-06-09
0
bigcode/starcoder2-7b
bigscience_bloom-1b1_bfloat16
bfloat16
🟢 pretrained
🟢
Original
BloomForCausalLM
<a target="_blank" href="https://huggingface.co/bigscience/bloom-1b1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigscience/bloom-1b1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigscience__bloom-1b1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigscience/bloom-1b1
eb3dd7399312f5f94fd13f41d2f318117d3eb1e4
3.962215
bigscience-bloom-rail-1.0
62
1
true
false
false
false
0.717021
0.137338
13.733782
0.310728
4.042705
0.001511
0.151057
0.259228
1.230425
0.37
3.416667
0.110788
1.198655
false
true
2022-05-19
2024-06-13
0
bigscience/bloom-1b1
bigscience_bloom-1b7_bfloat16
bfloat16
🟢 pretrained
🟢
Original
BloomForCausalLM
<a target="_blank" href="https://huggingface.co/bigscience/bloom-1b7" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigscience/bloom-1b7</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigscience__bloom-1b7-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigscience/bloom-1b7
cc72a88036c2fb937d65efeacc57a0c2ef5d6fe5
3.971226
bigscience-bloom-rail-1.0
118
1
true
false
false
false
0.81836
0.10439
10.438969
0.314055
4.397453
0.000755
0.075529
0.258389
1.118568
0.388573
6.838281
0.108627
0.958555
false
true
2022-05-19
2024-06-13
0
bigscience/bloom-1b7
bigscience_bloom-3b_bfloat16
bfloat16
🟢 pretrained
🟢
Original
BloomForCausalLM
<a target="_blank" href="https://huggingface.co/bigscience/bloom-3b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigscience/bloom-3b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigscience__bloom-3b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigscience/bloom-3b
52bc5b43010b4844513826b8be3f78c7344c37d7
4.262013
bigscience-bloom-rail-1.0
88
3
true
false
false
false
0.996056
0.127096
12.709611
0.306292
3.420098
0.000755
0.075529
0.239933
0
0.398063
7.891146
0.113281
1.475694
false
true
2022-05-19
2024-06-13
0
bigscience/bloom-3b
bigscience_bloom-560m_bfloat16
bfloat16
🟢 pretrained
🟢
Original
BloomForCausalLM
<a target="_blank" href="https://huggingface.co/bigscience/bloom-560m" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigscience/bloom-560m</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigscience__bloom-560m-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigscience/bloom-560m
ac2ae5fab2ce3f9f40dc79b5ca9f637430d24971
3.456891
bigscience-bloom-rail-1.0
347
0
true
false
false
false
0.762716
0.062024
6.202432
0.302595
2.885364
0.000755
0.075529
0.261745
1.565996
0.403083
8.185417
0.116439
1.826611
false
true
2022-05-19
2024-06-13
0
bigscience/bloom-560m
bigscience_bloom-7b1_float16
float16
🟢 pretrained
🟢
Original
BloomForCausalLM
<a target="_blank" href="https://huggingface.co/bigscience/bloom-7b1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bigscience/bloom-7b1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bigscience__bloom-7b1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bigscience/bloom-7b1
6232703e399354503377bf59dfbb8397fd569e4a
3.707393
bigscience-bloom-rail-1.0
199
7
true
false
false
false
1.005775
0.132217
13.221696
0.311372
4.038809
0
0
0.264262
1.901566
0.348698
1.920573
0.110455
1.161717
false
true
2022-05-19
2024-06-13
0
bigscience/bloom-7b1
bosonai_Higgs-Llama-3-70B_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/bosonai/Higgs-Llama-3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bosonai/Higgs-Llama-3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bosonai__Higgs-Llama-3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bosonai/Higgs-Llama-3-70B
b2c7540768046dfdae7a0cb846a7da6c41d826b1
32.216234
other
215
70
true
false
false
true
13.726847
0.556068
55.60679
0.625766
45.897406
0.173716
17.371601
0.366611
15.548098
0.447083
15.51875
0.490193
43.354758
false
false
2024-06-05
2024-08-30
1
meta-llama/Meta-Llama-3-70B
brgx53_3Bgeneral-ECE-PRYMMAL-Martial_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Phi3ForCausalLM
<a target="_blank" href="https://huggingface.co/brgx53/3Bgeneral-ECE-PRYMMAL-Martial" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">brgx53/3Bgeneral-ECE-PRYMMAL-Martial</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/brgx53__3Bgeneral-ECE-PRYMMAL-Martial-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
brgx53/3Bgeneral-ECE-PRYMMAL-Martial
78ee3bde02df349ee7161f9c2a5b36161c294009
23.167894
apache-2.0
0
3
true
false
false
false
0.653652
0.328931
32.893057
0.545801
36.673582
0.124622
12.462236
0.324664
9.955257
0.437281
14.426823
0.393368
32.59641
true
false
2024-10-23
2024-10-23
1
brgx53/3Bgeneral-ECE-PRYMMAL-Martial (Merge)
brgx53_3Bgeneralv2-ECE-PRYMMAL-Martial_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/brgx53/3Bgeneralv2-ECE-PRYMMAL-Martial" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">brgx53/3Bgeneralv2-ECE-PRYMMAL-Martial</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/brgx53__3Bgeneralv2-ECE-PRYMMAL-Martial-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
brgx53/3Bgeneralv2-ECE-PRYMMAL-Martial
8525f801c47b2bce2ca4dad360ce71b2cb6b370b
30.777462
apache-2.0
0
3
true
false
false
false
1.341162
0.567708
56.770813
0.56072
37.250633
0.307402
30.740181
0.311242
8.165548
0.435635
12.78776
0.450549
38.949837
true
false
2024-11-08
2024-11-08
1
brgx53/3Bgeneralv2-ECE-PRYMMAL-Martial (Merge)
brgx53_3Blareneg-ECE-PRYMMAL-Martial_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Phi3ForCausalLM
<a target="_blank" href="https://huggingface.co/brgx53/3Blareneg-ECE-PRYMMAL-Martial" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">brgx53/3Blareneg-ECE-PRYMMAL-Martial</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/brgx53__3Blareneg-ECE-PRYMMAL-Martial-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
brgx53/3Blareneg-ECE-PRYMMAL-Martial
abac4757125a66a427fb82751bf171dabaea3458
21.33386
apache-2.0
0
3
true
false
false
false
0.807534
0.287639
28.763902
0.535846
35.452586
0.035498
3.549849
0.334732
11.297539
0.442896
15.428646
0.401596
33.510638
true
false
2024-10-23
2024-10-23
1
brgx53/3Blareneg-ECE-PRYMMAL-Martial (Merge)
brgx53_3Blarenegv2-ECE-PRYMMAL-Martial_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/brgx53/3Blarenegv2-ECE-PRYMMAL-Martial" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">brgx53/3Blarenegv2-ECE-PRYMMAL-Martial</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/brgx53__3Blarenegv2-ECE-PRYMMAL-Martial-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
brgx53/3Blarenegv2-ECE-PRYMMAL-Martial
304038fc2b2527e31c738f9091206253a0d40f6c
30.752066
apache-2.0
0
7
true
false
false
false
0.686896
0.566184
56.618439
0.56072
37.250633
0.307402
30.740181
0.311242
8.165548
0.435635
12.78776
0.450549
38.949837
true
false
2024-11-08
2024-11-08
1
brgx53/3Blarenegv2-ECE-PRYMMAL-Martial (Merge)
bunnycore_Best-Mix-Llama-3.1-8B_float16
float16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Best-Mix-Llama-3.1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Best-Mix-Llama-3.1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Best-Mix-Llama-3.1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Best-Mix-Llama-3.1-8B
4bde0e60ac20d6944b1fbdfb3456efea8ba59ae9
8.624959
apache-2.0
0
8
true
false
false
false
0.905003
0.206706
20.670598
0.343178
7.255276
0.14426
14.425982
0.265101
2.013423
0.292854
1.106771
0.156499
6.277704
true
false
2024-10-10
2024-10-10
0
bunnycore/Best-Mix-Llama-3.1-8B
bunnycore_CyberCore-Qwen-2.1-7B_float16
float16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/CyberCore-Qwen-2.1-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/CyberCore-Qwen-2.1-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__CyberCore-Qwen-2.1-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/CyberCore-Qwen-2.1-7B
98e69ba1cd70444b90178e1253e904d1892593c8
27.24554
2
7
false
false
false
true
0.672684
0.576576
57.657571
0.557209
36.966533
0.134441
13.444109
0.307886
7.718121
0.41449
9.411198
0.444481
38.275709
false
false
2024-11-21
2024-11-23
1
bunnycore/CyberCore-Qwen-2.1-7B (Merge)
bunnycore_HyperLlama-3.1-8B_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/HyperLlama-3.1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/HyperLlama-3.1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__HyperLlama-3.1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/HyperLlama-3.1-8B
659b18ffaee2c1e8dbe8a9a56a44502325d71696
28.398623
apache-2.0
4
8
true
false
false
true
0.894545
0.788301
78.83006
0.510339
29.806656
0.179758
17.975831
0.286913
4.9217
0.382927
7.932552
0.378324
30.924941
true
false
2024-09-04
2024-09-05
0
bunnycore/HyperLlama-3.1-8B
bunnycore_Llama-3.1-8B-TitanFusion-Mix_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.1-8B-TitanFusion-Mix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.1-8B-TitanFusion-Mix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.1-8B-TitanFusion-Mix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.1-8B-TitanFusion-Mix
9eb89de7df048276ccbc4405ce4f005f9185f82e
24.961895
2
8
false
false
false
false
0.933085
0.492495
49.249547
0.575596
39.535483
0.125378
12.537764
0.295302
6.040268
0.431698
12.46224
0.369515
29.94607
false
false
2024-09-23
2024-09-23
1
bunnycore/Llama-3.1-8B-TitanFusion-Mix (Merge)
bunnycore_Llama-3.1-8B-TitanFusion-v3_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.1-8B-TitanFusion-v3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.1-8B-TitanFusion-v3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.1-8B-TitanFusion-v3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.1-8B-TitanFusion-v3
ea8269ac3b2e9c0dc855a9089251ebdb273ada16
24.231721
2
8
false
false
false
false
0.887824
0.480955
48.095498
0.526211
32.072941
0.142749
14.274924
0.308725
7.829978
0.430208
11.942708
0.380568
31.174276
false
false
2024-09-22
2024-09-22
1
bunnycore/Llama-3.1-8B-TitanFusion-v3 (Merge)
bunnycore_Llama-3.2-3B-All-Mix_float16
float16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.2-3B-All-Mix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.2-3B-All-Mix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.2-3B-All-Mix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.2-3B-All-Mix
adacdd571c4073990ecf05a23277793e9e5f0410
22.416478
1
3
false
false
false
true
0.740309
0.722605
72.260491
0.450834
22.516311
0.11858
11.858006
0.262584
1.677852
0.328698
2.18724
0.315991
23.998966
false
false
2024-10-20
2024-10-20
1
bunnycore/Llama-3.2-3B-All-Mix (Merge)
bunnycore_Llama-3.2-3B-Booval_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.2-3B-Booval" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.2-3B-Booval</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.2-3B-Booval-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.2-3B-Booval
d7f3449f89fa86d8e2c411aa4ca10ad552a62803
21.3011
2
3
false
false
false
true
0.655241
0.666926
66.692598
0.451439
22.515991
0.111027
11.102719
0.266779
2.237136
0.339427
2.395052
0.305768
22.863106
false
false
2024-10-27
2024-10-28
1
bunnycore/Llama-3.2-3B-Booval (Merge)
bunnycore_Llama-3.2-3B-Long-Think_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.2-3B-Long-Think" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.2-3B-Long-Think</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.2-3B-Long-Think-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.2-3B-Long-Think
a8522bfc03657b41b0541b164a98ddff302a6fd2
19.549052
1
3
false
false
false
true
1.381029
0.54735
54.734992
0.461039
24.226803
0.129154
12.915408
0.260906
1.454139
0.339552
1.210677
0.304771
22.75229
false
false
2024-10-24
2024-10-24
1
bunnycore/Llama-3.2-3B-Long-Think (Merge)
bunnycore_Llama-3.2-3B-Mix-Skill_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.2-3B-Mix-Skill" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.2-3B-Mix-Skill</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.2-3B-Mix-Skill-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.2-3B-Mix-Skill
d07d6e733aaeaf48cb6616228d00104b05b52afd
21.399687
1
3
false
false
false
true
0.685067
0.640423
64.042297
0.458184
23.784247
0.126888
12.688822
0.261745
1.565996
0.339615
2.751823
0.312084
23.564938
false
false
2024-10-24
2024-10-24
1
bunnycore/Llama-3.2-3B-Mix-Skill (Merge)
bunnycore_Llama-3.2-3B-ProdigyPlus_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlus" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.2-3B-ProdigyPlus</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.2-3B-ProdigyPlus-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.2-3B-ProdigyPlus
799f7669701ecf27f4c3e29998dd839b4d54c408
16.079069
2
3
false
false
false
true
0.71394
0.40152
40.152019
0.439228
20.622989
0.098943
9.89426
0.268456
2.46085
0.358
3.15
0.281749
20.194297
false
false
2024-10-25
2024-10-25
1
bunnycore/Llama-3.2-3B-ProdigyPlus (Merge)
bunnycore_Llama-3.2-3B-ProdigyPlusPlus_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Llama-3.2-3B-ProdigyPlusPlus" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Llama-3.2-3B-ProdigyPlusPlus</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Llama-3.2-3B-ProdigyPlusPlus-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Llama-3.2-3B-ProdigyPlusPlus
512865708a7ec9754997fb404b1ffc0752b099d7
6.443826
0
3
false
false
false
true
0.671576
0.164516
16.451571
0.368993
11.561978
0.029456
2.945619
0.253356
0.447427
0.354125
1.698958
0.150017
5.557402
false
false
2024-10-28
2024-10-28
1
bunnycore/Llama-3.2-3B-ProdigyPlusPlus (Merge)
bunnycore_Phi-3.5-mini-TitanFusion-0.1_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Phi3ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Phi-3.5-mini-TitanFusion-0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Phi-3.5-mini-TitanFusion-0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Phi-3.5-mini-TitanFusion-0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Phi-3.5-mini-TitanFusion-0.1
72939b8b75e23b22b1758bb05a842e5834f75d96
25.392388
0
3
false
false
false
true
0.796885
0.522795
52.279507
0.537373
35.446219
0.067976
6.797583
0.331376
10.850112
0.445313
15.797396
0.380652
31.183511
false
false
2024-10-13
2024-10-13
1
bunnycore/Phi-3.5-mini-TitanFusion-0.1 (Merge)
bunnycore_Qandora-2.5-7B-Creative_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Qandora-2.5-7B-Creative" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Qandora-2.5-7B-Creative</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Qandora-2.5-7B-Creative-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Qandora-2.5-7B-Creative
fdb174364d4a4f323ed1cb01219ac4d87708219d
30.931131
1
7
false
false
false
true
0.709795
0.680315
68.03149
0.554176
36.424652
0.23565
23.564955
0.310403
8.053691
0.421188
10.848438
0.447972
38.663564
false
false
2024-11-20
2024-11-20
1
bunnycore/Qandora-2.5-7B-Creative (Merge)
bunnycore_QandoraExp-7B_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/QandoraExp-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/QandoraExp-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__QandoraExp-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/QandoraExp-7B
74906d5518c7feb7df9b168763dabc1b0167942f
28.51072
1
7
false
false
false
true
0.670528
0.750906
75.090648
0.547796
35.924742
0.009063
0.906344
0.310403
8.053691
0.431208
13.201042
0.440991
37.887855
false
false
2024-11-11
2024-11-11
1
bunnycore/QandoraExp-7B (Merge)
bunnycore_QandoraExp-7B-Persona_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/QandoraExp-7B-Persona" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/QandoraExp-7B-Persona</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__QandoraExp-7B-Persona-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/QandoraExp-7B-Persona
21bd6c2e270358b70f9af98bcccd6ec9c2cfce88
29.188506
2
7
false
false
false
true
0.687971
0.624686
62.468583
0.555834
36.832709
0.160121
16.012085
0.314597
8.612975
0.437156
13.344531
0.440741
37.860151
false
false
2024-11-12
2024-11-12
1
bunnycore/QandoraExp-7B-Persona (Merge)
bunnycore_QandoraExp-7B-v2_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/QandoraExp-7B-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/QandoraExp-7B-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__QandoraExp-7B-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/QandoraExp-7B-v2
017594240f9b3c4262e23de6d550453a1a3d5540
23.274654
1
7
false
false
false
true
0.693649
0.560689
56.068897
0.544486
34.944967
0
0
0.302852
7.04698
0.404542
9.267708
0.390874
32.319371
false
false
2024-11-12
2024-11-12
1
bunnycore/QandoraExp-7B-v2 (Merge)
bunnycore_Qwen2.5-3B-RP-Mix_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Qwen2.5-3B-RP-Mix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Qwen2.5-3B-RP-Mix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Qwen2.5-3B-RP-Mix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Qwen2.5-3B-RP-Mix
0e8f3b56f9270fdcdd4badfd7b925dc8fc4902c7
23.377815
2
3
false
false
false
true
0.919701
0.572054
57.205437
0.489438
28.305923
0.087613
8.761329
0.27349
3.131991
0.428448
12.55599
0.372756
30.30622
false
false
2024-10-22
2024-10-22
1
bunnycore/Qwen2.5-3B-RP-Mix (Merge)
bunnycore_Qwen2.5-7B-CyberRombos_float16
float16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Qwen2.5-7B-CyberRombos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Qwen2.5-7B-CyberRombos</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Qwen2.5-7B-CyberRombos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Qwen2.5-7B-CyberRombos
dfd4d30fc6956ffecb9fb3c59fad51875552f7f9
27.692747
2
7
false
false
false
true
0.710105
0.751831
75.18307
0.546496
35.884025
0.000755
0.075529
0.30453
7.270694
0.412542
10.067708
0.439079
37.675458
false
false
2024-11-04
2024-11-05
1
bunnycore/Qwen2.5-7B-CyberRombos (Merge)
bunnycore_Qwen2.5-7B-Instruct-Fusion_float16
float16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Qwen2.5-7B-Instruct-Fusion" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Qwen2.5-7B-Instruct-Fusion</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Qwen2.5-7B-Instruct-Fusion-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Qwen2.5-7B-Instruct-Fusion
6313c0b3de799ab48720c3b828e322a77cf8d023
30.747252
3
7
false
false
false
true
0.66295
0.696202
69.620163
0.54919
36.179859
0.199396
19.939577
0.30453
7.270694
0.429719
12.948177
0.446725
38.525044
false
false
2024-10-31
2024-11-02
1
bunnycore/Qwen2.5-7B-Instruct-Fusion (Merge)
bunnycore_QwenMosaic-7B_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/QwenMosaic-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/QwenMosaic-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__QwenMosaic-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/QwenMosaic-7B
1eab0bbe701195ba26f60a284f74e3c6dfe5c139
25.308428
1
7
false
false
false
true
0.750287
0.581922
58.192152
0.556413
36.75052
0.084592
8.459215
0.260906
1.454139
0.416385
10.214844
0.431017
36.779699
false
false
2024-12-01
2024-12-02
1
bunnycore/QwenMosaic-7B (Merge)
bunnycore_SmolLM2-1.7-Persona_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/SmolLM2-1.7-Persona" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/SmolLM2-1.7-Persona</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__SmolLM2-1.7-Persona-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/SmolLM2-1.7-Persona
ebeaa6f284c044bd54e3e66cc5458d974d92523e
14.25041
apache-2.0
0
1
true
false
false
true
0.331332
0.546525
54.652544
0.362321
11.203753
0.04003
4.003021
0.263423
1.789709
0.334125
3.032292
0.19739
10.821144
true
false
2024-11-15
2024-11-15
0
bunnycore/SmolLM2-1.7-Persona
bunnycore_SmolLM2-1.7B-roleplay-lora_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Adapter
?
<a target="_blank" href="https://huggingface.co/bunnycore/SmolLM2-1.7B-roleplay-lora" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/SmolLM2-1.7B-roleplay-lora</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__SmolLM2-1.7B-roleplay-lora-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/SmolLM2-1.7B-roleplay-lora
bbab860a4ffdd8e48f600192947ad3504bb0a944
14.252474
apache-2.0
0
3
true
false
false
true
0.700026
0.538208
53.820751
0.361034
10.907238
0.039275
3.927492
0.275168
3.355705
0.339458
2.765625
0.196642
10.738032
false
false
2024-11-15
2024-11-15
3
HuggingFaceTB/SmolLM2-1.7B-Instruct (Merge)
bunnycore_Tulu-3.1-8B-SuperNova_float16
float16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/bunnycore/Tulu-3.1-8B-SuperNova" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">bunnycore/Tulu-3.1-8B-SuperNova</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Tulu-3.1-8B-SuperNova-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
bunnycore/Tulu-3.1-8B-SuperNova
bbbfb910ca8d8f7ae35ecaf4824ad68713bf8d86
30.941023
3
8
false
false
false
true
0.693924
0.819375
81.937481
0.525412
32.499171
0.243202
24.320242
0.302013
6.935123
0.3935
8.6875
0.3814
31.266622
false
false
2024-11-22
2024-11-23
1
bunnycore/Tulu-3.1-8B-SuperNova (Merge)
byroneverson_Mistral-Small-Instruct-2409-abliterated_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/byroneverson/Mistral-Small-Instruct-2409-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">byroneverson/Mistral-Small-Instruct-2409-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/byroneverson__Mistral-Small-Instruct-2409-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
byroneverson/Mistral-Small-Instruct-2409-abliterated
5e24aaef2a37f9cb69f70ae9fe714f9d9599fd6e
27.169391
other
10
22
true
false
false
true
1.402286
0.697076
69.707598
0.523786
31.2557
0.149547
14.954683
0.333054
11.073826
0.369719
3.548177
0.392287
32.476359
false
false
2024-09-23
2024-10-13
1
mistralai/Mistral-Small-Instruct-2409
byroneverson_Yi-1.5-9B-Chat-16K-abliterated_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/byroneverson/Yi-1.5-9B-Chat-16K-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">byroneverson/Yi-1.5-9B-Chat-16K-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/byroneverson__Yi-1.5-9B-Chat-16K-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
byroneverson/Yi-1.5-9B-Chat-16K-abliterated
84a6eaa723633bbefc7cfac9c64bf0e0a4d39065
26.532728
apache-2.0
4
8
true
false
false
true
1.090103
0.552845
55.284534
0.528205
32.843259
0.116314
11.63142
0.312919
8.389262
0.473438
19.679687
0.382314
31.368203
false
false
2024-09-03
2024-09-03
1
01-ai/Yi-1.5-9B-Chat-16K
byroneverson_Yi-1.5-9B-Chat-abliterated_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/byroneverson/Yi-1.5-9B-Chat-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">byroneverson/Yi-1.5-9B-Chat-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/byroneverson__Yi-1.5-9B-Chat-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
byroneverson/Yi-1.5-9B-Chat-abliterated
4e26c200cdf2dc50dd50cdd9fe5b74887e9fa94a
25.300721
apache-2.0
2
8
true
false
false
true
0.844572
0.572329
57.23292
0.540122
34.352187
0.108006
10.800604
0.291946
5.592841
0.438865
13.658073
0.371509
30.167701
false
false
2024-09-04
2024-09-17
1
01-ai/Yi-1.5-9B-Chat
c10x_Q-Pluse_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/c10x/Q-Pluse" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">c10x/Q-Pluse</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/c10x__Q-Pluse-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
c10x/Q-Pluse
3.634371
0
7
false
false
false
true
1.311691
0.112283
11.228319
0.287511
1.947945
0
0
0.246644
0
0.393812
7.126563
0.113531
1.503398
false
false
2024-10-10
0
Removed
c10x_longthinker_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/c10x/longthinker" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">c10x/longthinker</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/c10x__longthinker-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
c10x/longthinker
e1bb4a2c2782ab52be7a8fa2e5905f08b7cfd464
19.686074
0
8
false
false
false
true
0.944774
0.360879
36.087913
0.492749
28.424737
0.169184
16.918429
0.264262
1.901566
0.390958
6.703125
0.352726
28.080674
false
false
2024-10-10
2024-10-10
1
c10x/longthinker (Merge)
carsenk_flippa-v6_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Adapter
?
<a target="_blank" href="https://huggingface.co/carsenk/flippa-v6" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">carsenk/flippa-v6</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/carsenk__flippa-v6-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
carsenk/flippa-v6
5206a32e0bd3067aef1ce90f5528ade7d866253f
20.738603
llama3.1
1
16
true
false
false
false
1.0648
0.343943
34.394296
0.504697
29.993501
0.138218
13.821752
0.292785
5.704698
0.408875
10.876042
0.366772
29.641327
false
false
2024-08-24
2024-08-24
2
meta-llama/Meta-Llama-3.1-8B
carsenk_phi3.5_mini_exp_825_uncensored_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/carsenk/phi3.5_mini_exp_825_uncensored" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">carsenk/phi3.5_mini_exp_825_uncensored</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/carsenk__phi3.5_mini_exp_825_uncensored-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
carsenk/phi3.5_mini_exp_825_uncensored
6b208dc3df02e0d5ef0c3fe5899f9f31618f2e94
3.466875
apache-2.0
2
3
true
false
false
true
0.487818
0.136414
13.64136
0.296473
1.827813
0
0
0.249161
0
0.364417
3.385417
0.11752
1.946661
false
false
2024-08-29
2024-08-29
2
microsoft/Phi-3.5-mini-instruct
cat-searcher_gemma-2-9b-it-sppo-iter-1_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Gemma2ForCausalLM
<a target="_blank" href="https://huggingface.co/cat-searcher/gemma-2-9b-it-sppo-iter-1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cat-searcher/gemma-2-9b-it-sppo-iter-1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cat-searcher__gemma-2-9b-it-sppo-iter-1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cat-searcher/gemma-2-9b-it-sppo-iter-1
b29a3a5cef93ee044e2297fcb40bd2976415e900
20.553948
0
9
false
false
false
true
2.768039
0.301477
30.147675
0.597187
41.676308
0
0
0.344799
12.639821
0.392667
7.15
0.385389
31.709885
false
false
2024-08-09
0
Removed
cat-searcher_gemma-2-9b-it-sppo-iter-1-evol-1_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Gemma2ForCausalLM
<a target="_blank" href="https://huggingface.co/cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cat-searcher__gemma-2-9b-it-sppo-iter-1-evol-1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1
c2d7b76786151aecfa5972a2a3e937feb2d2c48b
20.103006
0
9
false
false
false
true
2.787812
0.294183
29.418277
0.593937
41.10464
0
0
0.340604
12.080537
0.392573
6.904948
0.379987
31.109634
false
false
2024-08-09
0
Removed
cgato_TheSalt-L3-8b-v0.3.2_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cgato/TheSalt-L3-8b-v0.3.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cgato/TheSalt-L3-8b-v0.3.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cgato__TheSalt-L3-8b-v0.3.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cgato/TheSalt-L3-8b-v0.3.2
5cf08e2bf9590ebcd14ba021e113def28c65afa2
7.248832
cc-by-nc-4.0
1
8
true
false
false
true
0.940294
0.270503
27.050338
0.296797
2.612714
0.03852
3.851964
0.26594
2.12528
0.389625
6.303125
0.113946
1.549572
false
false
2024-06-18
2024-06-26
0
cgato/TheSalt-L3-8b-v0.3.2
chargoddard_prometheus-2-llama-3-8b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/chargoddard/prometheus-2-llama-3-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">chargoddard/prometheus-2-llama-3-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/chargoddard__prometheus-2-llama-3-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
chargoddard/prometheus-2-llama-3-8b
90a728ac98e5b4169f88ae4945e357cf45477568
19.281097
apache-2.0
2
8
true
false
false
true
0.945115
0.52889
52.889001
0.493114
27.803839
0.08006
8.006042
0.272651
3.020134
0.339583
0.78125
0.308677
23.186318
true
false
2024-05-26
2024-06-26
1
chargoddard/prometheus-2-llama-3-8b (Merge)
chujiezheng_Llama-3-Instruct-8B-SimPO-ExPO_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/chujiezheng__Llama-3-Instruct-8B-SimPO-ExPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
3fcaa9fe99691659eb197487e9a343f601bf63f2
21.972344
llama3
16
8
true
false
false
true
0.720481
0.643371
64.33707
0.476452
25.868282
0.005287
0.528701
0.286913
4.9217
0.39201
9.501302
0.340093
26.677009
false
false
2024-05-26
2024-06-26
0
chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
chujiezheng_Mistral7B-PairRM-SPPO-ExPO_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/chujiezheng/Mistral7B-PairRM-SPPO-ExPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">chujiezheng/Mistral7B-PairRM-SPPO-ExPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/chujiezheng__Mistral7B-PairRM-SPPO-ExPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
chujiezheng/Mistral7B-PairRM-SPPO-ExPO
d3e8342a63e5ae096f450f2467a92168db12768c
13.491268
apache-2.0
0
7
true
false
false
true
0.509034
0.367349
36.734863
0.388219
13.678636
0.010574
1.057402
0.276846
3.579418
0.405531
8.658073
0.255153
17.239214
false
false
2024-05-04
2024-09-21
0
chujiezheng/Mistral7B-PairRM-SPPO-ExPO
cloudyu_Llama-3-70Bx2-MOE_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/cloudyu/Llama-3-70Bx2-MOE" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cloudyu/Llama-3-70Bx2-MOE</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cloudyu__Llama-3-70Bx2-MOE-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cloudyu/Llama-3-70Bx2-MOE
b8bd85e8db8e4ec352b93441c92e0ae1334bf5a7
35.666465
llama3
1
126
true
true
false
false
21.539555
0.548249
54.824865
0.663623
51.422138
0.217523
21.752266
0.393456
19.127517
0.481188
20.848437
0.514212
46.023567
false
false
2024-05-20
2024-06-27
0
cloudyu/Llama-3-70Bx2-MOE
cloudyu_Mixtral_34Bx2_MoE_60B_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/cloudyu/Mixtral_34Bx2_MoE_60B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cloudyu/Mixtral_34Bx2_MoE_60B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cloudyu__Mixtral_34Bx2_MoE_60B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cloudyu/Mixtral_34Bx2_MoE_60B
d01642769ccc782e1db1fc26cb25097aecb98e23
27.598581
apache-2.0
111
60
true
true
false
false
7.332588
0.453777
45.377709
0.58697
41.209129
0.076284
7.628399
0.338087
11.744966
0.462521
17.781771
0.476646
41.849512
false
false
2024-01-05
2024-08-22
0
cloudyu/Mixtral_34Bx2_MoE_60B
cloudyu_Yi-34Bx2-MoE-60B-DPO_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/cloudyu/Yi-34Bx2-MoE-60B-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cloudyu/Yi-34Bx2-MoE-60B-DPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cloudyu__Yi-34Bx2-MoE-60B-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cloudyu/Yi-34Bx2-MoE-60B-DPO
5c2d31042229ee06246064100b781dd926cb0ffd
26.043502
apache-2.0
3
60
true
true
false
true
7.339247
0.531888
53.188761
0.516831
31.259298
0.070242
7.024169
0.322148
9.619687
0.437469
14.316927
0.46767
40.852172
false
false
2024-01-23
2024-08-06
0
cloudyu/Yi-34Bx2-MoE-60B-DPO
cluebbers_Llama-3.1-8B-paraphrase-type-generation-apty-ipo_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-ipo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-ipo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cluebbers__Llama-3.1-8B-paraphrase-type-generation-apty-ipo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-ipo
eb04613997875935cb667a517e518874bb716169
9.749648
apache-2.0
0
8
true
false
false
false
0.719831
0.132667
13.266688
0.380022
12.669478
0.006798
0.679758
0.263423
1.789709
0.433219
12.41901
0.259059
17.673242
false
false
2024-11-14
2024-11-15
1
cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-ipo (Merge)
cluebbers_Llama-3.1-8B-paraphrase-type-generation-apty-sigmoid_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-sigmoid" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-sigmoid</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cluebbers__Llama-3.1-8B-paraphrase-type-generation-apty-sigmoid-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-sigmoid
2c8b52e8db11a6ff57cccf890ee26688e858f9fb
9.743408
apache-2.0
0
8
true
false
false
false
0.72356
0.131842
13.18424
0.37889
12.757325
0.006798
0.679758
0.268456
2.46085
0.430552
12.01901
0.256233
17.359264
false
false
2024-11-15
2024-11-15
1
cluebbers/Llama-3.1-8B-paraphrase-type-generation-apty-sigmoid (Merge)
cluebbers_Llama-3.1-8B-paraphrase-type-generation-etpc_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/cluebbers/Llama-3.1-8B-paraphrase-type-generation-etpc" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cluebbers/Llama-3.1-8B-paraphrase-type-generation-etpc</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cluebbers__Llama-3.1-8B-paraphrase-type-generation-etpc-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cluebbers/Llama-3.1-8B-paraphrase-type-generation-etpc
a003a227aed5c1ad67cd4a653b13a0dd7acb7ed5
9.430026
apache-2.0
0
8
true
false
false
false
0.741527
0.120852
12.085156
0.378081
12.694579
0.004532
0.453172
0.265101
2.013423
0.431854
12.048437
0.255568
17.285387
false
false
2024-11-04
2024-11-15
1
cluebbers/Llama-3.1-8B-paraphrase-type-generation-etpc (Merge)
cognitivecomputations_dolphin-2.9-llama3-8b_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9-llama3-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9-llama3-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9-llama3-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9-llama3-8b
5aeb036f9215c558b483a654a8c6e1cc22e841bf
18.390285
other
422
8
true
false
false
true
0.73912
0.385034
38.503393
0.494992
27.858929
0.055891
5.589124
0.286913
4.9217
0.437531
13.791406
0.277094
19.677157
false
true
2024-04-20
2024-06-12
1
meta-llama/Meta-Llama-3-8B
cognitivecomputations_dolphin-2.9.1-llama-3-70b_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-70b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.1-llama-3-70b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.1-llama-3-70b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.1-llama-3-70b
31adf616c3c9176d147e0a62e9fedb7bf97678ac
23.444759
llama3
40
70
true
false
false
true
12.149088
0.376017
37.601675
0.520492
31.101152
0.056647
5.664653
0.308725
7.829978
0.497562
23.695312
0.412982
34.775783
false
true
2024-05-22
2024-06-27
1
meta-llama/Meta-Llama-3-70B
cognitivecomputations_dolphin-2.9.1-yi-1.5-34b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.1-yi-1.5-34b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.1-yi-1.5-34b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.1-yi-1.5-34b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.1-yi-1.5-34b
1ec522298a6935c881df6dc29d3669833bd8672d
27.904384
apache-2.0
34
34
true
false
false
true
2.992653
0.385259
38.525889
0.607623
44.174089
0.162387
16.238671
0.343121
12.416107
0.459792
16.973958
0.451878
39.097592
false
true
2024-05-18
2024-07-27
1
01-ai/Yi-1.5-34B
cognitivecomputations_dolphin-2.9.1-yi-1.5-9b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.1-yi-1.5-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.1-yi-1.5-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.1-yi-1.5-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.1-yi-1.5-9b
91f0a521e3e2a0675a3549aa5d3f40717068de94
24.93479
apache-2.0
26
8
true
false
false
true
1.050866
0.446533
44.653298
0.548431
35.77609
0.109517
10.951662
0.338087
11.744966
0.434802
13.516927
0.396692
32.965795
false
true
2024-05-18
2024-08-02
1
01-ai/Yi-1.5-9B
cognitivecomputations_dolphin-2.9.2-Phi-3-Medium_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.2-Phi-3-Medium</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.2-Phi-3-Medium-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.2-Phi-3-Medium
0470c5b912b51fa6e27d87a8ea7feafacd8cb101
25.668897
mit
18
-1
true
false
false
true
0.840482
0.424776
42.477626
0.645674
49.72194
0.006042
0.60423
0.327181
10.290828
0.419052
11.414844
0.455535
39.503915
false
true
2024-05-31
2024-08-05
1
cognitivecomputations/dolphin-2.9.2-Phi-3-Medium (Merge)
cognitivecomputations_dolphin-2.9.2-Phi-3-Medium-abliterated_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.2-Phi-3-Medium-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated
d50be5f22ca9745a2a3175996611d6a840318b7f
25.590064
mit
16
13
true
false
false
false
0.843954
0.361254
36.12537
0.612323
45.441267
0.123867
12.386707
0.32802
10.402685
0.411177
10.363802
0.449385
38.820553
false
true
2024-06-03
2024-06-27
1
cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated (Merge)
cognitivecomputations_dolphin-2.9.2-Phi-3-Medium-abliterated_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.2-Phi-3-Medium-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated
d50be5f22ca9745a2a3175996611d6a840318b7f
25.618429
mit
16
13
true
false
false
true
0.820797
0.412361
41.236142
0.638289
48.385347
0.006798
0.679758
0.328859
10.514541
0.434927
13.732552
0.45246
39.162234
false
true
2024-06-03
2024-08-05
1
cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated (Merge)
cognitivecomputations_dolphin-2.9.2-qwen2-72b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.2-qwen2-72b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.2-qwen2-72b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.2-qwen2-72b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.2-qwen2-72b
e79582577c2bf2af304221af0e8308b7e7d46ca1
35.745292
other
123
72
true
false
false
true
25.115554
0.634378
63.43779
0.629636
47.696174
0.206193
20.619335
0.369966
15.995526
0.452073
17.042448
0.547124
49.680482
false
true
2024-05-27
2024-10-20
1
Qwen/Qwen2-72B
cognitivecomputations_dolphin-2.9.2-qwen2-7b_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.2-qwen2-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.2-qwen2-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.2-qwen2-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.2-qwen2-7b
c443c4eb5138ed746ac49ed98bf3c183dc5380ac
21.183965
apache-2.0
64
7
true
false
false
true
1.279197
0.35346
35.345993
0.489383
27.914875
0.129154
12.915408
0.290268
5.369128
0.419146
11.659896
0.405086
33.898493
false
true
2024-05-24
2024-07-10
1
Qwen/Qwen2-7B
cognitivecomputations_dolphin-2.9.3-Yi-1.5-34B-32k_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-Yi-1.5-34B-32k-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k
ff4eee6438194a670a95dff3118b5231eb568610
27.073206
apache-2.0
18
34
true
false
false
true
3.245261
0.363927
36.39266
0.6047
43.406476
0.165408
16.540785
0.343121
12.416107
0.431052
13.348177
0.463015
40.335033
false
true
2024-06-23
2024-07-27
1
01-ai/Yi-1.5-34B-32k
cognitivecomputations_dolphin-2.9.3-mistral-7B-32k_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.3-mistral-7B-32k" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.3-mistral-7B-32k</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
4f4273ee8e7930dd64e2c6121c79d12546b883e2
19.373872
apache-2.0
46
7
true
false
false
true
0.600083
0.412636
41.263625
0.481254
26.906354
0.052115
5.21148
0.285235
4.697987
0.46426
17.932552
0.282081
20.231235
false
true
2024-06-25
2024-07-04
1
mistralai/Mistral-7B-v0.3
cognitivecomputations_dolphin-2.9.3-mistral-nemo-12b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-nemo-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b
7b535c900688fc836fbeebaeb7133910b09bafda
24.682904
apache-2.0
86
12
true
false
false
true
1.375142
0.560089
56.008945
0.548037
36.082759
0.056647
5.664653
0.315436
8.724832
0.44299
15.207031
0.337683
26.409205
false
true
2024-07-23
2024-07-26
1
mistralai/Mistral-Nemo-Base-2407
cognitivecomputations_dolphin-2.9.4-gemma2-2b_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Gemma2ForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.4-gemma2-2b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.4-gemma2-2b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.4-gemma2-2b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.4-gemma2-2b
5c0854beb88a6711221771d1b13d51f733e6ca06
9.797441
gemma
33
2
true
false
false
true
1.511248
0.089551
8.955128
0.408132
17.367633
0.046828
4.682779
0.284396
4.58613
0.417969
10.91276
0.210522
12.280216
false
true
2024-08-24
2024-08-25
1
google/gemma-2-2b
cognitivecomputations_dolphin-2.9.4-llama3.1-8b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cognitivecomputations/dolphin-2.9.4-llama3.1-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cognitivecomputations/dolphin-2.9.4-llama3.1-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cognitivecomputations__dolphin-2.9.4-llama3.1-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cognitivecomputations/dolphin-2.9.4-llama3.1-8b
7b73d1b7760bf9abac168de3d388b30d1ca1a138
6.955628
llama3.1
92
8
true
false
false
true
1.756313
0.275724
27.572397
0.352363
8.972089
0.001511
0.151057
0.263423
1.789709
0.323615
0.61849
0.12367
2.630024
false
true
2024-08-04
2024-09-17
1
meta-llama/Meta-Llama-3.1-8B
collaiborateorg_Collaiborator-MEDLLM-Llama-3-8B-v2_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/collaiborateorg__Collaiborator-MEDLLM-Llama-3-8B-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v2
2560556d655d0ecaefec10f579c92292d65fb28b
17.951635
0
8
false
false
false
false
0.705789
0.380887
38.088716
0.464803
23.648503
0.057402
5.740181
0.333054
11.073826
0.343427
1.595052
0.348072
27.563534
false
false
2024-06-27
0
Removed
cpayne1303_cp2024_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cpayne1303/cp2024" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cpayne1303/cp2024</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cpayne1303__cp2024-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cpayne1303/cp2024
fb354aaa73c40b4f1fc6e86beea733e4f3929470
3.614016
apache-2.0
0
0
true
false
false
false
0.047613
0.165814
16.581448
0.298539
2.739141
0
0
0.255872
0.782998
0.338313
0.455729
0.110123
1.124778
false
false
2024-11-26
2024-11-26
0
cpayne1303/cp2024
cpayne1303_cp2024-instruct_bfloat16
bfloat16
🟩 continuously pretrained
🟩
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cpayne1303/cp2024-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cpayne1303/cp2024-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cpayne1303__cp2024-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cpayne1303/cp2024-instruct
ac4cfbc28479f8a94e3eb745526620be9b75edfa
4.319731
apache-2.0
0
0
true
false
false
true
0.032162
0.170611
17.061065
0.294678
2.4813
0
0
0.260067
1.342282
0.368635
3.179427
0.116689
1.854314
false
false
2024-11-27
2024-11-27
1
cpayne1303/cp2024
cpayne1303_llama-43m-beta_bfloat16
bfloat16
🟩 continuously pretrained
🟩
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cpayne1303/llama-43m-beta" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cpayne1303/llama-43m-beta</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cpayne1303__llama-43m-beta-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cpayne1303/llama-43m-beta
1f85bec8c3541ed58fc2fcf4e6f98c1c34d72f60
5.288332
apache-2.0
0
0
true
false
false
false
0.058392
0.191568
19.156837
0.297678
2.482041
0
0
0.268456
2.46085
0.387177
6.163802
0.113198
1.46646
false
false
2024-11-30
2024-11-30
1
JackFram/llama-68m
cpayne1303_llama-43m-beta_float16
float16
🟩 continuously pretrained
🟩
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cpayne1303/llama-43m-beta" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cpayne1303/llama-43m-beta</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cpayne1303__llama-43m-beta-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cpayne1303/llama-43m-beta
1f85bec8c3541ed58fc2fcf4e6f98c1c34d72f60
5.3471
apache-2.0
0
0
true
false
false
false
0.059916
0.194891
19.489067
0.296463
2.496048
0
0
0.268456
2.46085
0.388542
6.401042
0.11112
1.235594
false
false
2024-11-30
2024-12-04
1
JackFram/llama-68m
cpayne1303_smallcp2024_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cpayne1303/smallcp2024" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cpayne1303/smallcp2024</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cpayne1303__smallcp2024-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cpayne1303/smallcp2024
ef995127242553e4126190e7f70f927504834360
3.455732
apache-2.0
0
0
true
false
false
false
0.047308
0.158196
15.819581
0.302705
3.118178
0
0
0.230705
0
0.342469
0.533333
0.11137
1.263298
false
false
2024-11-27
2024-11-27
0
cpayne1303/smallcp2024
cstr_llama3.1-8b-spaetzle-v90_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cstr/llama3.1-8b-spaetzle-v90" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cstr/llama3.1-8b-spaetzle-v90</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cstr__llama3.1-8b-spaetzle-v90-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cstr/llama3.1-8b-spaetzle-v90
717e5c3d31ed2465cd7cf927327adf677a9420b5
27.830191
llama3
0
8
true
false
false
true
0.778908
0.735619
73.561927
0.530286
32.763666
0.148036
14.803625
0.282718
4.362416
0.413437
11.146354
0.373088
30.343159
true
false
2024-09-15
2024-09-15
1
cstr/llama3.1-8b-spaetzle-v90 (Merge)
cyberagent_calm3-22b-chat_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/cyberagent/calm3-22b-chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">cyberagent/calm3-22b-chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/cyberagent__calm3-22b-chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
cyberagent/calm3-22b-chat
055922aa0f0fb1fbfbc97a2e31134532485ee99b
21.37559
apache-2.0
70
22
true
false
false
true
1.774248
0.509131
50.913133
0.499168
29.520884
0.064955
6.495468
0.276846
3.579418
0.455323
16.082031
0.294963
21.662603
false
false
2024-07-01
2024-07-04
0
cyberagent/calm3-22b-chat
darkc0de_BuddyGlassNeverSleeps_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/darkc0de/BuddyGlassNeverSleeps" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">darkc0de/BuddyGlassNeverSleeps</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/darkc0de__BuddyGlassNeverSleeps-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
darkc0de/BuddyGlassNeverSleeps
f8849498f02c94b68ef0308a7bf6637264949a7d
19.845818
2
8
false
false
false
false
1.354149
0.423902
42.390191
0.497723
28.477953
0.064199
6.41994
0.294463
5.928412
0.399271
8.608854
0.345246
27.249557
false
false
2024-09-16
2024-09-16
1
darkc0de/BuddyGlassNeverSleeps (Merge)
darkc0de_BuddyGlass_v0.3_Xortron7MethedUpSwitchedUp_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/darkc0de/BuddyGlass_v0.3_Xortron7MethedUpSwitchedUp" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">darkc0de/BuddyGlass_v0.3_Xortron7MethedUpSwitchedUp</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/darkc0de__BuddyGlass_v0.3_Xortron7MethedUpSwitchedUp-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
darkc0de/BuddyGlass_v0.3_Xortron7MethedUpSwitchedUp
57367fefe01c7d9653c303b28449b416fc777d93
22.265315
1
0
false
false
false
false
0.898182
0.435842
43.584245
0.524309
31.869311
0.124622
12.462236
0.298658
6.487696
0.414333
9.491667
0.367271
29.696735
false
false
2024-09-10
2024-09-15
1
darkc0de/BuddyGlass_v0.3_Xortron7MethedUpSwitchedUp (Merge)
databricks_dbrx-instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
DbrxForCausalLM
<a target="_blank" href="https://huggingface.co/databricks/dbrx-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">databricks/dbrx-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/databricks__dbrx-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
databricks/dbrx-instruct
c0a9245908c187da8f43a81e538e67ff360904ea
25.19901
other
1,104
131
true
false
false
true
47.958027
0.54158
54.157968
0.542896
35.96382
0.068731
6.873112
0.341443
12.192394
0.426927
12.199219
0.368268
29.80755
false
true
2024-03-26
2024-06-12
0
databricks/dbrx-instruct
databricks_dolly-v1-6b_bfloat16
bfloat16
🟢 pretrained
🟢
Original
GPTJForCausalLM
<a target="_blank" href="https://huggingface.co/databricks/dolly-v1-6b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">databricks/dolly-v1-6b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/databricks__dolly-v1-6b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
databricks/dolly-v1-6b
c9a85b3a322b402e20c839c702c725afe0cb454d
6.918291
cc-by-nc-4.0
310
6
true
false
false
false
0.66078
0.222443
22.244312
0.317209
4.781309
0.015106
1.510574
0.264262
1.901566
0.400417
8.11875
0.126579
2.953236
false
true
2023-03-23
2024-06-12
0
databricks/dolly-v1-6b
databricks_dolly-v2-12b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
GPTNeoXForCausalLM
<a target="_blank" href="https://huggingface.co/databricks/dolly-v2-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">databricks/dolly-v2-12b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/databricks__dolly-v2-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
databricks/dolly-v2-12b
19308160448536e378e3db21a73a751579ee7fdd
6.383024
mit
1,950
12
true
false
false
false
1.397119
0.235507
23.550734
0.331997
6.377894
0.01435
1.435045
0.240772
0
0.373906
5.504948
0.112866
1.429521
false
true
2023-04-11
2024-06-12
0
databricks/dolly-v2-12b
databricks_dolly-v2-3b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
GPTNeoXForCausalLM
<a target="_blank" href="https://huggingface.co/databricks/dolly-v2-3b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">databricks/dolly-v2-3b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/databricks__dolly-v2-3b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
databricks/dolly-v2-3b
f6c9be08f16fe4d3a719bee0a4a7c7415b5c65df
5.461189
mit
287
3
true
false
false
false
0.758084
0.224716
22.471598
0.307928
3.324769
0.006798
0.679758
0.260906
1.454139
0.333781
3.222656
0.114528
1.614214
false
true
2023-04-13
2024-06-12
0
databricks/dolly-v2-3b
databricks_dolly-v2-7b_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
GPTNeoXForCausalLM
<a target="_blank" href="https://huggingface.co/databricks/dolly-v2-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">databricks/dolly-v2-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/databricks__dolly-v2-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
databricks/dolly-v2-7b
d632f0c8b75b1ae5b26b250d25bfba4e99cb7c6f
5.571832
mit
148
7
true
false
false
false
0.830206
0.200986
20.098561
0.317306
5.449893
0.009819
0.981873
0.268456
2.46085
0.355302
2.779427
0.114943
1.660387
false
true
2023-04-13
2024-06-12
0
databricks/dolly-v2-7b
davidkim205_Rhea-72b-v0.5_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/davidkim205/Rhea-72b-v0.5" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">davidkim205/Rhea-72b-v0.5</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/davidkim205__Rhea-72b-v0.5-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
davidkim205/Rhea-72b-v0.5
bc3806efb23d2713e6630a748d9747fd76b27169
4.224031
apache-2.0
134
72
true
false
false
false
8.688691
0.014538
1.453809
0.307834
3.670747
0.067221
6.722054
0.252517
0.33557
0.424135
11.316927
0.116606
1.84508
false
false
2024-03-22
2024-09-15
0
davidkim205/Rhea-72b-v0.5
davidkim205_nox-solar-10.7b-v4_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/davidkim205/nox-solar-10.7b-v4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">davidkim205/nox-solar-10.7b-v4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/davidkim205__nox-solar-10.7b-v4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
davidkim205/nox-solar-10.7b-v4
5f4be6cb7d8398b84689148d15f3838f2e01e104
18.489145
apache-2.0
11
10
true
false
false
true
0.848976
0.375342
37.534187
0.481404
26.631088
0.006798
0.679758
0.307047
7.606264
0.429844
12.563802
0.333278
25.91977
false
false
2024-03-16
2024-10-04
0
davidkim205/nox-solar-10.7b-v4
deepseek-ai_deepseek-llm-67b-chat_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/deepseek-ai/deepseek-llm-67b-chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">deepseek-ai/deepseek-llm-67b-chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/deepseek-ai__deepseek-llm-67b-chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
deepseek-ai/deepseek-llm-67b-chat
79648bef7658bb824e4630740f6e1484c1b0620b
26.995929
other
177
67
true
false
false
true
59.821809
0.558715
55.871532
0.524342
33.225242
0.074018
7.401813
0.316275
8.836689
0.505865
23.933073
0.394365
32.707225
false
true
2023-11-29
2024-06-12
0
deepseek-ai/deepseek-llm-67b-chat
deepseek-ai_deepseek-llm-7b-base_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/deepseek-ai/deepseek-llm-7b-base" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">deepseek-ai/deepseek-llm-7b-base</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/deepseek-ai__deepseek-llm-7b-base-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
deepseek-ai/deepseek-llm-7b-base
7683fea62db869066ddaff6a41d032262c490d4f
8.138982
other
36
7
true
false
false
false
0.822536
0.217872
21.787191
0.350303
9.767925
0.01435
1.435045
0.27349
3.131991
0.373781
3.75599
0.180602
8.955748
false
true
2023-11-29
2024-06-12
0
deepseek-ai/deepseek-llm-7b-base
deepseek-ai_deepseek-llm-7b-chat_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/deepseek-ai/deepseek-llm-7b-chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">deepseek-ai/deepseek-llm-7b-chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/deepseek-ai__deepseek-llm-7b-chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
deepseek-ai/deepseek-llm-7b-chat
afbda8b347ec881666061fa67447046fc5164ec8
14.785393
other
76
7
true
false
false
true
0.774483
0.417082
41.708223
0.363208
11.258949
0.018127
1.812689
0.26594
2.12528
0.466771
19.213021
0.213348
12.594193
false
true
2023-11-29
2024-06-12
0
deepseek-ai/deepseek-llm-7b-chat
deepseek-ai_deepseek-moe-16b-base_bfloat16
bfloat16
🟢 pretrained
🟢
Original
DeepseekForCausalLM
<a target="_blank" href="https://huggingface.co/deepseek-ai/deepseek-moe-16b-base" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">deepseek-ai/deepseek-moe-16b-base</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/deepseek-ai__deepseek-moe-16b-base-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
deepseek-ai/deepseek-moe-16b-base
521d2bc4fb69a3f3ae565310fcc3b65f97af2580
7.390805
other
84
16
true
true
false
false
7.002465
0.244974
24.497445
0.340946
8.355556
0.019637
1.963746
0.254195
0.559284
0.365781
3.35599
0.150515
5.61281
false
true
2024-01-08
2024-06-12
0
deepseek-ai/deepseek-moe-16b-base