eval_name
stringlengths
12
111
Precision
stringclasses
3 values
Type
stringclasses
6 values
T
stringclasses
6 values
Weight type
stringclasses
2 values
Architecture
stringclasses
52 values
Model
stringlengths
355
689
fullname
stringlengths
4
102
Model sha
stringlengths
0
40
Average ⬆️
float64
1.03
52
Hub License
stringclasses
26 values
Hub ❤️
int64
0
5.9k
#Params (B)
int64
-1
140
Available on the hub
bool
2 classes
MoE
bool
2 classes
Flagged
bool
2 classes
Chat Template
bool
2 classes
CO₂ cost (kg)
float64
0.03
107
IFEval Raw
float64
0
0.9
IFEval
float64
0
90
BBH Raw
float64
0.27
0.75
BBH
float64
0.81
63.5
MATH Lvl 5 Raw
float64
0
0.51
MATH Lvl 5
float64
0
50.7
GPQA Raw
float64
0.22
0.44
GPQA
float64
0
24.9
MUSR Raw
float64
0.29
0.6
MUSR
float64
0
38.5
MMLU-PRO Raw
float64
0.1
0.73
MMLU-PRO
float64
0
70
Merged
bool
2 classes
Official Providers
bool
2 classes
Upload To Hub Date
stringclasses
424 values
Submission Date
stringclasses
169 values
Generation
int64
0
10
Base Model
stringlengths
4
102
FlofloB_83k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit_float16
float16
🟩 continuously pretrained
🟩
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/FlofloB/83k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FlofloB/83k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FlofloB__83k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FlofloB/83k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit
4c4d3660d0288295f89880a3a86f4eb9ecc9d344
7.923936
apache-2.0
1
0
true
false
false
true
0.492186
0.28694
28.693976
0.334653
8.132273
0
0
0.27349
3.131991
0.328948
1.41849
0.155502
6.166888
false
false
2024-11-26
2024-11-26
3
Qwen/Qwen2.5-0.5B
FlofloB_test_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit_float16
float16
🟩 continuously pretrained
🟩
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/FlofloB/test_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FlofloB/test_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FlofloB__test_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FlofloB/test_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit
cfd97ca5927a2e09ec30001a576d82dd8b635e09
24.460526
apache-2.0
1
16
true
false
false
true
1.008801
0.521546
52.154616
0.524083
32.882433
0.108761
10.876133
0.311242
8.165548
0.424417
12.452083
0.372091
30.232343
false
false
2024-11-21
2024-11-21
1
unsloth/phi-3-mini-4k-instruct-bnb-4bit
FuJhen_ft-openhermes-25-mistral-7b-irca-dpo-pairs_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/FuJhen/ft-openhermes-25-mistral-7b-irca-dpo-pairs" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FuJhen/ft-openhermes-25-mistral-7b-irca-dpo-pairs</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FuJhen__ft-openhermes-25-mistral-7b-irca-dpo-pairs-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FuJhen/ft-openhermes-25-mistral-7b-irca-dpo-pairs
24c0bea14d53e6f67f1fbe2eca5bfe7cae389b33
19.615525
apache-2.0
0
14
true
false
false
true
1.002048
0.542004
54.20041
0.477303
26.596861
0.001511
0.151057
0.278523
3.803132
0.417375
11.205208
0.295628
21.73648
false
false
2024-09-12
2024-09-12
1
FuJhen/ft-openhermes-25-mistral-7b-irca-dpo-pairs (Merge)
FuJhen_mistral-instruct-7B-DPO_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/FuJhen/mistral-instruct-7B-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FuJhen/mistral-instruct-7B-DPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FuJhen__mistral-instruct-7B-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FuJhen/mistral-instruct-7B-DPO
e0bc86c23ce5aae1db576c8cca6f06f1f73af2db
19.016943
apache-2.0
0
14
true
false
false
true
1.009647
0.496842
49.684171
0.462391
24.925827
0.037764
3.776435
0.277685
3.691275
0.401563
9.428646
0.303358
22.595301
false
false
2024-09-12
2024-09-12
1
FuJhen/mistral-instruct-7B-DPO (Merge)
FuJhen_mistral_7b_v0.1_structedData_e2e_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Adapter
?
<a target="_blank" href="https://huggingface.co/FuJhen/mistral_7b_v0.1_structedData_e2e" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FuJhen/mistral_7b_v0.1_structedData_e2e</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FuJhen__mistral_7b_v0.1_structedData_e2e-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FuJhen/mistral_7b_v0.1_structedData_e2e
7231864981174d9bee8c7687c24c8344414eae6b
10.871547
apache-2.0
0
7
true
false
false
false
1.080246
0.172684
17.268403
0.411391
18.062424
0.002266
0.226586
0.279362
3.914989
0.372292
5.636458
0.281084
20.12042
false
false
2024-09-13
2024-09-13
1
FuJhen/mistral_7b_v0.1_structedData_e2e (Merge)
FuJhen_mistral_7b_v0.1_structedData_viggo_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Adapter
?
<a target="_blank" href="https://huggingface.co/FuJhen/mistral_7b_v0.1_structedData_viggo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FuJhen/mistral_7b_v0.1_structedData_viggo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FuJhen__mistral_7b_v0.1_structedData_viggo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FuJhen/mistral_7b_v0.1_structedData_viggo
7231864981174d9bee8c7687c24c8344414eae6b
12.352466
apache-2.0
0
14
true
false
false
false
1.076114
0.178329
17.832906
0.452386
23.960172
0.023414
2.34139
0.283557
4.474273
0.373813
3.926563
0.294215
21.579492
false
false
2024-09-13
2024-09-13
1
FuJhen/mistral_7b_v0.1_structedData_viggo (Merge)
FuseAI_FuseChat-7B-v2.0_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/FuseAI/FuseChat-7B-v2.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FuseAI/FuseChat-7B-v2.0</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/FuseAI__FuseChat-7B-v2.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
FuseAI/FuseChat-7B-v2.0
65fdb310c09f56b9aca01b89a849f06f39faeb75
20.184132
apache-2.0
9
7
true
false
false
false
0.443306
0.342319
34.231949
0.495421
29.341638
0.063444
6.344411
0.302013
6.935123
0.479667
20.225
0.31624
24.02667
false
false
2024-08-13
2024-11-21
1
openchat/openchat_3.5
GalrionSoftworks_MN-LooseCannon-12B-v1_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/GalrionSoftworks/MN-LooseCannon-12B-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GalrionSoftworks/MN-LooseCannon-12B-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GalrionSoftworks__MN-LooseCannon-12B-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GalrionSoftworks/MN-LooseCannon-12B-v1
21.885253
8
12
false
false
false
true
1.52902
0.541779
54.177915
0.512818
29.976062
0.070997
7.099698
0.285235
4.697987
0.413844
10.963802
0.319564
24.396055
false
false
2024-08-09
2024-09-05
1
GalrionSoftworks/MN-LooseCannon-12B-v1 (Merge)
GalrionSoftworks_MagnusIntellectus-12B-v1_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/GalrionSoftworks/MagnusIntellectus-12B-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GalrionSoftworks/MagnusIntellectus-12B-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GalrionSoftworks__MagnusIntellectus-12B-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GalrionSoftworks/MagnusIntellectus-12B-v1
fc83cb3eec2f8328448c5fe3cb830fc77983a6b9
21.622238
apache-2.0
5
12
true
false
false
true
1.624264
0.442137
44.213686
0.532301
33.262254
0.055891
5.589124
0.284396
4.58613
0.442802
15.183594
0.342088
26.898641
true
false
2024-08-13
2024-09-05
1
GalrionSoftworks/MagnusIntellectus-12B-v1 (Merge)
GoToCompany_gemma2-9b-cpt-sahabatai-v1-instruct_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Gemma2ForCausalLM
<a target="_blank" href="https://huggingface.co/GoToCompany/gemma2-9b-cpt-sahabatai-v1-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GoToCompany/gemma2-9b-cpt-sahabatai-v1-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GoToCompany__gemma2-9b-cpt-sahabatai-v1-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GoToCompany/gemma2-9b-cpt-sahabatai-v1-instruct
ca19cec82a7d2bdba20020e1bebf296417cfc3ee
32.342379
gemma
26
9
true
false
false
false
1.931095
0.655061
65.506079
0.595455
41.866504
0.197885
19.78852
0.334732
11.297539
0.477865
19.333073
0.426363
36.262559
false
false
2024-11-06
2024-11-20
1
GoToCompany/gemma2-9b-cpt-sahabatai-v1-instruct (Merge)
GoToCompany_llama3-8b-cpt-sahabatai-v1-instruct_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/GoToCompany/llama3-8b-cpt-sahabatai-v1-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GoToCompany/llama3-8b-cpt-sahabatai-v1-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GoToCompany__llama3-8b-cpt-sahabatai-v1-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GoToCompany/llama3-8b-cpt-sahabatai-v1-instruct
20fd3cff1dc86553d11b5c4b2fdbb6f2dd1ede55
22.908342
llama3
6
8
true
false
false
true
0.673411
0.523845
52.384451
0.495129
28.539529
0.11858
11.858006
0.266779
2.237136
0.448844
15.172135
0.345329
27.258791
false
false
2024-11-06
2024-11-20
1
GoToCompany/llama3-8b-cpt-sahabatai-v1-instruct (Merge)
Goekdeniz-Guelmez_Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1
bfc0e7dc6add02baecd9b6f84a078f7f3d164315
8.320603
apache-2.0
1
0
true
false
false
true
0.487642
0.34719
34.71899
0.326831
6.845786
0.002266
0.226586
0.251678
0.223714
0.32625
0.78125
0.164146
7.12729
false
false
2024-11-17
2024-11-18
2
Qwen/Qwen2.5-0.5B
Goekdeniz-Guelmez_Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-0.5B-Instruct-abliterated-v1
bfc0e7dc6add02baecd9b6f84a078f7f3d164315
8.415919
apache-2.0
1
0
true
false
false
true
0.498004
0.341694
34.169448
0.32921
7.221169
0.002266
0.226586
0.25755
1.006711
0.324917
0.78125
0.163813
7.090352
false
false
2024-11-17
2024-11-18
2
Qwen/Qwen2.5-0.5B
Goekdeniz-Guelmez_Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1
eca7edeba61e894597e9940348e8d90817c1ad79
15.294146
apache-2.0
4
1
true
false
false
true
0.783381
0.476858
47.685807
0.418601
18.306013
0.019637
1.963746
0.243289
0
0.36749
4.002865
0.278258
19.806442
false
false
2024-09-20
2024-09-28
1
Qwen/Qwen2.5-1.5B
Goekdeniz-Guelmez_Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v2
ff4a6eff69adb015dfcfbff7a2d2dc43b34afe89
13.665944
apache-2.0
1
1
true
false
false
true
0.719243
0.421554
42.15537
0.404189
16.499503
0.01284
1.283988
0.239933
0
0.376854
4.706771
0.25615
17.35003
false
false
2024-09-28
2024-09-28
2
Qwen/Qwen2.5-1.5B
Goekdeniz-Guelmez_Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v3
03ffa6f7a6ada9d63d838707c597297f048d409b
13.540924
apache-2.0
1
1
true
false
false
true
0.706201
0.425251
42.525056
0.405345
16.439712
0.007553
0.755287
0.243289
0
0.370187
4.240104
0.255568
17.285387
false
false
2024-09-28
2024-09-28
3
Qwen/Qwen2.5-1.5B
Goekdeniz-Guelmez_Josiefied-Qwen2.5-14B-Instruct-abliterated-v4_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-14B-Instruct-abliterated-v4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4
00afd27eef16e835fcb0d8e687435dca3c185bdf
33.511798
apache-2.0
12
14
true
false
false
true
1.747117
0.829167
82.916661
0.635564
48.05227
0
0
0.342282
12.304251
0.428667
13.15
0.501828
44.647606
false
false
2024-10-21
2024-10-23
2
Qwen/Qwen2.5-14B
Goekdeniz-Guelmez_Josiefied-Qwen2.5-7B-Instruct-abliterated-v2_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__Josiefied-Qwen2.5-7B-Instruct-abliterated-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2
ecf4024048ea1be2f0840a50080fb79b88aacde9
27.763763
apache-2.0
4
7
true
false
false
true
1.201506
0.781381
78.138118
0.530967
33.333986
0
0
0.298658
6.487696
0.435396
13.957813
0.411985
34.664967
false
false
2024-09-20
2024-10-08
1
Qwen/Qwen2.5-7B
Goekdeniz-Guelmez_j.o.s.i.e.v4o-1.5b-dpo-stage1-v1_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Goekdeniz-Guelmez/j.o.s.i.e.v4o-1.5b-dpo-stage1-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Goekdeniz-Guelmez/j.o.s.i.e.v4o-1.5b-dpo-stage1-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Goekdeniz-Guelmez__j.o.s.i.e.v4o-1.5b-dpo-stage1-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Goekdeniz-Guelmez/j.o.s.i.e.v4o-1.5b-dpo-stage1-v1
d5ddad290d83b1ba8a7612a6c1cfad6fb4346fe4
13.567474
apache-2.0
1
1
true
false
false
true
0.791153
0.418831
41.883092
0.412421
17.748017
0.029456
2.945619
0.250839
0.111857
0.352854
1.440104
0.255485
17.276152
false
false
2024-10-07
2024-10-08
2
Qwen/Qwen2.5-1.5B
GreenNode_GreenNode-small-9B-it_float16
float16
🟩 continuously pretrained
🟩
Original
Gemma2ForCausalLM
<a target="_blank" href="https://huggingface.co/GreenNode/GreenNode-small-9B-it" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GreenNode/GreenNode-small-9B-it</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GreenNode__GreenNode-small-9B-it-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GreenNode/GreenNode-small-9B-it
1ba4ce8e2267c7fcc820961a9bfc13ab80150866
28.286651
0
9
false
false
false
true
2.645944
0.743613
74.36125
0.599384
41.899926
0
0
0.319631
9.284116
0.420417
11.652083
0.392703
32.522533
false
false
2024-10-14
0
Removed
GritLM_GritLM-7B-KTO_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/GritLM/GritLM-7B-KTO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GritLM/GritLM-7B-KTO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GritLM__GritLM-7B-KTO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GritLM/GritLM-7B-KTO
b5c48669508c1de18c698460c187f64e90e7df44
19.172954
apache-2.0
4
7
true
false
false
true
0.639864
0.531013
53.101327
0.485294
27.904318
0.023414
2.34139
0.297819
6.375839
0.371021
6.644271
0.268035
18.670582
false
false
2024-04-16
2024-08-04
0
GritLM/GritLM-7B-KTO
GritLM_GritLM-8x7B-KTO_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/GritLM/GritLM-8x7B-KTO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">GritLM/GritLM-8x7B-KTO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/GritLM__GritLM-8x7B-KTO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
GritLM/GritLM-8x7B-KTO
938913477064fcc498757c5136d9899bb6e713ed
25.838485
apache-2.0
3
46
true
false
false
true
4.604463
0.571405
57.140498
0.58203
40.826162
0.098187
9.818731
0.296141
6.152125
0.421656
11.673698
0.364777
29.419696
false
false
2024-04-17
2024-08-04
0
GritLM/GritLM-8x7B-KTO
Gryphe_Pantheon-RP-1.0-8b-Llama-3_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Gryphe/Pantheon-RP-1.0-8b-Llama-3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gryphe/Pantheon-RP-1.0-8b-Llama-3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gryphe__Pantheon-RP-1.0-8b-Llama-3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gryphe/Pantheon-RP-1.0-8b-Llama-3
70a6df202c9df9abdc6928bec5a5ab47f2667aee
16.772417
apache-2.0
46
8
true
false
false
true
0.720836
0.393252
39.325213
0.453908
23.631915
0.057402
5.740181
0.276007
3.467562
0.38324
5.504948
0.306682
22.964687
false
false
2024-05-08
2024-06-27
1
meta-llama/Meta-Llama-3-8B
Gryphe_Pantheon-RP-1.5-12b-Nemo_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Gryphe/Pantheon-RP-1.5-12b-Nemo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gryphe/Pantheon-RP-1.5-12b-Nemo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gryphe__Pantheon-RP-1.5-12b-Nemo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gryphe/Pantheon-RP-1.5-12b-Nemo
00107381f05f69666772d88a1b11affe77c94a47
21.311159
apache-2.0
29
12
true
false
false
true
1.685583
0.476308
47.630842
0.519582
31.750144
0.048338
4.833837
0.272651
3.020134
0.442031
15.053906
0.330203
25.578088
false
false
2024-07-25
2024-08-04
1
mistralai/Mistral-Nemo-Base-2407
Gryphe_Pantheon-RP-1.6-12b-Nemo_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Gryphe/Pantheon-RP-1.6-12b-Nemo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gryphe/Pantheon-RP-1.6-12b-Nemo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gryphe__Pantheon-RP-1.6-12b-Nemo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gryphe/Pantheon-RP-1.6-12b-Nemo
60cf38ae0367baf314e3cce748d9a199adfea557
20.365189
apache-2.0
11
12
true
false
false
true
1.737253
0.448057
44.805671
0.520401
31.687344
0.033988
3.398792
0.277685
3.691275
0.42876
12.928385
0.331117
25.679669
false
false
2024-08-18
2024-08-31
1
mistralai/Mistral-Nemo-Base-2407
Gryphe_Pantheon-RP-1.6-12b-Nemo-KTO_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Gryphe/Pantheon-RP-1.6-12b-Nemo-KTO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gryphe/Pantheon-RP-1.6-12b-Nemo-KTO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gryphe__Pantheon-RP-1.6-12b-Nemo-KTO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gryphe/Pantheon-RP-1.6-12b-Nemo-KTO
6cb6d8d9a7352d71f539ab5053987e058c090443
21.407541
apache-2.0
5
12
true
false
false
true
1.682026
0.463619
46.361875
0.527698
33.0322
0.043807
4.380665
0.295302
6.040268
0.424792
12.165625
0.338182
26.464613
false
false
2024-08-28
2024-08-31
1
mistralai/Mistral-Nemo-Base-2407
Gryphe_Pantheon-RP-Pure-1.6.2-22b-Small_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Gryphe/Pantheon-RP-Pure-1.6.2-22b-Small" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gryphe/Pantheon-RP-Pure-1.6.2-22b-Small</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gryphe__Pantheon-RP-Pure-1.6.2-22b-Small-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gryphe/Pantheon-RP-Pure-1.6.2-22b-Small
d031830dcb3bc5ad9634374db4dd15b3ef6ebe0f
27.823932
other
16
22
true
false
false
true
1.45332
0.693104
69.31043
0.530454
31.683163
0.183535
18.353474
0.328859
10.514541
0.376479
4.393229
0.394199
32.688756
false
false
2024-10-13
2024-10-15
1
mistralai/Mistral-Small-Instruct-2409
Gunulhona_Gemma-Ko-Merge_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Gemma2ForCausalLM
<a target="_blank" href="https://huggingface.co/Gunulhona/Gemma-Ko-Merge" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gunulhona/Gemma-Ko-Merge</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gunulhona__Gemma-Ko-Merge-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gunulhona/Gemma-Ko-Merge
ca6b0eb1405f21db6a7a9cce3b112d21fcfdde97
25.935394
0
10
false
false
false
true
3.137248
0.641572
64.157214
0.581303
38.787197
0.001511
0.151057
0.33557
11.409396
0.404698
9.120573
0.387882
31.986924
false
false
2024-09-04
2024-10-23
1
Gunulhona/Gemma-Ko-Merge (Merge)
Gunulhona_Gemma-Ko-Merge-PEFT_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Adapter
?
<a target="_blank" href="https://huggingface.co/Gunulhona/Gemma-Ko-Merge-PEFT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gunulhona/Gemma-Ko-Merge-PEFT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gunulhona__Gemma-Ko-Merge-PEFT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gunulhona/Gemma-Ko-Merge-PEFT
ca6b0eb1405f21db6a7a9cce3b112d21fcfdde97
18.169495
0
20
false
false
false
false
5.876477
0.288039
28.803907
0.515409
30.186273
0
0
0.324664
9.955257
0.40801
8.767969
0.381732
31.303561
false
false
2024-09-30
2024-10-17
0
Gunulhona/Gemma-Ko-Merge-PEFT
Gunulhona_Gemma-Ko-Merge-PEFT_float16
float16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/Gunulhona/Gemma-Ko-Merge-PEFT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Gunulhona/Gemma-Ko-Merge-PEFT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Gunulhona__Gemma-Ko-Merge-PEFT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Gunulhona/Gemma-Ko-Merge-PEFT
ca6b0eb1405f21db6a7a9cce3b112d21fcfdde97
18.06624
0
20
false
false
false
true
9.394334
0.444135
44.41349
0.486299
26.015069
0
0
0.307047
7.606264
0.398583
7.05625
0.309757
23.306368
false
false
2024-09-30
2024-10-23
0
Gunulhona/Gemma-Ko-Merge-PEFT
HPAI-BSC_Llama3-Aloe-8B-Alpha_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HPAI-BSC/Llama3-Aloe-8B-Alpha" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HPAI-BSC/Llama3-Aloe-8B-Alpha</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HPAI-BSC__Llama3-Aloe-8B-Alpha-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HPAI-BSC/Llama3-Aloe-8B-Alpha
f0bce5c1fee5ea2a6679bb3dc9de8548e7262c9e
20.104566
cc-by-nc-4.0
53
8
true
false
false
true
0.795245
0.508107
50.810738
0.483085
27.145978
0.053625
5.362538
0.294463
5.928412
0.367271
5.875521
0.329538
25.504211
false
false
2024-04-26
2024-10-29
0
HPAI-BSC/Llama3-Aloe-8B-Alpha
HPAI-BSC_Llama3.1-Aloe-Beta-8B_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HPAI-BSC/Llama3.1-Aloe-Beta-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HPAI-BSC/Llama3.1-Aloe-Beta-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HPAI-BSC__Llama3.1-Aloe-Beta-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HPAI-BSC/Llama3.1-Aloe-Beta-8B
3f2f0bbfb03cb0a8310efa50659688c1f2c02da0
23.754809
llama3.1
10
8
true
false
false
true
1.398697
0.725328
72.532769
0.509276
30.369625
0.016616
1.661631
0.268456
2.46085
0.383458
6.832292
0.358045
28.67169
false
false
2024-10-30
2024-11-07
0
HPAI-BSC/Llama3.1-Aloe-Beta-8B
Hastagaras_Llama-3.1-Jamet-8B-MK.I_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Hastagaras/Llama-3.1-Jamet-8B-MK.I" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Hastagaras/Llama-3.1-Jamet-8B-MK.I</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Hastagaras__Llama-3.1-Jamet-8B-MK.I-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Hastagaras/Llama-3.1-Jamet-8B-MK.I
26cb97042b04fee7d0140375a7babbf92278f8ac
25.39863
llama3.1
1
8
true
false
false
true
0.71874
0.733821
73.382071
0.504867
29.503905
0.125378
12.537764
0.274329
3.243848
0.372604
6.142188
0.348238
27.582004
false
false
2024-11-18
2024-11-18
0
Hastagaras/Llama-3.1-Jamet-8B-MK.I
Hastagaras_Zabuza-8B-Llama-3.1_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Hastagaras/Zabuza-8B-Llama-3.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Hastagaras/Zabuza-8B-Llama-3.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Hastagaras__Zabuza-8B-Llama-3.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Hastagaras/Zabuza-8B-Llama-3.1
57ffa92f229b8308916aae1d64d8f0dc9baa0a34
19.711829
llama3.1
0
8
true
false
false
true
0.675287
0.626534
62.653426
0.453892
23.220321
0.042296
4.229607
0.264262
1.901566
0.356792
4.898958
0.292304
21.367095
true
false
2024-11-05
2024-11-05
1
Hastagaras/Zabuza-8B-Llama-3.1 (Merge)
HiroseKoichi_Llama-Salad-4x8B-V3_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/HiroseKoichi/Llama-Salad-4x8B-V3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HiroseKoichi/Llama-Salad-4x8B-V3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HiroseKoichi__Llama-Salad-4x8B-V3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HiroseKoichi/Llama-Salad-4x8B-V3
a343915429779efbd1478f01ba1f7fd9d8d226c0
24.93529
llama3
5
24
true
true
false
true
2.137695
0.665352
66.535238
0.524465
31.928849
0.096677
9.667674
0.302852
7.04698
0.374031
6.453906
0.351812
27.979093
true
false
2024-06-17
2024-06-26
0
HiroseKoichi/Llama-Salad-4x8B-V3
HuggingFaceH4_zephyr-7b-alpha_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceH4/zephyr-7b-alpha</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceH4__zephyr-7b-alpha-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceH4/zephyr-7b-alpha
2ce2d025864af849b3e5029e2ec9d568eeda892d
18.571864
mit
1,102
7
true
false
false
true
0.795675
0.519148
51.914808
0.458786
23.955291
0.017372
1.73716
0.297819
6.375839
0.394958
7.503125
0.279505
19.944962
false
true
2023-10-09
2024-06-12
1
mistralai/Mistral-7B-v0.1
HuggingFaceH4_zephyr-7b-beta_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceH4/zephyr-7b-beta" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceH4/zephyr-7b-beta</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceH4__zephyr-7b-beta-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceH4/zephyr-7b-beta
b70e0c9a2d9e14bd1e812d3c398e5f313e93b473
17.767061
mit
1,617
7
true
false
false
true
0.555023
0.495043
49.504315
0.431582
21.487542
0.02719
2.719033
0.290268
5.369128
0.392542
7.734375
0.278092
19.787973
false
true
2023-10-26
2024-06-12
1
mistralai/Mistral-7B-v0.1
HuggingFaceH4_zephyr-7b-gemma-v0.1_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
GemmaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceH4/zephyr-7b-gemma-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceH4/zephyr-7b-gemma-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceH4__zephyr-7b-gemma-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceH4/zephyr-7b-gemma-v0.1
03b3427d0ed07d2e0f86c0a7e53d82d4beef9540
15.929338
other
121
8
true
false
false
true
1.481775
0.336374
33.637415
0.462374
23.751163
0.075529
7.55287
0.294463
5.928412
0.373969
4.179427
0.284741
20.526743
false
true
2024-03-01
2024-06-12
2
google/gemma-7b
HuggingFaceH4_zephyr-orpo-141b-A35b-v0.1_float16
float16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceH4__zephyr-orpo-141b-A35b-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1
a3be084543d278e61b64cd600f28157afc79ffd3
34.063023
apache-2.0
261
140
true
false
false
true
42.067786
0.651089
65.108911
0.629044
47.503796
0.200906
20.090634
0.378356
17.114094
0.446521
14.715104
0.45861
39.845597
false
true
2024-04-10
2024-06-12
1
mistral-community/Mixtral-8x22B-v0.1
HuggingFaceTB_SmolLM-1.7B_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM-1.7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM-1.7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM-1.7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM-1.7B
673a07602ca1191e5bc2ddda428e2f608a0a14c0
5.425399
apache-2.0
164
1
true
false
false
false
0.324307
0.236157
23.615673
0.318052
4.411128
0.007553
0.755287
0.241611
0
0.342094
2.128385
0.114777
1.641918
false
true
2024-07-14
2024-07-18
0
HuggingFaceTB/SmolLM-1.7B
HuggingFaceTB_SmolLM-1.7B-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM-1.7B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM-1.7B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM-1.7B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM-1.7B-Instruct
0ad161e59935a9a691dfde2818df8b98786f30a7
5.138222
apache-2.0
107
1
true
false
false
true
0.317023
0.234783
23.47826
0.288511
2.080374
0
0
0.260067
1.342282
0.348667
2.083333
0.116606
1.84508
false
true
2024-07-15
2024-07-18
1
HuggingFaceTB/SmolLM-1.7B
HuggingFaceTB_SmolLM-135M_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM-135M" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM-135M</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM-135M-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM-135M
eec6e461571fba3e197a57c298f60b75422eae02
6.838197
apache-2.0
180
0
true
false
false
false
0.343378
0.212476
21.247623
0.304605
3.2854
0.006798
0.679758
0.258389
1.118568
0.436604
13.342188
0.112201
1.355644
false
true
2024-07-14
2024-07-18
0
HuggingFaceTB/SmolLM-135M
HuggingFaceTB_SmolLM-135M-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM-135M-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM-135M-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM-135M-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM-135M-Instruct
8ca7af58e27777cae460ad8ca3ab9db15f5c160d
3.564171
apache-2.0
98
0
true
false
false
true
0.467805
0.121401
12.140122
0.301508
2.692958
0
0
0.259228
1.230425
0.363458
3.365625
0.117603
1.955895
false
true
2024-07-15
2024-10-12
1
HuggingFaceTB/SmolLM-135M
HuggingFaceTB_SmolLM-360M_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM-360M" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM-360M</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM-360M-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM-360M
318cc630b73730bfd712e5873063156ffb8936b5
6.147596
apache-2.0
62
0
true
false
false
false
0.36526
0.213351
21.335058
0.306452
3.284915
0.004532
0.453172
0.267617
2.348993
0.401781
8.089323
0.112367
1.374113
false
true
2024-07-14
2024-07-18
0
HuggingFaceTB/SmolLM-360M
HuggingFaceTB_SmolLM-360M-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM-360M-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM-360M-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM-360M-Instruct
8e951de8c220295ea4f85d078c4e320df7137535
4.706784
apache-2.0
77
0
true
false
false
true
0.366501
0.195165
19.516549
0.288511
2.080374
0
0
0.264262
1.901566
0.347177
2.897135
0.116606
1.84508
false
true
2024-07-15
2024-08-20
1
HuggingFaceTB/SmolLM-360M
HuggingFaceTB_SmolLM2-1.7B_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-1.7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-1.7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-1.7B
4fa12cab4f5f53670b05125fb9d2873af587d231
9.495504
apache-2.0
83
1
true
false
false
false
0.325026
0.244
24.400036
0.345259
9.301788
0.021148
2.114804
0.279362
3.914989
0.348542
4.601042
0.213763
12.640366
false
true
2024-10-30
2024-11-06
0
HuggingFaceTB/SmolLM2-1.7B
HuggingFaceTB_SmolLM2-1.7B-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-1.7B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-1.7B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-1.7B-Instruct
d1bb90bcfbe0f211109880f4da18da66f229c4f6
14.745339
apache-2.0
429
1
true
false
false
true
0.324961
0.536784
53.678351
0.359862
10.917989
0.041541
4.154079
0.279362
3.914989
0.342125
4.098958
0.205369
11.707668
false
true
2024-10-31
2024-11-06
1
HuggingFaceTB/SmolLM2-1.7B-Instruct (Merge)
HuggingFaceTB_SmolLM2-135M_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-135M" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-135M</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-135M-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-135M
28e66ca6931668447a3bac213f23d990ad3b0e2b
5.557677
apache-2.0
37
0
true
false
false
false
0.333905
0.1833
18.330031
0.304423
3.708078
0.002266
0.226586
0.248322
0
0.411177
10.030469
0.109458
1.050901
false
true
2024-10-31
2024-11-06
0
HuggingFaceTB/SmolLM2-135M
HuggingFaceTB_SmolLM2-135M-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-135M-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-135M-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-135M-Instruct
5a33ba103645800d7b3790c4448546c1b73efc71
6.467365
apache-2.0
72
0
true
false
false
true
0.338376
0.288314
28.83139
0.312432
4.720808
0.003021
0.302115
0.235738
0
0.366219
3.677344
0.111453
1.272533
false
true
2024-10-31
2024-11-06
1
HuggingFaceTB/SmolLM2-135M-Instruct (Merge)
HuggingFaceTB_SmolLM2-135M-Instruct_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-135M-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-135M-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-135M-Instruct
5a33ba103645800d7b3790c4448546c1b73efc71
2.992599
apache-2.0
72
0
true
false
false
false
0.348754
0.059252
5.925167
0.313475
4.796276
0.001511
0.151057
0.23406
0
0.387146
6.059896
0.109209
1.023197
false
true
2024-10-31
2024-11-14
1
HuggingFaceTB/SmolLM2-135M-Instruct (Merge)
HuggingFaceTB_SmolLM2-360M_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-360M" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-360M</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-360M-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-360M
3ce05f63c246c44616da500b47b01f082f4d3bcc
6.100225
apache-2.0
27
0
true
false
false
false
0.386658
0.211452
21.145228
0.323348
5.543603
0.003021
0.302115
0.245805
0
0.395427
7.728385
0.116938
1.882018
false
true
2024-10-31
2024-11-06
0
HuggingFaceTB/SmolLM2-360M
HuggingFaceTB_SmolLM2-360M-Instruct_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-360M-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-360M-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-360M-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-360M-Instruct
4873f67095301d304753fae05bc09ec766634e50
3.10002
apache-2.0
60
0
true
false
false
false
0.392382
0.083032
8.303191
0.30527
3.299047
0.008308
0.830816
0.265101
2.013423
0.342281
2.751823
0.112616
1.401817
false
true
2024-10-31
2024-11-14
0
HuggingFaceTB/SmolLM2-360M-Instruct
HuggingFaceTB_SmolLM2-360M-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HuggingFaceTB/SmolLM2-360M-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HuggingFaceTB/SmolLM2-360M-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HuggingFaceTB__SmolLM2-360M-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HuggingFaceTB/SmolLM2-360M-Instruct
4873f67095301d304753fae05bc09ec766634e50
8.001097
apache-2.0
60
0
true
false
false
true
0.375819
0.38416
38.415959
0.314351
4.173864
0.006798
0.679758
0.255034
0.671141
0.346125
2.765625
0.111702
1.300236
false
true
2024-10-31
2024-11-06
0
HuggingFaceTB/SmolLM2-360M-Instruct
HumanLLMs_Humanish-LLama3-8B-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/HumanLLMs/Humanish-LLama3-8B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HumanLLMs/Humanish-LLama3-8B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HumanLLMs__Humanish-LLama3-8B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HumanLLMs/Humanish-LLama3-8B-Instruct
42f73ada2b7fb16f18a75404d72b7911bf1e65ce
22.564911
llama3
2
8
true
false
false
true
0.748278
0.64979
64.979033
0.496771
28.012477
0.095921
9.592145
0.255872
0.782998
0.358156
2.002865
0.37018
30.019947
false
false
2024-10-04
2024-10-05
1
meta-llama/Meta-Llama-3-8B-Instruct
HumanLLMs_Humanish-Mistral-Nemo-Instruct-2407_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/HumanLLMs/Humanish-Mistral-Nemo-Instruct-2407" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HumanLLMs/Humanish-Mistral-Nemo-Instruct-2407</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HumanLLMs__Humanish-Mistral-Nemo-Instruct-2407-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HumanLLMs/Humanish-Mistral-Nemo-Instruct-2407
45b80bdce8d447ef494af06751904afcc607eb37
23.0069
apache-2.0
3
12
true
false
false
true
1.620283
0.545127
54.512693
0.526178
32.709613
0.083837
8.383686
0.287752
5.033557
0.39676
9.395052
0.352061
28.006797
false
false
2024-10-06
2024-10-06
2
mistralai/Mistral-Nemo-Base-2407
HumanLLMs_Humanish-Qwen2.5-7B-Instruct_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/HumanLLMs/Humanish-Qwen2.5-7B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">HumanLLMs/Humanish-Qwen2.5-7B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/HumanLLMs__Humanish-Qwen2.5-7B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
HumanLLMs/Humanish-Qwen2.5-7B-Instruct
7d2c71d926832d6e257ad2776011494dbac2d151
26.665374
apache-2.0
3
7
true
false
false
true
1.193393
0.728425
72.842502
0.536368
34.478998
0
0
0.298658
6.487696
0.398063
8.424479
0.439827
37.75857
false
false
2024-10-05
2024-10-05
2
Qwen/Qwen2.5-7B
IDEA-CCNL_Ziya-LLaMA-13B-v1_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/IDEA-CCNL/Ziya-LLaMA-13B-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">IDEA-CCNL/Ziya-LLaMA-13B-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/IDEA-CCNL__Ziya-LLaMA-13B-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
IDEA-CCNL/Ziya-LLaMA-13B-v1
64d931f346e1a49ea3bbca07a83137075bab1c66
3.906425
gpl-3.0
273
13
true
false
false
false
1.108257
0.169686
16.968643
0.287703
1.463617
0
0
0.249161
0
0.375052
3.88151
0.110123
1.124778
false
true
2023-05-16
2024-06-12
0
IDEA-CCNL/Ziya-LLaMA-13B-v1
Infinirc_Infinirc-Llama3-8B-2G-Release-v1.0_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Infinirc/Infinirc-Llama3-8B-2G-Release-v1.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Infinirc/Infinirc-Llama3-8B-2G-Release-v1.0</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Infinirc__Infinirc-Llama3-8B-2G-Release-v1.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Infinirc/Infinirc-Llama3-8B-2G-Release-v1.0
9c542d9ec3f86e145ae445c200c6ebe9066e8cd6
13.087133
llama3
1
8
true
false
false
false
1.818723
0.202434
20.243399
0.435074
20.831165
0.012085
1.208459
0.299497
6.599553
0.460938
16.750521
0.216007
12.889702
false
false
2024-06-26
2024-09-29
0
Infinirc/Infinirc-Llama3-8B-2G-Release-v1.0
Intel_neural-chat-7b-v3_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Intel/neural-chat-7b-v3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Intel/neural-chat-7b-v3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Intel__neural-chat-7b-v3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Intel/neural-chat-7b-v3
fc679274dfcd28a8b6087634f71af7ed2a0659c4
17.943646
apache-2.0
67
7
true
false
false
false
0.48929
0.277797
27.779736
0.504832
30.205692
0.021903
2.190332
0.291946
5.592841
0.50549
23.019531
0.269864
18.873744
false
true
2023-10-25
2024-06-12
1
mistralai/Mistral-7B-v0.1
Intel_neural-chat-7b-v3-1_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Intel/neural-chat-7b-v3-1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Intel/neural-chat-7b-v3-1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Intel__neural-chat-7b-v3-1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Intel/neural-chat-7b-v3-1
c0d379a49c1c0579529d5e6f2e936ddb759552a8
21.004986
apache-2.0
545
7
true
false
false
false
0.563692
0.46869
46.868974
0.505157
29.739752
0.031722
3.172205
0.290268
5.369128
0.497896
22.236979
0.267786
18.642878
false
true
2023-11-14
2024-06-12
1
mistralai/Mistral-7B-v0.1
Intel_neural-chat-7b-v3-2_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Intel/neural-chat-7b-v3-2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Intel/neural-chat-7b-v3-2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Intel__neural-chat-7b-v3-2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Intel/neural-chat-7b-v3-2
0d8f77647810d21d935ea90c66d6339b85e65a75
21.433647
apache-2.0
56
7
true
false
false
false
0.560441
0.49884
49.883975
0.503223
30.237458
0.045317
4.531722
0.290268
5.369128
0.489521
20.056771
0.266705
18.522828
false
true
2023-11-21
2024-06-12
0
Intel/neural-chat-7b-v3-2
Intel_neural-chat-7b-v3-3_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Intel/neural-chat-7b-v3-3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Intel/neural-chat-7b-v3-3</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Intel__neural-chat-7b-v3-3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Intel/neural-chat-7b-v3-3
bdd31cf498d13782cc7497cba5896996ce429f91
19.99112
apache-2.0
75
7
true
false
false
false
0.559524
0.476259
47.625855
0.487662
27.753851
0.006798
0.679758
0.28943
5.257271
0.485958
20.578125
0.262467
18.051862
false
true
2023-12-09
2024-06-12
2
mistralai/Mistral-7B-v0.1
IntervitensInc_internlm2_5-20b-llamafied_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/IntervitensInc/internlm2_5-20b-llamafied" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">IntervitensInc/internlm2_5-20b-llamafied</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/IntervitensInc__internlm2_5-20b-llamafied-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
IntervitensInc/internlm2_5-20b-llamafied
0b6fc3cc0b9bf3529816061eb508483c20b77fe9
29.204293
apache-2.0
2
19
true
false
false
false
1.381128
0.340995
34.099523
0.747847
63.47058
0.170695
17.069486
0.338087
11.744966
0.447542
14.942708
0.405086
33.898493
false
false
2024-08-06
2024-11-11
0
IntervitensInc/internlm2_5-20b-llamafied
Isaak-Carter_JOSIEv4o-8b-stage1-v4_float16
float16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Isaak-Carter/JOSIEv4o-8b-stage1-v4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Isaak-Carter/JOSIEv4o-8b-stage1-v4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Isaak-Carter__JOSIEv4o-8b-stage1-v4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Isaak-Carter/JOSIEv4o-8b-stage1-v4
a8380a7be51b547761824e524b3d95ac73203122
15.567377
apache-2.0
1
8
true
false
false
false
0.890582
0.255266
25.526603
0.472497
25.787276
0.046828
4.682779
0.291946
5.592841
0.365438
6.079687
0.331616
25.735077
false
false
2024-08-03
2024-08-03
0
Isaak-Carter/JOSIEv4o-8b-stage1-v4
Isaak-Carter_JOSIEv4o-8b-stage1-v4_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Isaak-Carter/JOSIEv4o-8b-stage1-v4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Isaak-Carter/JOSIEv4o-8b-stage1-v4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Isaak-Carter__JOSIEv4o-8b-stage1-v4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Isaak-Carter/JOSIEv4o-8b-stage1-v4
a8380a7be51b547761824e524b3d95ac73203122
15.419272
apache-2.0
1
8
true
false
false
false
0.879882
0.247697
24.769722
0.475807
25.919578
0.045317
4.531722
0.291107
5.480984
0.364104
6.346354
0.329205
25.467272
false
false
2024-08-03
2024-08-03
0
Isaak-Carter/JOSIEv4o-8b-stage1-v4
Isaak-Carter_Josiefied-Qwen2.5-7B-Instruct-abliterated_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Isaak-Carter__Josiefied-Qwen2.5-7B-Instruct-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated
879168f9ce9fac315a19dd4f4c7df5253bb660f2
26.857295
0
7
false
false
false
true
1.076791
0.731747
73.174732
0.539638
34.904316
0
0
0.302852
7.04698
0.408667
9.616667
0.42761
36.401079
false
false
2024-09-21
0
Removed
Isaak-Carter_Josiefied-Qwen2.5-7B-Instruct-abliterated-v2_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Isaak-Carter__Josiefied-Qwen2.5-7B-Instruct-abliterated-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2
5d07f58562422feb9f25c9c048e40356d2cf7e4b
27.81796
apache-2.0
4
7
true
false
false
true
1.130915
0.784104
78.410396
0.531092
33.29454
0
0
0.298658
6.487696
0.435396
13.957813
0.412816
34.757314
false
false
2024-09-20
2024-09-21
1
Qwen/Qwen2.5-7B
J-LAB_Thynk_orpo_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/J-LAB/Thynk_orpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">J-LAB/Thynk_orpo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/J-LAB__Thynk_orpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
J-LAB/Thynk_orpo
c6606d402f26d005b9f1a71a1cde9139d1cffb2a
16.974407
0
3
false
false
false
false
1.214764
0.210178
21.017788
0.446311
22.062784
0.130665
13.066465
0.292785
5.704698
0.451479
15.201563
0.323138
24.793144
false
false
2024-10-14
0
Removed
JackFram_llama-160m_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/JackFram/llama-160m" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">JackFram/llama-160m</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/JackFram__llama-160m-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
JackFram/llama-160m
aca9b687d1425f863dcf5de9a4c96e3fe36266dd
4.599661
apache-2.0
34
0
true
false
false
false
0.093474
0.179104
17.910367
0.288802
2.033606
0
0
0.261745
1.565996
0.379208
4.667708
0.112783
1.420287
false
false
2023-05-26
2024-11-30
0
JackFram/llama-160m
JackFram_llama-68m_bfloat16
bfloat16
🟢 pretrained
🟢
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/JackFram/llama-68m" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">JackFram/llama-68m</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/JackFram__llama-68m-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
JackFram/llama-68m
964a5d77df908b69f8d6476fb70e940425b04cb5
4.862635
apache-2.0
25
0
true
false
false
false
0.060558
0.172634
17.263417
0.29363
2.591048
0
0
0.258389
1.118568
0.39099
6.607031
0.114362
1.595745
false
false
2023-07-19
2024-11-30
0
JackFram/llama-68m
Jacoby746_Casual-Magnum-34B_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Casual-Magnum-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Casual-Magnum-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Casual-Magnum-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Casual-Magnum-34B
b628c6959441db75460cfd49536322b1ea46130e
23.571335
apache-2.0
1
34
true
false
false
false
3.426697
0.193017
19.301675
0.603205
43.051568
0.07855
7.854985
0.372483
16.331096
0.40776
8.403385
0.518368
46.485298
true
false
2024-10-01
2024-10-23
1
Jacoby746/Casual-Magnum-34B (Merge)
Jacoby746_Inf-Silent-Kunoichi-v0.1-2x7B_float16
float16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Inf-Silent-Kunoichi-v0.1-2x7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Inf-Silent-Kunoichi-v0.1-2x7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Inf-Silent-Kunoichi-v0.1-2x7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Inf-Silent-Kunoichi-v0.1-2x7B
9ab68beb6fe16cab2ab708b9af4417c89751d297
20.009948
apache-2.0
0
12
true
false
false
false
1.860053
0.387982
38.798167
0.518546
32.387004
0.060423
6.042296
0.28943
5.257271
0.428042
12.338542
0.327128
25.236407
false
false
2024-09-19
2024-09-20
1
Jacoby746/Inf-Silent-Kunoichi-v0.1-2x7B (Merge)
Jacoby746_Inf-Silent-Kunoichi-v0.2-2x7B_float16
float16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Inf-Silent-Kunoichi-v0.2-2x7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Inf-Silent-Kunoichi-v0.2-2x7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Inf-Silent-Kunoichi-v0.2-2x7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Inf-Silent-Kunoichi-v0.2-2x7B
711263c24f812676eb382a31a5f0fed9bd8c16e4
19.917523
apache-2.0
0
12
true
false
false
false
0.866265
0.363602
36.360191
0.520942
32.259184
0.056647
5.664653
0.300336
6.711409
0.431979
13.264062
0.327211
25.245641
false
false
2024-09-19
2024-09-21
1
Jacoby746/Inf-Silent-Kunoichi-v0.2-2x7B (Merge)
Jacoby746_Proto-Athena-4x7B_float16
float16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Proto-Athena-4x7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Proto-Athena-4x7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Proto-Athena-4x7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Proto-Athena-4x7B
450fcba7a630fb61a662f71936d37979226fced8
19.649696
apache-2.0
0
24
true
false
false
false
1.676614
0.370296
37.029637
0.510655
30.870823
0.057402
5.740181
0.294463
5.928412
0.434771
13.813021
0.320645
24.516105
false
false
2024-09-21
2024-09-21
1
Jacoby746/Proto-Athena-4x7B (Merge)
Jacoby746_Proto-Athena-v0.2-4x7B_bfloat16
bfloat16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Proto-Athena-v0.2-4x7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Proto-Athena-v0.2-4x7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Proto-Athena-v0.2-4x7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Proto-Athena-v0.2-4x7B
01feeded217ea83a8794e7968c8850859b5f0b14
19.143898
apache-2.0
0
24
true
false
false
false
1.651372
0.375242
37.524214
0.506773
30.340844
0.05136
5.135952
0.298658
6.487696
0.421281
10.960156
0.319731
24.414524
false
false
2024-09-21
2024-09-21
1
Jacoby746/Proto-Athena-v0.2-4x7B (Merge)
Jacoby746_Proto-Harpy-Blazing-Light-v0.1-2x7B_float16
float16
🤝 base merges and moerges
🤝
Original
MixtralForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Proto-Harpy-Blazing-Light-v0.1-2x7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Proto-Harpy-Blazing-Light-v0.1-2x7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Proto-Harpy-Blazing-Light-v0.1-2x7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Proto-Harpy-Blazing-Light-v0.1-2x7B
bbb5d7c7a0c9e999e057ffa71eaa93d59d95b36b
22.292392
0
12
false
false
false
false
0.881841
0.490472
49.047195
0.518685
32.63253
0.063444
6.344411
0.295302
6.040268
0.444969
14.121094
0.33012
25.568853
false
false
2024-09-22
2024-09-30
1
Jacoby746/Proto-Harpy-Blazing-Light-v0.1-2x7B (Merge)
Jacoby746_Proto-Harpy-Spark-v0.1-7B_float16
float16
🤝 base merges and moerges
🤝
Original
MistralForCausalLM
<a target="_blank" href="https://huggingface.co/Jacoby746/Proto-Harpy-Spark-v0.1-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jacoby746/Proto-Harpy-Spark-v0.1-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jacoby746__Proto-Harpy-Spark-v0.1-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jacoby746/Proto-Harpy-Spark-v0.1-7B
984cca02cd930b2e1b7b2a7d53471d32d9821cdd
19.862588
apache-2.0
0
7
true
false
false
false
0.595805
0.433269
43.326928
0.473577
26.91311
0.062689
6.268882
0.305369
7.38255
0.431667
12.291667
0.306932
22.992391
true
false
2024-09-22
2024-09-30
1
Jacoby746/Proto-Harpy-Spark-v0.1-7B (Merge)
Jimmy19991222_Llama-3-Instruct-8B-SimPO-v0.2_float16
float16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/Llama-3-Instruct-8B-SimPO-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/Llama-3-Instruct-8B-SimPO-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__Llama-3-Instruct-8B-SimPO-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/Llama-3-Instruct-8B-SimPO-v0.2
53a517ceaef324efc3626be44140b4f18a010591
24.279948
0
8
false
false
false
true
0.513152
0.654037
65.403684
0.498371
29.123823
0.043051
4.305136
0.314597
8.612975
0.40125
8.389583
0.3686
29.844489
false
false
2024-09-06
0
Removed
Jimmy19991222_llama-3-8b-instruct-gapo-v2-bert-f1-beta10-gamma0.3-lr1.0e-6-1minus-rerun_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert-f1-beta10-gamma0.3-lr1.0e-6-1minus-rerun" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert-f1-beta10-gamma0.3-lr1.0e-6-1minus-rerun</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-bert-f1-beta10-gamma0.3-lr1.0e-6-1minus-rerun-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert-f1-beta10-gamma0.3-lr1.0e-6-1minus-rerun
00c02a823b4ff1a6cfcded6085ba9630df633998
23.817704
llama3
0
8
true
false
false
true
0.481791
0.671722
67.172214
0.48798
27.755229
0.040785
4.07855
0.294463
5.928412
0.404073
8.709115
0.363364
29.262707
false
false
2024-09-17
2024-09-18
1
meta-llama/Meta-Llama-3-8B-Instruct
Jimmy19991222_llama-3-8b-instruct-gapo-v2-bert_f1-beta10-gamma0.3-lr1.0e-6-scale-log_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert_f1-beta10-gamma0.3-lr1.0e-6-scale-log" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert_f1-beta10-gamma0.3-lr1.0e-6-scale-log</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-bert_f1-beta10-gamma0.3-lr1.0e-6-scale-log-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert_f1-beta10-gamma0.3-lr1.0e-6-scale-log
99d9e31df5b7e88b1da78b1bd335cac3215dfd6e
23.75627
llama3
0
8
true
false
false
true
0.478535
0.655561
65.556058
0.493458
28.613597
0.033988
3.398792
0.30453
7.270694
0.40001
8.167969
0.365775
29.530511
false
false
2024-09-22
2024-09-22
1
meta-llama/Meta-Llama-3-8B-Instruct
Jimmy19991222_llama-3-8b-instruct-gapo-v2-bert_p-beta10-gamma0.3-lr1.0e-6-scale-log_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert_p-beta10-gamma0.3-lr1.0e-6-scale-log" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert_p-beta10-gamma0.3-lr1.0e-6-scale-log</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-bert_p-beta10-gamma0.3-lr1.0e-6-scale-log-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-bert_p-beta10-gamma0.3-lr1.0e-6-scale-log
49a029ea2605d768e89b638ad78a59fd62d192ab
22.797979
llama3
0
8
true
false
false
true
0.522485
0.631506
63.150552
0.491641
27.666184
0.050604
5.060423
0.286074
4.809843
0.3935
7.0875
0.36112
29.013372
false
false
2024-09-22
2024-09-22
1
meta-llama/Meta-Llama-3-8B-Instruct
Jimmy19991222_llama-3-8b-instruct-gapo-v2-bleu-beta0.1-no-length-scale-gamma0.4_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-bleu-beta0.1-no-length-scale-gamma0.4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-bleu-beta0.1-no-length-scale-gamma0.4</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-bleu-beta0.1-no-length-scale-gamma0.4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-bleu-beta0.1-no-length-scale-gamma0.4
de8bb28ad7a9d1158f318a4461dc47ad03e6e560
22.827312
0
8
false
false
false
true
0.480371
0.628458
62.845805
0.498609
29.329732
0.017372
1.73716
0.292785
5.704698
0.401375
9.071875
0.354471
28.274601
false
false
2024-09-06
0
Removed
Jimmy19991222_llama-3-8b-instruct-gapo-v2-rouge2-beta10-1minus-gamma0.3-rerun_bfloat16
bfloat16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-rouge2-beta10-1minus-gamma0.3-rerun" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-rouge2-beta10-1minus-gamma0.3-rerun</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-rouge2-beta10-1minus-gamma0.3-rerun-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-rouge2-beta10-1minus-gamma0.3-rerun
e9692d8dbe30273839763757aa9ef07a5fcf0c59
24.159026
llama3
0
8
true
false
false
true
1.009359
0.66775
66.775046
0.494046
28.390676
0.047583
4.758308
0.306208
7.494407
0.398708
8.005208
0.365775
29.530511
false
false
2024-09-14
2024-09-15
1
meta-llama/Meta-Llama-3-8B-Instruct
Jimmy19991222_llama-3-8b-instruct-gapo-v2-rouge2-beta10-gamma0.3-lr1.0e-6-scale-log_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-rouge2-beta10-gamma0.3-lr1.0e-6-scale-log" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-rouge2-beta10-gamma0.3-lr1.0e-6-scale-log</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-rouge2-beta10-gamma0.3-lr1.0e-6-scale-log-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-rouge2-beta10-gamma0.3-lr1.0e-6-scale-log
9ff0ce408abb8dbcf7efb9b6533338f2c344a355
23.858383
llama3
0
8
true
false
false
true
0.501994
0.660506
66.050635
0.491601
28.075036
0.044562
4.456193
0.303691
7.158837
0.400042
7.805208
0.366439
29.604388
false
false
2024-09-22
2024-09-22
1
meta-llama/Meta-Llama-3-8B-Instruct
Jimmy19991222_llama-3-8b-instruct-gapo-v2-rougeL-beta10-gamma0.3-lr1.0e-6-scale-log_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Jimmy19991222/llama-3-8b-instruct-gapo-v2-rougeL-beta10-gamma0.3-lr1.0e-6-scale-log" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Jimmy19991222/llama-3-8b-instruct-gapo-v2-rougeL-beta10-gamma0.3-lr1.0e-6-scale-log</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Jimmy19991222__llama-3-8b-instruct-gapo-v2-rougeL-beta10-gamma0.3-lr1.0e-6-scale-log-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Jimmy19991222/llama-3-8b-instruct-gapo-v2-rougeL-beta10-gamma0.3-lr1.0e-6-scale-log
ec67f95c4d1813a34bbde52d0ad14824fd7111a0
23.742269
llama3
0
8
true
false
false
true
0.486586
0.649191
64.919081
0.495249
28.562567
0.045317
4.531722
0.302013
6.935123
0.396135
7.383594
0.371094
30.121528
false
false
2024-09-22
2024-09-22
1
meta-llama/Meta-Llama-3-8B-Instruct
Joseph717171_Hermes-3-Llama-3.1-8B_TIES_with_Base_Embeds_Initialized_to_Special_Instruct_Toks_dtypeF32_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Joseph717171/Hermes-3-Llama-3.1-8B_TIES_with_Base_Embeds_Initialized_to_Special_Instruct_Toks_dtypeF32" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Joseph717171/Hermes-3-Llama-3.1-8B_TIES_with_Base_Embeds_Initialized_to_Special_Instruct_Toks_dtypeF32</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Joseph717171__Hermes-3-Llama-3.1-8B_TIES_with_Base_Embeds_Initialized_to_Special_Instruct_Toks_dtypeF32-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Joseph717171/Hermes-3-Llama-3.1-8B_TIES_with_Base_Embeds_Initialized_to_Special_Instruct_Toks_dtypeF32
823930851c57b11fd2e25cd65b5c53f909209d0e
23.252877
llama3.1
1
8
true
false
false
true
0.707545
0.618541
61.854103
0.517745
30.724097
0.05136
5.135952
0.282718
4.362416
0.436938
13.617187
0.314412
23.823508
true
false
2024-10-23
2024-10-25
0
Joseph717171/Hermes-3-Llama-3.1-8B_TIES_with_Base_Embeds_Initialized_to_Special_Instruct_Toks_dtypeF32
Joseph717171_Llama-3.1-SuperNova-8B-Lite_TIES_with_Base_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Joseph717171/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Joseph717171/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Joseph717171__Llama-3.1-SuperNova-8B-Lite_TIES_with_Base-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Joseph717171/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base
f1e2cad4dca10f948fd2ee9588f80df0b40d7232
30.081383
llama3.1
8
8
true
false
false
true
0.874731
0.809633
80.963289
0.514742
31.465813
0.173716
17.371601
0.309564
7.941834
0.41099
10.740365
0.388049
32.005393
true
false
2024-10-02
2024-10-03
0
Joseph717171/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base
Josephgflowers_Cinder-Phi-2-V1-F16-gguf_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
PhiForCausalLM
<a target="_blank" href="https://huggingface.co/Josephgflowers/Cinder-Phi-2-V1-F16-gguf" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Josephgflowers/Cinder-Phi-2-V1-F16-gguf</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Josephgflowers__Cinder-Phi-2-V1-F16-gguf-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Josephgflowers/Cinder-Phi-2-V1-F16-gguf
85629ec9b18efee31d07630664e7a3815121badf
10.855703
mit
4
2
true
false
false
true
0.471404
0.235657
23.565695
0.439662
22.453402
0
0
0.281879
4.250559
0.343458
1.965625
0.21609
12.898936
false
false
2024-02-25
2024-06-26
0
Josephgflowers/Cinder-Phi-2-V1-F16-gguf
Josephgflowers_Differential-Attention-Liquid-Metal-Tinyllama_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Josephgflowers/Differential-Attention-Liquid-Metal-Tinyllama" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Josephgflowers/Differential-Attention-Liquid-Metal-Tinyllama</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Josephgflowers__Differential-Attention-Liquid-Metal-Tinyllama-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Josephgflowers/Differential-Attention-Liquid-Metal-Tinyllama
bdb6c63ff1025241e8e10b1858d67dc410f0a702
4.709671
mit
0
1
true
false
false
true
0.173794
0.222692
22.269246
0.292556
2.552224
0
0
0.250839
0.111857
0.335552
0.94401
0.121426
2.380689
false
false
2024-11-05
2024-11-07
0
Josephgflowers/Differential-Attention-Liquid-Metal-Tinyllama
Josephgflowers_TinyLlama-Cinder-Agent-v1_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Josephgflowers/TinyLlama-Cinder-Agent-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Josephgflowers/TinyLlama-Cinder-Agent-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Josephgflowers__TinyLlama-Cinder-Agent-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Josephgflowers/TinyLlama-Cinder-Agent-v1
a9cd8b48bfe30f29bb1f819213da9a4c41eee67f
5.816564
mit
1
1
true
false
false
true
0.237832
0.266956
26.695612
0.311604
3.804167
0.003776
0.377644
0.244128
0
0.339458
2.232292
0.116107
1.789672
false
false
2024-05-21
2024-06-26
4
Josephgflowers/TinyLlama-3T-Cinder-v1.2
Josephgflowers_TinyLlama-v1.1-Cinders-World_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Josephgflowers/TinyLlama-v1.1-Cinders-World" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Josephgflowers/TinyLlama-v1.1-Cinders-World</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Josephgflowers__TinyLlama-v1.1-Cinders-World-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Josephgflowers/TinyLlama-v1.1-Cinders-World
11a2c305f787a7908dd87c4e5a7d0f1e314a1f05
5.129125
mit
0
1
true
false
false
true
0.257383
0.246923
24.692261
0.299797
3.107714
0.001511
0.151057
0.244128
0
0.335615
0.61849
0.119847
2.20523
false
false
2024-10-12
2024-10-13
0
Josephgflowers/TinyLlama-v1.1-Cinders-World
Josephgflowers_TinyLlama_v1.1_math_code-world-test-1_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Josephgflowers/TinyLlama_v1.1_math_code-world-test-1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Josephgflowers/TinyLlama_v1.1_math_code-world-test-1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Josephgflowers__TinyLlama_v1.1_math_code-world-test-1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Josephgflowers/TinyLlama_v1.1_math_code-world-test-1
6f7c2aaf0b8723bc6a1dc23a4a1ff0ec24dc11ec
1.839166
mit
0
1
true
false
false
false
0.272944
0.007844
0.784363
0.314635
4.164017
0.009819
0.981873
0.23406
0
0.349906
3.638281
0.113198
1.46646
false
false
2024-06-23
2024-09-09
0
Josephgflowers/TinyLlama_v1.1_math_code-world-test-1
Josephgflowers_Tinyllama-STEM-Cinder-Agent-v1_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/Josephgflowers/Tinyllama-STEM-Cinder-Agent-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Josephgflowers/Tinyllama-STEM-Cinder-Agent-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Josephgflowers__Tinyllama-STEM-Cinder-Agent-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Josephgflowers/Tinyllama-STEM-Cinder-Agent-v1
c6880b94e72dddbe591fdf30fa15fe42ea60b924
4.575881
mit
0
1
true
false
false
true
0.163386
0.212576
21.257597
0.308438
3.731313
0.000755
0.075529
0.234899
0
0.334125
1.432292
0.108627
0.958555
false
false
2024-11-27
2024-11-27
0
Josephgflowers/Tinyllama-STEM-Cinder-Agent-v1
Junhoee_Qwen-Megumin_float16
float16
💬 chat models (RLHF, DPO, IFT, ...)
💬
Adapter
?
<a target="_blank" href="https://huggingface.co/Junhoee/Qwen-Megumin" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Junhoee/Qwen-Megumin</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Junhoee__Qwen-Megumin-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Junhoee/Qwen-Megumin
bb46c15ee4bb56c5b63245ef50fd7637234d6f75
25.822493
0
15
false
false
false
true
1.889396
0.714112
71.411189
0.528527
33.642144
0
0
0.296141
6.152125
0.398031
8.18724
0.41988
35.542258
false
false
2024-11-26
2024-11-26
2
Qwen/Qwen2.5-7B
KSU-HW-SEC_Llama3-70b-SVA-FT-1415_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/KSU-HW-SEC/Llama3-70b-SVA-FT-1415" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">KSU-HW-SEC/Llama3-70b-SVA-FT-1415</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/KSU-HW-SEC__Llama3-70b-SVA-FT-1415-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
KSU-HW-SEC/Llama3-70b-SVA-FT-1415
1c09728455567898116d2d9cfb6cbbbbd4ee730c
36.119233
0
70
false
false
false
false
9.601029
0.617991
61.799137
0.665015
51.328741
0.219789
21.978852
0.375
16.666667
0.456542
17.801042
0.524269
47.140957
false
false
2024-09-08
0
Removed
KSU-HW-SEC_Llama3-70b-SVA-FT-500_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/KSU-HW-SEC/Llama3-70b-SVA-FT-500" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">KSU-HW-SEC/Llama3-70b-SVA-FT-500</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/KSU-HW-SEC__Llama3-70b-SVA-FT-500-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
KSU-HW-SEC/Llama3-70b-SVA-FT-500
856a23f28aeada23d1135c86a37e05524307e8ed
35.953712
0
70
false
false
false
false
9.473738
0.610522
61.05223
0.669224
51.887026
0.213746
21.374622
0.380872
17.449664
0.451146
16.993229
0.522689
46.965499
false
false
2024-09-08
0
Removed
KSU-HW-SEC_Llama3-70b-SVA-FT-final_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/KSU-HW-SEC/Llama3-70b-SVA-FT-final" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">KSU-HW-SEC/Llama3-70b-SVA-FT-final</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/KSU-HW-SEC__Llama3-70b-SVA-FT-final-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
KSU-HW-SEC/Llama3-70b-SVA-FT-final
391bbd94173b34975d1aa2c7356977a630253b75
36.093837
0
70
false
false
false
false
9.656199
0.616468
61.646764
0.665015
51.328741
0.219789
21.978852
0.375
16.666667
0.456542
17.801042
0.524269
47.140957
false
false
2024-09-08
0
Removed
KSU-HW-SEC_Llama3.1-70b-SVA-FT-1000step_bfloat16
bfloat16
🔶 fine-tuned on domain-specific datasets
🔶
Original
LlamaForCausalLM
<a target="_blank" href="https://huggingface.co/KSU-HW-SEC/Llama3.1-70b-SVA-FT-1000step" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">KSU-HW-SEC/Llama3.1-70b-SVA-FT-1000step</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/KSU-HW-SEC__Llama3.1-70b-SVA-FT-1000step-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
KSU-HW-SEC/Llama3.1-70b-SVA-FT-1000step
b195fea0d8f350ff29243d4e88654b1baa5af79e
40.750259
0
70
false
false
false
false
12.554447
0.723804
72.380395
0.690312
55.485365
0.320997
32.099698
0.395973
19.463087
0.459177
17.830469
0.525183
47.242538
false
false
2024-09-08
0
Removed
Kimargin_GPT-NEO-1.3B-wiki_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
GPTNeoForCausalLM
<a target="_blank" href="https://huggingface.co/Kimargin/GPT-NEO-1.3B-wiki" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Kimargin/GPT-NEO-1.3B-wiki</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Kimargin__GPT-NEO-1.3B-wiki-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
Kimargin/GPT-NEO-1.3B-wiki
92fa51fa6589f6e8fdfcc83f085216b3dae11da5
5.248478
apache-2.0
1
1
true
false
false
false
0.832745
0.192068
19.206816
0.302634
3.423612
0.008308
0.830816
0.244966
0
0.38826
6.932552
0.109874
1.097074
false
false
2024-10-23
2024-10-24
1
Kimargin/GPT-NEO-1.3B-wiki (Merge)
KingNish_Qwen2.5-0.5b-Test-ft_float16
float16
🔶 fine-tuned on domain-specific datasets
🔶
Original
Qwen2ForCausalLM
<a target="_blank" href="https://huggingface.co/KingNish/Qwen2.5-0.5b-Test-ft" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">KingNish/Qwen2.5-0.5b-Test-ft</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/KingNish__Qwen2.5-0.5b-Test-ft-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
KingNish/Qwen2.5-0.5b-Test-ft
f905bb1d37c7853fb5c7157d8d3ad0f062b65c0f
7.475184
apache-2.0
5
0
true
false
false
false
0.66869
0.267081
26.708134
0.323153
6.058845
0.012085
1.208459
0.263423
1.789709
0.342125
1.432292
0.168883
7.653664
false
false
2024-09-26
2024-09-29
1
KingNish/Qwen2.5-0.5b-Test-ft (Merge)