Novaciano commited on
Commit
ddd59db
·
verified ·
1 Parent(s): 59c2c68

Adding Evaluation Results (#1)

Browse files

- Adding Evaluation Results (18ab4ded488b2a14464248caaa97ae66e33c1f4f)

Files changed (1) hide show
  1. README.md +114 -1
README.md CHANGED
@@ -6,7 +6,105 @@ library_name: transformers
6
  tags:
7
  - mergekit
8
  - merge
9
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
  # merge
12
 
@@ -38,3 +136,18 @@ parameters:
38
  t: [0, 0.5, 1, 0.5, 0]
39
 
40
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  tags:
7
  - mergekit
8
  - merge
9
+ model-index:
10
+ - name: Cultist-3.2-1B
11
+ results:
12
+ - task:
13
+ type: text-generation
14
+ name: Text Generation
15
+ dataset:
16
+ name: IFEval (0-Shot)
17
+ type: wis-k/instruction-following-eval
18
+ split: train
19
+ args:
20
+ num_few_shot: 0
21
+ metrics:
22
+ - type: inst_level_strict_acc and prompt_level_strict_acc
23
+ value: 52.95
24
+ name: averaged accuracy
25
+ source:
26
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Novaciano%2FCultist-3.2-1B
27
+ name: Open LLM Leaderboard
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: BBH (3-Shot)
33
+ type: SaylorTwift/bbh
34
+ split: test
35
+ args:
36
+ num_few_shot: 3
37
+ metrics:
38
+ - type: acc_norm
39
+ value: 7.52
40
+ name: normalized accuracy
41
+ source:
42
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Novaciano%2FCultist-3.2-1B
43
+ name: Open LLM Leaderboard
44
+ - task:
45
+ type: text-generation
46
+ name: Text Generation
47
+ dataset:
48
+ name: MATH Lvl 5 (4-Shot)
49
+ type: lighteval/MATH-Hard
50
+ split: test
51
+ args:
52
+ num_few_shot: 4
53
+ metrics:
54
+ - type: exact_match
55
+ value: 5.89
56
+ name: exact match
57
+ source:
58
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Novaciano%2FCultist-3.2-1B
59
+ name: Open LLM Leaderboard
60
+ - task:
61
+ type: text-generation
62
+ name: Text Generation
63
+ dataset:
64
+ name: GPQA (0-shot)
65
+ type: Idavidrein/gpqa
66
+ split: train
67
+ args:
68
+ num_few_shot: 0
69
+ metrics:
70
+ - type: acc_norm
71
+ value: 1.45
72
+ name: acc_norm
73
+ source:
74
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Novaciano%2FCultist-3.2-1B
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: MuSR (0-shot)
81
+ type: TAUR-Lab/MuSR
82
+ args:
83
+ num_few_shot: 0
84
+ metrics:
85
+ - type: acc_norm
86
+ value: 1.89
87
+ name: acc_norm
88
+ source:
89
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Novaciano%2FCultist-3.2-1B
90
+ name: Open LLM Leaderboard
91
+ - task:
92
+ type: text-generation
93
+ name: Text Generation
94
+ dataset:
95
+ name: MMLU-PRO (5-shot)
96
+ type: TIGER-Lab/MMLU-Pro
97
+ config: main
98
+ split: test
99
+ args:
100
+ num_few_shot: 5
101
+ metrics:
102
+ - type: acc
103
+ value: 7.93
104
+ name: accuracy
105
+ source:
106
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Novaciano%2FCultist-3.2-1B
107
+ name: Open LLM Leaderboard
108
  ---
109
  # merge
110
 
 
136
  t: [0, 0.5, 1, 0.5, 0]
137
 
138
  ```
139
+
140
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
141
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/Novaciano__Cultist-3.2-1B-details)!
142
+ Summarized results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/contents/viewer/default/train?q=Novaciano%2FCultist-3.2-1B&sort[column]=Average%20%E2%AC%86%EF%B8%8F&sort[direction]=desc)!
143
+
144
+ | Metric |Value (%)|
145
+ |-------------------|--------:|
146
+ |**Average** | 12.94|
147
+ |IFEval (0-Shot) | 52.95|
148
+ |BBH (3-Shot) | 7.52|
149
+ |MATH Lvl 5 (4-Shot)| 5.89|
150
+ |GPQA (0-shot) | 1.45|
151
+ |MuSR (0-shot) | 1.89|
152
+ |MMLU-PRO (5-shot) | 7.93|
153
+