Spaces:
Running
Running
GECTurk model response white space problem fix
Browse files- results/zero-shot/CerebrumTech__cere-llama-3-8b-tr.json +1 -1
- results/zero-shot/Llama-3.3-70B-Instruct.json +1 -1
- results/zero-shot/Ministral-8B-Instruct.json +1 -1
- results/zero-shot/Mixtral-8x7B-Instruct-v0.1.json +1 -1
- results/zero-shot/Qwen2.5-0.5B-Instruct.json +1 -1
- results/zero-shot/Qwen2.5-0.5B.json +1 -1
- results/zero-shot/Qwen2.5-1.5B-Instruct.json +1 -1
- results/zero-shot/Qwen2.5-1.5B.json +1 -1
- results/zero-shot/Qwen2.5-14B-Instruct.json +1 -1
- results/zero-shot/Qwen2.5-14B.json +1 -1
- results/zero-shot/Qwen2.5-3B-Instruct.json +1 -1
- results/zero-shot/Qwen2.5-3B.json +1 -1
- results/zero-shot/Qwen2.5-7B-Instruct.json +1 -1
- results/zero-shot/Qwen2.5-7B.json +1 -1
- results/zero-shot/aya-23-35B.json +1 -1
- results/zero-shot/aya-23-8b.json +1 -1
- results/zero-shot/aya-expanse-32b.json +1 -1
- results/zero-shot/aya-expanse-8b.json +1 -1
- results/zero-shot/aya101.json +1 -1
- results/zero-shot/commencis-7b.json +1 -1
- results/zero-shot/kanarya-2b.json +1 -1
- results/zero-shot/llama-3-8b-instruct.json +1 -1
- results/zero-shot/llama-3-8b.json +1 -1
- results/zero-shot/llama-3.1-8b-instruct.json +1 -1
- results/zero-shot/llama-3.1-8b.json +1 -1
- results/zero-shot/llama-3.2-1b.json +1 -1
- results/zero-shot/llama-3.2-3b-instruct.json +1 -1
- results/zero-shot/llama-3.2-3b.json +1 -1
- results/zero-shot/mistral-7b.json +1 -1
- results/zero-shot/trendyol-7b.json +1 -1
results/zero-shot/CerebrumTech__cere-llama-3-8b-tr.json
CHANGED
|
@@ -47,7 +47,7 @@
|
|
| 47 |
{
|
| 48 |
"name": "gecturk_generation",
|
| 49 |
"task": "grammatical_error_correction",
|
| 50 |
-
"exact_match": 0.
|
| 51 |
},
|
| 52 |
{
|
| 53 |
"name": "xquad_tr",
|
|
|
|
| 47 |
{
|
| 48 |
"name": "gecturk_generation",
|
| 49 |
"task": "grammatical_error_correction",
|
| 50 |
+
"exact_match": 0.4596
|
| 51 |
},
|
| 52 |
{
|
| 53 |
"name": "xquad_tr",
|
results/zero-shot/Llama-3.3-70B-Instruct.json
CHANGED
|
@@ -162,7 +162,7 @@
|
|
| 162 |
{
|
| 163 |
"name": "gecturk_generation",
|
| 164 |
"task": "grammatical_error_correction",
|
| 165 |
-
"exact_match": 0.
|
| 166 |
},
|
| 167 |
{
|
| 168 |
"name": "mlsum_tr",
|
|
|
|
| 162 |
{
|
| 163 |
"name": "gecturk_generation",
|
| 164 |
"task": "grammatical_error_correction",
|
| 165 |
+
"exact_match": 0.441
|
| 166 |
},
|
| 167 |
{
|
| 168 |
"name": "mlsum_tr",
|
results/zero-shot/Ministral-8B-Instruct.json
CHANGED
|
@@ -166,7 +166,7 @@
|
|
| 166 |
{
|
| 167 |
"name": "gecturk_generation",
|
| 168 |
"task": "grammatical_error_correction",
|
| 169 |
-
"exact_match": 0.
|
| 170 |
},
|
| 171 |
{
|
| 172 |
"name": "turkce_atasozleri",
|
|
|
|
| 166 |
{
|
| 167 |
"name": "gecturk_generation",
|
| 168 |
"task": "grammatical_error_correction",
|
| 169 |
+
"exact_match": 0.3915
|
| 170 |
},
|
| 171 |
{
|
| 172 |
"name": "turkce_atasozleri",
|
results/zero-shot/Mixtral-8x7B-Instruct-v0.1.json
CHANGED
|
@@ -168,7 +168,7 @@
|
|
| 168 |
{
|
| 169 |
"name": "gecturk_generation",
|
| 170 |
"task": "grammatical_error_correction",
|
| 171 |
-
"exact_match": 0.
|
| 172 |
},
|
| 173 |
{
|
| 174 |
"name": "turkce_atasozleri",
|
|
|
|
| 168 |
{
|
| 169 |
"name": "gecturk_generation",
|
| 170 |
"task": "grammatical_error_correction",
|
| 171 |
+
"exact_match": 0.0364
|
| 172 |
},
|
| 173 |
{
|
| 174 |
"name": "turkce_atasozleri",
|
results/zero-shot/Qwen2.5-0.5B-Instruct.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.0177
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-0.5B.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.0092
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-1.5B-Instruct.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.0361
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-1.5B.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.1417
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-14B-Instruct.json
CHANGED
|
@@ -168,7 +168,7 @@
|
|
| 168 |
{
|
| 169 |
"name": "gecturk_generation",
|
| 170 |
"task": "grammatical_error_correction",
|
| 171 |
-
"exact_match": 0.
|
| 172 |
},
|
| 173 |
{
|
| 174 |
"name": "turkce_atasozleri",
|
|
|
|
| 168 |
{
|
| 169 |
"name": "gecturk_generation",
|
| 170 |
"task": "grammatical_error_correction",
|
| 171 |
+
"exact_match": 0.1088
|
| 172 |
},
|
| 173 |
{
|
| 174 |
"name": "turkce_atasozleri",
|
results/zero-shot/Qwen2.5-14B.json
CHANGED
|
@@ -168,7 +168,7 @@
|
|
| 168 |
{
|
| 169 |
"name": "gecturk_generation",
|
| 170 |
"task": "grammatical_error_correction",
|
| 171 |
-
"exact_match": 0.
|
| 172 |
},
|
| 173 |
{
|
| 174 |
"name": "turkce_atasozleri",
|
|
|
|
| 168 |
{
|
| 169 |
"name": "gecturk_generation",
|
| 170 |
"task": "grammatical_error_correction",
|
| 171 |
+
"exact_match": 0.1888
|
| 172 |
},
|
| 173 |
{
|
| 174 |
"name": "turkce_atasozleri",
|
results/zero-shot/Qwen2.5-3B-Instruct.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.0406
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-3B.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.1466
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-7B-Instruct.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.0661
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/Qwen2.5-7B.json
CHANGED
|
@@ -46,7 +46,7 @@
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
-
"exact_match": 0.
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
|
|
|
| 46 |
{
|
| 47 |
"name": "gecturk_generation",
|
| 48 |
"task": "grammatical_error_correction",
|
| 49 |
+
"exact_match": 0.2227
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "xquad_tr",
|
results/zero-shot/aya-23-35B.json
CHANGED
|
@@ -134,7 +134,7 @@
|
|
| 134 |
{
|
| 135 |
"name": "gecturk_generation",
|
| 136 |
"task": "grammatical_error_correction",
|
| 137 |
-
"exact_match": 0.
|
| 138 |
},
|
| 139 |
{
|
| 140 |
"name": "xlsum_tr",
|
|
|
|
| 134 |
{
|
| 135 |
"name": "gecturk_generation",
|
| 136 |
"task": "grammatical_error_correction",
|
| 137 |
+
"exact_match": 0.3079
|
| 138 |
},
|
| 139 |
{
|
| 140 |
"name": "xlsum_tr",
|
results/zero-shot/aya-23-8b.json
CHANGED
|
@@ -128,7 +128,7 @@
|
|
| 128 |
{
|
| 129 |
"name": "gecturk_generation",
|
| 130 |
"task": "grammatical_error_correction",
|
| 131 |
-
"exact_match": 0.
|
| 132 |
},
|
| 133 |
{
|
| 134 |
"name": "mlsum_tr",
|
|
|
|
| 128 |
{
|
| 129 |
"name": "gecturk_generation",
|
| 130 |
"task": "grammatical_error_correction",
|
| 131 |
+
"exact_match": 0.0429
|
| 132 |
},
|
| 133 |
{
|
| 134 |
"name": "mlsum_tr",
|
results/zero-shot/aya-expanse-32b.json
CHANGED
|
@@ -167,7 +167,7 @@
|
|
| 167 |
{
|
| 168 |
"name": "gecturk_generation",
|
| 169 |
"task": "grammatical_error_correction",
|
| 170 |
-
"exact_match": 0
|
| 171 |
},
|
| 172 |
{
|
| 173 |
"name": "turkce_atasozleri",
|
|
|
|
| 167 |
{
|
| 168 |
"name": "gecturk_generation",
|
| 169 |
"task": "grammatical_error_correction",
|
| 170 |
+
"exact_match": 0.0447
|
| 171 |
},
|
| 172 |
{
|
| 173 |
"name": "turkce_atasozleri",
|
results/zero-shot/aya-expanse-8b.json
CHANGED
|
@@ -30,7 +30,7 @@
|
|
| 30 |
{
|
| 31 |
"name": "gecturk_generation",
|
| 32 |
"task": "grammatical_error_correction",
|
| 33 |
-
"exact_match": 0.
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "ironytr",
|
|
|
|
| 30 |
{
|
| 31 |
"name": "gecturk_generation",
|
| 32 |
"task": "grammatical_error_correction",
|
| 33 |
+
"exact_match": 0.0552
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "ironytr",
|
results/zero-shot/aya101.json
CHANGED
|
@@ -154,7 +154,7 @@
|
|
| 154 |
{
|
| 155 |
"name": "gecturk_generation",
|
| 156 |
"task": "grammatical_error_correction",
|
| 157 |
-
"exact_match": 0.
|
| 158 |
},
|
| 159 |
{
|
| 160 |
"name": "exams_tr",
|
|
|
|
| 154 |
{
|
| 155 |
"name": "gecturk_generation",
|
| 156 |
"task": "grammatical_error_correction",
|
| 157 |
+
"exact_match": 0.3239
|
| 158 |
},
|
| 159 |
{
|
| 160 |
"name": "exams_tr",
|
results/zero-shot/commencis-7b.json
CHANGED
|
@@ -154,7 +154,7 @@
|
|
| 154 |
{
|
| 155 |
"name": "gecturk_generation",
|
| 156 |
"task": "grammatical_error_correction",
|
| 157 |
-
"exact_match": 0.
|
| 158 |
},
|
| 159 |
{
|
| 160 |
"name": "exams_tr",
|
|
|
|
| 154 |
{
|
| 155 |
"name": "gecturk_generation",
|
| 156 |
"task": "grammatical_error_correction",
|
| 157 |
+
"exact_match": 0.2371
|
| 158 |
},
|
| 159 |
{
|
| 160 |
"name": "exams_tr",
|
results/zero-shot/kanarya-2b.json
CHANGED
|
@@ -23,7 +23,7 @@
|
|
| 23 |
{
|
| 24 |
"name": "gecturk_generation",
|
| 25 |
"task": "grammatical_error_correction",
|
| 26 |
-
"exact_match":
|
| 27 |
},
|
| 28 |
{
|
| 29 |
"name": "ironytr",
|
|
|
|
| 23 |
{
|
| 24 |
"name": "gecturk_generation",
|
| 25 |
"task": "grammatical_error_correction",
|
| 26 |
+
"exact_match": 0.0662
|
| 27 |
},
|
| 28 |
{
|
| 29 |
"name": "ironytr",
|
results/zero-shot/llama-3-8b-instruct.json
CHANGED
|
@@ -127,7 +127,7 @@
|
|
| 127 |
{
|
| 128 |
"name": "gecturk_generation",
|
| 129 |
"task": "grammatical_error_correction",
|
| 130 |
-
"exact_match": 0.
|
| 131 |
},
|
| 132 |
{
|
| 133 |
"name": "mlsum_tr",
|
|
|
|
| 127 |
{
|
| 128 |
"name": "gecturk_generation",
|
| 129 |
"task": "grammatical_error_correction",
|
| 130 |
+
"exact_match": 0.1001
|
| 131 |
},
|
| 132 |
{
|
| 133 |
"name": "mlsum_tr",
|
results/zero-shot/llama-3-8b.json
CHANGED
|
@@ -126,7 +126,7 @@
|
|
| 126 |
{
|
| 127 |
"name": "gecturk_generation",
|
| 128 |
"task": "grammatical_error_correction",
|
| 129 |
-
"exact_match": 0.
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "mlsum_tr",
|
|
|
|
| 126 |
{
|
| 127 |
"name": "gecturk_generation",
|
| 128 |
"task": "grammatical_error_correction",
|
| 129 |
+
"exact_match": 0.3412
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "mlsum_tr",
|
results/zero-shot/llama-3.1-8b-instruct.json
CHANGED
|
@@ -126,7 +126,7 @@
|
|
| 126 |
{
|
| 127 |
"name": "gecturk_generation",
|
| 128 |
"task": "grammatical_error_correction",
|
| 129 |
-
"exact_match": 0.
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "mlsum_tr",
|
|
|
|
| 126 |
{
|
| 127 |
"name": "gecturk_generation",
|
| 128 |
"task": "grammatical_error_correction",
|
| 129 |
+
"exact_match": 0.3149
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "mlsum_tr",
|
results/zero-shot/llama-3.1-8b.json
CHANGED
|
@@ -126,7 +126,7 @@
|
|
| 126 |
{
|
| 127 |
"name": "gecturk_generation",
|
| 128 |
"task": "grammatical_error_correction",
|
| 129 |
-
"exact_match": 0.
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "mlsum_tr",
|
|
|
|
| 126 |
{
|
| 127 |
"name": "gecturk_generation",
|
| 128 |
"task": "grammatical_error_correction",
|
| 129 |
+
"exact_match": 0.3527
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "mlsum_tr",
|
results/zero-shot/llama-3.2-1b.json
CHANGED
|
@@ -30,7 +30,7 @@
|
|
| 30 |
{
|
| 31 |
"name": "gecturk_generation",
|
| 32 |
"task": "grammatical_error_correction",
|
| 33 |
-
"exact_match": 0.
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "ironytr",
|
|
|
|
| 30 |
{
|
| 31 |
"name": "gecturk_generation",
|
| 32 |
"task": "grammatical_error_correction",
|
| 33 |
+
"exact_match": 0.181
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "ironytr",
|
results/zero-shot/llama-3.2-3b-instruct.json
CHANGED
|
@@ -30,7 +30,7 @@
|
|
| 30 |
{
|
| 31 |
"name": "gecturk_generation",
|
| 32 |
"task": "grammatical_error_correction",
|
| 33 |
-
"exact_match": 0.
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "ironytr",
|
|
|
|
| 30 |
{
|
| 31 |
"name": "gecturk_generation",
|
| 32 |
"task": "grammatical_error_correction",
|
| 33 |
+
"exact_match": 0.1672
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "ironytr",
|
results/zero-shot/llama-3.2-3b.json
CHANGED
|
@@ -120,7 +120,7 @@
|
|
| 120 |
{
|
| 121 |
"name": "gecturk_generation",
|
| 122 |
"task": "grammatical_error_correction",
|
| 123 |
-
"exact_match": 0.
|
| 124 |
},
|
| 125 |
{
|
| 126 |
"name": "mlsum_tr",
|
|
|
|
| 120 |
{
|
| 121 |
"name": "gecturk_generation",
|
| 122 |
"task": "grammatical_error_correction",
|
| 123 |
+
"exact_match": 0.2682
|
| 124 |
},
|
| 125 |
{
|
| 126 |
"name": "mlsum_tr",
|
results/zero-shot/mistral-7b.json
CHANGED
|
@@ -132,7 +132,7 @@
|
|
| 132 |
{
|
| 133 |
"name": "gecturk_generation",
|
| 134 |
"task": "grammatical_error_correction",
|
| 135 |
-
"exact_match": 0.
|
| 136 |
},
|
| 137 |
{
|
| 138 |
"name": "mlsum_tr",
|
|
|
|
| 132 |
{
|
| 133 |
"name": "gecturk_generation",
|
| 134 |
"task": "grammatical_error_correction",
|
| 135 |
+
"exact_match": 0.2084
|
| 136 |
},
|
| 137 |
{
|
| 138 |
"name": "mlsum_tr",
|
results/zero-shot/trendyol-7b.json
CHANGED
|
@@ -154,7 +154,7 @@
|
|
| 154 |
{
|
| 155 |
"name": "gecturk_generation",
|
| 156 |
"task": "grammatical_error_correction",
|
| 157 |
-
"exact_match": 0.
|
| 158 |
},
|
| 159 |
{
|
| 160 |
"name": "exams_tr",
|
|
|
|
| 154 |
{
|
| 155 |
"name": "gecturk_generation",
|
| 156 |
"task": "grammatical_error_correction",
|
| 157 |
+
"exact_match": 0.0007
|
| 158 |
},
|
| 159 |
{
|
| 160 |
"name": "exams_tr",
|