init
Browse files- get_stats.py +3 -2
- super_tweet_eval.py +1 -1
get_stats.py
CHANGED
|
@@ -7,9 +7,10 @@ task_description = {
|
|
| 7 |
'tweet_ner7': "sequence labeling",
|
| 8 |
'tweet_qa': "generation",
|
| 9 |
'tweet_similarity': "regression on two texts",
|
| 10 |
-
'tweet_topic': "multi-label classification"
|
|
|
|
| 11 |
}
|
| 12 |
-
for task in
|
| 13 |
data = load_dataset("cardiffnlp/super_tweet_eval", task)
|
| 14 |
tmp_table = {"task": task, "description": task_description[task]}
|
| 15 |
tmp_table['number of instances'] = " / ".join([str(len(data[s])) for s in ['train', 'validation', 'test']])
|
|
|
|
| 7 |
'tweet_ner7': "sequence labeling",
|
| 8 |
'tweet_qa': "generation",
|
| 9 |
'tweet_similarity': "regression on two texts",
|
| 10 |
+
'tweet_topic': "multi-label classification",
|
| 11 |
+
"tempo_wic": "binary classification on two texts"
|
| 12 |
}
|
| 13 |
+
for task in task_description.keys():
|
| 14 |
data = load_dataset("cardiffnlp/super_tweet_eval", task)
|
| 15 |
tmp_table = {"task": task, "description": task_description[task]}
|
| 16 |
tmp_table['number of instances'] = " / ".join([str(len(data[s])) for s in ['train', 'validation', 'test']])
|
super_tweet_eval.py
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
import json
|
| 3 |
import datasets
|
| 4 |
|
| 5 |
-
_VERSION = "0.0.
|
| 6 |
_SUPER_TWEET_EVAL_CITATION = """TBA"""
|
| 7 |
_SUPER_TWEET_EVAL_DESCRIPTION = """TBA"""
|
| 8 |
_TWEET_TOPIC_DESCRIPTION = """
|
|
|
|
| 2 |
import json
|
| 3 |
import datasets
|
| 4 |
|
| 5 |
+
_VERSION = "0.0.5"
|
| 6 |
_SUPER_TWEET_EVAL_CITATION = """TBA"""
|
| 7 |
_SUPER_TWEET_EVAL_DESCRIPTION = """TBA"""
|
| 8 |
_TWEET_TOPIC_DESCRIPTION = """
|