Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -144,8 +144,6 @@ There are several dataset versions available:
|
|
144 |
* [AdapTable-cluster29](https://huggingface.co/datasets/MicPie/adaptable_cluster29)
|
145 |
* [AdapTable-cluster-noise](https://huggingface.co/datasets/MicPie/adaptable_cluster-noise)
|
146 |
|
147 |
-
|
148 |
-
|
149 |
### Supported Tasks and Leaderboards
|
150 |
|
151 |
Since the tables come from the web, the distribution of tasks and topics is very broad. The shape of our dataset is very wide, i.e., we have 1000's of tasks, while each task has only a few examples, compared to most current NLP datasets which are very deep, i.e., 10s of tasks with many examples. This implies that our dataset covers a broad range of potential tasks, e.g., multiple-choice, question-answering, table-question-answering, text-classification, etc.
|
@@ -220,14 +218,11 @@ The data was extracted from [WDC Web Table Corpora](http://webdatacommons.org/we
|
|
220 |
|
221 |
### Social Impact of Dataset
|
222 |
|
223 |
-
|
224 |
-
|
225 |
-
While tables have a similar structure to few-shot tasks and we do see an improved performance on few-shot tasks in our paper, we want to make clear that fine-tuning on tables also has its risks. First of all, since the tables are extracted from the web, they may contain user identities or otherwise sensitive information which a model might reveal at inference, or which could influence the learning process of a model in a negative way. Second, since tables are very diverse in nature, the model also trains on low-quality data or data with an unusual structure. While it is interesting that training on such data improves few-shot performance on downstream tasks, this could also imply that the model learns concepts that are very dissimilar to human concepts that would be useful for a certain downstream task. In other words, it is possible that the model learns weird things that are helpful on the evaluated downstream tasks, but might lead to bad out-of-distribution behavior.
|
226 |
|
227 |
### Discussion of Biases
|
228 |
|
229 |
-
Since our dataset contains tables that are scraped from the web, it will also contain many toxic, racist, sexist, and otherwise harmful biases and texts. We have not run any analysis on the biases prevalent in our datasets. Neither have we explicitly filtered the content.
|
230 |
-
This implies that a model trained on our dataset will potentially reinforce harmful biases and toxic text that exist in our dataset.
|
231 |
|
232 |
### Other Known Limitations
|
233 |
|
|
|
144 |
* [AdapTable-cluster29](https://huggingface.co/datasets/MicPie/adaptable_cluster29)
|
145 |
* [AdapTable-cluster-noise](https://huggingface.co/datasets/MicPie/adaptable_cluster-noise)
|
146 |
|
|
|
|
|
147 |
### Supported Tasks and Leaderboards
|
148 |
|
149 |
Since the tables come from the web, the distribution of tasks and topics is very broad. The shape of our dataset is very wide, i.e., we have 1000's of tasks, while each task has only a few examples, compared to most current NLP datasets which are very deep, i.e., 10s of tasks with many examples. This implies that our dataset covers a broad range of potential tasks, e.g., multiple-choice, question-answering, table-question-answering, text-classification, etc.
|
|
|
218 |
|
219 |
### Social Impact of Dataset
|
220 |
|
221 |
+
This dataset is intended for use as a research resource to investigate the relationship between training data and few-shot learning. As such, it contains high- and low-quality data, as well as diverse content that may be untruthful or inappropriate. Without careful investigation, it should not be used for training models that will be deployed for use in decision-critical or user-facing situations.
|
|
|
|
|
222 |
|
223 |
### Discussion of Biases
|
224 |
|
225 |
+
Since our dataset contains tables that are scraped from the web, it will also contain many toxic, racist, sexist, and otherwise harmful biases and texts. We have not run any analysis on the biases prevalent in our datasets. Neither have we explicitly filtered the content. This implies that a model trained on our dataset may potentially reflect harmful biases and toxic text that exist in our dataset.
|
|
|
226 |
|
227 |
### Other Known Limitations
|
228 |
|