Update README.md
Browse files
README.md
CHANGED
@@ -27,3 +27,32 @@ configs:
|
|
27 |
- split: test
|
28 |
path: data/test-*
|
29 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
- split: test
|
28 |
path: data/test-*
|
29 |
---
|
30 |
+
## Dataset Description
|
31 |
+
|
32 |
+
This dataset is designed for fine-tuning language models, particularly the [Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) model, for the task of hate speech detection in social media text (tweets). It focuses on both **implicit** and **explicit** forms of hate speech, aiming to improve the performance of smaller language models in this challenging task.
|
33 |
+
|
34 |
+
The dataset is a combination of two existing datasets:
|
35 |
+
|
36 |
+
* **Hate Speech Examples:** Examples of implicit hate speech are sourced from the [SALT-NLP/ImplicitHate](https://huggingface.co/datasets/SALT-NLP/ImplicitHate) dataset. This dataset contains tweets annotated as containing implicit hate speech, categorized into types like grievance, incitement, inferiority, irony, stereotyping, and threatening.
|
37 |
+
* **Non-Hate Speech Examples:** Examples of non-hate speech are sourced from the [TweetEval](https://huggingface.co/datasets/tweet_eval) dataset, specifically the `hate` configuration. This configuration provides tweets labeled as 'non-hate'.
|
38 |
+
|
39 |
+
By combining these two sources, we create a dataset suitable for binary classification of tweets into "hate speech" and "not hate speech" categories.
|
40 |
+
|
41 |
+
## Dataset Splits
|
42 |
+
|
43 |
+
The dataset is divided into the following splits:
|
44 |
+
|
45 |
+
* **`train`**: Contains 2500 examples for training the model.
|
46 |
+
* **`validation`**: Contains 150 examples for evaluating and tuning the model during training.
|
47 |
+
* **`test`**: Contains 100 examples for final evaluation of the trained model's performance.
|
48 |
+
|
49 |
+
These splits are designed to be relatively balanced in terms of class distribution (hate vs. not hate) to ensure fair evaluation.
|
50 |
+
|
51 |
+
## Dataset Fields
|
52 |
+
|
53 |
+
Each example in the dataset consists of the following fields:
|
54 |
+
|
55 |
+
* **`text`**: (`string`) The text content of the tweet.
|
56 |
+
* **`label`**: (`int`) The label for the tweet, with the following mapping:
|
57 |
+
* `0`: Not Hate Speech
|
58 |
+
* `1`: Hate Speech
|