Commit
·
25d241b
1
Parent(s):
23e7044
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- badmatr11x/hate-offensive-speech
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
metrics:
|
8 |
+
- accuracy
|
9 |
+
library_name: adapter-transformers
|
10 |
+
pipeline_tag: text-classification
|
11 |
+
tags:
|
12 |
+
- code
|
13 |
+
widget:
|
14 |
+
- text: "People are fun to talk."
|
15 |
+
example_title: "Neither Speech"
|
16 |
+
- text: "Black people are good at running."
|
17 |
+
example_title: "Hate Speech"
|
18 |
+
- text: "And I'm goin back to school, only for the hoes and a class or two."
|
19 |
+
example_title: "Offensive Speech"
|
20 |
+
---
|
21 |
+
|
22 |
+
|
23 |
+
This is the **Offensive and Hateful Speech Detection** mode fine-tuned on the **distilroberta-base** model available on the huggingface pre-trained models. This model is trained with the [dataset](https://huggingface.co/datasets/badmatr11x/hate-offensive-speech/) which contains around 55K annotated tweets; classified into three different categories, Hateful, Offensive and Neither.
|
24 |
+
|
25 |
+
This is the example of the dataset instance:
|
26 |
+
```
|
27 |
+
{
|
28 |
+
"label": {
|
29 |
+
0: "Hate Speech",
|
30 |
+
1: "Offensive Speech",
|
31 |
+
2: "Neither"
|
32 |
+
}
|
33 |
+
"tweet": <string>
|
34 |
+
}
|
35 |
+
```
|
36 |
+
|
37 |
+
Model is fine-tuned on epochs number 5 with over than 15500 rounds of training. The self-verified evaluation accuracy of the models is **95.60%** with the evaluation lost **17.02%**. The testing accuracy of the model is recored **95.04%**, self stated.
|