Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
task_categories:
|
3 |
+
- text-classification
|
4 |
+
- token-classification
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
---
|
8 |
+
|
9 |
+
# tl;dr:
|
10 |
+
|
11 |
+
This is a dataset largely based on CleanCoNLL but with some augmentations.
|
12 |
+
|
13 |
+
# Details:
|
14 |
+
|
15 |
+
## Base:
|
16 |
+
We started with the CoNLL-2003 dataset, a standard NER benchmark containing English and German text annotated with four entity types: person, location, organization, and miscellaneous. For our evaluation, we focused solely on examples containing the ORG (organization) entity, as these are most relevant to competitor detection.
|
17 |
+
We then applied corrections from CleanCoNLL, a 2023 revision by Rücker and Akbik that addresses annotation errors in the original CoNLL-2003. CleanCoNLL corrects 7.0% of labels in the English dataset, adds entity linking annotations, and maintains the original four entity types. This improved dataset enables more accurate evaluation, with top NER models achieving F1-scores up to 97.1%.
|
18 |
+
|
19 |
+
## Augmentations:
|
20 |
+
|
21 |
+
We created two augmented datasets to test specific aspects of competitor detection:
|
22 |
+
|
23 |
+
Positive Dataset (with typographical errors):
|
24 |
+
We selected random examples and used the ORG entity as the "competitor" to be detected. We introduced typographical errors to the competitor names by:
|
25 |
+
- Omission: Removing a letter
|
26 |
+
- Transposition: Swapping two adjacent letters
|
27 |
+
- Substitution: Swapping a letter with one found nearby on a US ANSI keyboard layout
|
28 |
+
- Duplication: Selecting a character at random and doubling it
|
29 |
+
|
30 |
+
This dataset tests the guardrail's ability to detect variations of competitor names, which is particularly relevant as our solution does not implement fuzzy matching.
|
31 |
+
|
32 |
+
Negative Dataset (with distractors)
|
33 |
+
For the negative dataset, we used the original examples containing ORG entities but created a list of "competitors" by randomly selecting companies from the Fortune 500 index (2024), excluding the actual ORG entity in the text. We set the 'has_competitor' flag to 'false' for all examples in this dataset. This evaluates the guardrail's precision in avoiding false positives when no actual competitors are mentioned.
|
34 |
+
|
35 |
+
# Citations
|
36 |
+
|
37 |
+
```
|
38 |
+
@inproceedings{rucker-akbik-2023-cleanconll,
|
39 |
+
title = "{C}lean{C}o{NLL}: A Nearly Noise-Free Named Entity Recognition Dataset",
|
40 |
+
author = {R{\"u}cker, Susanna and Akbik, Alan},
|
41 |
+
editor = "Bouamor, Houda and Pino, Juan and Bali, Kalika",
|
42 |
+
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
|
43 |
+
month = dec,
|
44 |
+
year = "2023",
|
45 |
+
address = "Singapore",
|
46 |
+
publisher = "Association for Computational Linguistics",
|
47 |
+
url = "https://aclanthology.org/2023.emnlp-main.533",
|
48 |
+
doi = "10.18653/v1/2023.emnlp-main.533",
|
49 |
+
pages = "8628--8645",
|
50 |
+
}
|
51 |
+
|
52 |
+
@misc{rücker2023cleanconll,
|
53 |
+
title={{C}lean{C}o{NLL}: A Nearly Noise-Free Named Entity Recognition Dataset},
|
54 |
+
author={Susanna R{\"u}cker and Alan Akbik},
|
55 |
+
year={2023},
|
56 |
+
eprint={2310.16225},
|
57 |
+
archivePrefix={arXiv},
|
58 |
+
primaryClass={cs.CL}
|
59 |
+
}
|
60 |
+
```
|