Update README.md
Browse files
README.md
CHANGED
@@ -92,7 +92,7 @@ pprint(ner_results)
|
|
92 |
## Training
|
93 |
|
94 |
### Training data
|
95 |
-
We used the NER dataset in Catalan called [
|
96 |
|
97 |
### Training Procedure
|
98 |
The model was trained with a batch size of 16 and a learning rate of 5e-5 for 5 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.
|
@@ -104,9 +104,9 @@ The model was trained with a batch size of 16 and a learning rate of 5e-5 for 5
|
|
104 |
This model was finetuned maximizing F1 score.
|
105 |
|
106 |
### Evaluation results
|
107 |
-
We evaluated the _roberta-base-ca-v2-cased-ner_ on the
|
108 |
|
109 |
-
| Model |
|
110 |
| ------------|:-------------|
|
111 |
| roberta-base-ca-v2-cased-ner | 89.29 |
|
112 |
| roberta-base-ca-cased-ner | **89.76** |
|
|
|
92 |
## Training
|
93 |
|
94 |
### Training data
|
95 |
+
We used the NER dataset in Catalan called [AnCora-Ca-NER](https://huggingface.co/datasets/projecte-aina/ancora-ca-ner) for training and evaluation.
|
96 |
|
97 |
### Training Procedure
|
98 |
The model was trained with a batch size of 16 and a learning rate of 5e-5 for 5 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.
|
|
|
104 |
This model was finetuned maximizing F1 score.
|
105 |
|
106 |
### Evaluation results
|
107 |
+
We evaluated the _roberta-base-ca-v2-cased-ner_ on the AnCora-Ca-NER test set against standard multilingual and monolingual baselines:
|
108 |
|
109 |
+
| Model | AnCora-Ca-NER (F1)|
|
110 |
| ------------|:-------------|
|
111 |
| roberta-base-ca-v2-cased-ner | 89.29 |
|
112 |
| roberta-base-ca-cased-ner | **89.76** |
|