Update README.md
Browse files
README.md
CHANGED
@@ -34,17 +34,12 @@ It achieves the following results on the evaluation set:
|
|
34 |
|
35 |
## Model description
|
36 |
|
37 |
-
|
38 |
|
39 |
-
|
40 |
|
41 |
-
|
42 |
|
43 |
-
## Training and evaluation data
|
44 |
-
|
45 |
-
More information needed
|
46 |
-
|
47 |
-
## Training procedure
|
48 |
|
49 |
### Training hyperparameters
|
50 |
|
@@ -73,3 +68,23 @@ The following hyperparameters were used during training:
|
|
73 |
- Pytorch 1.9.0+cu111
|
74 |
- Datasets 1.10.2
|
75 |
- Tokenizers 0.10.3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
## Model description
|
36 |
|
37 |
+
This model was fine-tunned on token classification task (NER) on Portuguese archival documents. The annotated labels are: Date, Profession, Person, Place, Organization
|
38 |
|
39 |
+
### Datasets
|
40 |
|
41 |
+
All the training and evaluation data is available at: http://ner.epl.di.uminho.pt/
|
42 |
|
|
|
|
|
|
|
|
|
|
|
43 |
|
44 |
### Training hyperparameters
|
45 |
|
|
|
68 |
- Pytorch 1.9.0+cu111
|
69 |
- Datasets 1.10.2
|
70 |
- Tokenizers 0.10.3
|
71 |
+
### Citation
|
72 |
+
|
73 |
+
@InProceedings{10.1007/978-3-031-04819-7_33,
|
74 |
+
author="da Costa Cunha, Lu{\'i}s Filipe
|
75 |
+
and Ramalho, Jos{\'e} Carlos",
|
76 |
+
editor="Rocha, Alvaro
|
77 |
+
and Adeli, Hojjat
|
78 |
+
and Dzemyda, Gintautas
|
79 |
+
and Moreira, Fernando",
|
80 |
+
title="NER in Archival Finding Aids: Next Level",
|
81 |
+
booktitle="Information Systems and Technologies",
|
82 |
+
year="2022",
|
83 |
+
publisher="Springer International Publishing",
|
84 |
+
address="Cham",
|
85 |
+
pages="333--342",
|
86 |
+
abstract="Currently, there is a vast amount of archival finding aids in Portuguese archives, however, these documents lack structure (are not annotated) making them hard to process and work with. In this way, we intend to extract and classify entities of interest, like geographicallocations, people's names, dates, etc. For this, we will use an architecture that has been revolutionizing several NLP tasks, Transformers, presenting several models in order to achieve high results. It is also intended to understand what will be the degree of improvement that this new mechanism will present in comparison with previous architectures. Can Transformer-based models replace the LSTMs in NER? We intend to answer this question along this paper.",
|
87 |
+
isbn="978-3-031-04819-7"
|
88 |
+
}
|
89 |
+
|
90 |
+
|