lfcc commited on
Commit
ec93c35
1 Parent(s): db87061

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -8
README.md CHANGED
@@ -34,17 +34,12 @@ It achieves the following results on the evaluation set:
34
 
35
  ## Model description
36
 
37
- More information needed
38
 
39
- ## Intended uses & limitations
40
 
41
- More information needed
42
 
43
- ## Training and evaluation data
44
-
45
- More information needed
46
-
47
- ## Training procedure
48
 
49
  ### Training hyperparameters
50
 
@@ -73,3 +68,23 @@ The following hyperparameters were used during training:
73
  - Pytorch 1.9.0+cu111
74
  - Datasets 1.10.2
75
  - Tokenizers 0.10.3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  ## Model description
36
 
37
+ This model was fine-tunned on token classification task (NER) on Portuguese archival documents. The annotated labels are: Date, Profession, Person, Place, Organization
38
 
39
+ ### Datasets
40
 
41
+ All the training and evaluation data is available at: http://ner.epl.di.uminho.pt/
42
 
 
 
 
 
 
43
 
44
  ### Training hyperparameters
45
 
 
68
  - Pytorch 1.9.0+cu111
69
  - Datasets 1.10.2
70
  - Tokenizers 0.10.3
71
+ ### Citation
72
+
73
+ @InProceedings{10.1007/978-3-031-04819-7_33,
74
+ author="da Costa Cunha, Lu{\'i}s Filipe
75
+ and Ramalho, Jos{\'e} Carlos",
76
+ editor="Rocha, Alvaro
77
+ and Adeli, Hojjat
78
+ and Dzemyda, Gintautas
79
+ and Moreira, Fernando",
80
+ title="NER in Archival Finding Aids: Next Level",
81
+ booktitle="Information Systems and Technologies",
82
+ year="2022",
83
+ publisher="Springer International Publishing",
84
+ address="Cham",
85
+ pages="333--342",
86
+ abstract="Currently, there is a vast amount of archival finding aids in Portuguese archives, however, these documents lack structure (are not annotated) making them hard to process and work with. In this way, we intend to extract and classify entities of interest, like geographicallocations, people's names, dates, etc. For this, we will use an architecture that has been revolutionizing several NLP tasks, Transformers, presenting several models in order to achieve high results. It is also intended to understand what will be the degree of improvement that this new mechanism will present in comparison with previous architectures. Can Transformer-based models replace the LSTMs in NER? We intend to answer this question along this paper.",
87
+ isbn="978-3-031-04819-7"
88
+ }
89
+
90
+