Update README.md
Browse files
README.md
CHANGED
@@ -23,4 +23,18 @@ This is a fine-tuned version of the [BERTić](https://huggingface.co/CLASSLA/bcm
|
|
23 |
- the [ReLDI-hr](http://hdl.handle.net/11356/1241) dataset, 89 thousand tokens in size, Internet (Twitter) Croatian
|
24 |
- the [ReLDI-sr](http://hdl.handle.net/11356/1240) dataset, 92 thousand tokens in size, Internet (Twitter) Serbian
|
25 |
|
26 |
-
The data was augmented with missing diacritics and standard data was additionally over-represented. The F1 obtained on dev data (train and test was merged into train) is 91.38. For a more detailed per-dataset evaluation of the BERTić model on the NER task have a look at the [main model page](https://huggingface.co/CLASSLA/bcms-bertic).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
- the [ReLDI-hr](http://hdl.handle.net/11356/1241) dataset, 89 thousand tokens in size, Internet (Twitter) Croatian
|
24 |
- the [ReLDI-sr](http://hdl.handle.net/11356/1240) dataset, 92 thousand tokens in size, Internet (Twitter) Serbian
|
25 |
|
26 |
+
The data was augmented with missing diacritics and standard data was additionally over-represented. The F1 obtained on dev data (train and test was merged into train) is 91.38. For a more detailed per-dataset evaluation of the BERTić model on the NER task have a look at the [main model page](https://huggingface.co/CLASSLA/bcms-bertic).
|
27 |
+
|
28 |
+
If you use this fine-tuned model, please cite the following paper:
|
29 |
+
|
30 |
+
```
|
31 |
+
@inproceedings{ljubesic-lauc-2021-bertic,
|
32 |
+
title = "{BERTić} - The Transformer Language Model for {B}osnian, {C}roatian, {M}ontenegrin and {S}erbian",
|
33 |
+
author = "Ljube{\v{s}}i{\'c}, Nikola and
|
34 |
+
Lauc, Davor",
|
35 |
+
booktitle = "Proceedings of the 8th Workshop on Balto-Slavic Natural Language Processing",
|
36 |
+
year = "2021",
|
37 |
+
address = "Kiev, Ukraine",
|
38 |
+
publisher = "Association for Computational Linguistics"
|
39 |
+
}
|
40 |
+
```
|