Update README.md
Browse files
README.md
CHANGED
@@ -30,17 +30,17 @@ For training, we used all Turkish data that was present in the monolingual Turki
|
|
30 |
|
31 |
# Benchmark performance
|
32 |
|
33 |
-
We tested the performance of **XLMR-MaCoCu-tr** on benchmarks of XPOS, UPOS and NER from the [Universal Dependencies](https://universaldependencies.org/) project.
|
34 |
|
35 |
Scores are averages of three runs, except for COPA, for which we use 10 runs. We use the same hyperparameter settings for all models for POS/NER, for COPA we optimized each learning rate on the dev set.
|
36 |
|
37 |
-
| | **UPOS** | **UPOS** | **XPOS** | **XPOS** | **NER** | **NER** | **COPA** |
|
38 |
-
|--------------------|:--------:|:--------:|:--------:|:--------:|---------|----------| ----------|
|
39 |
-
| | **Dev** | **Test** | **Dev** | **Test** | **Dev** | **Test** | **Test** |
|
40 |
-
| **XLM-R-base** | 89.0 | 89.0 | 90.4 | 90.6 | 92.8 | 92.6 | 56.0 |
|
41 |
-
| **XLM-R-large** | 89.4 | 89.3 | 90.8 | 90.7 | 94.1 | 94.1 | 52.1 |
|
42 |
-
| **BERTurk** | 88.2 | 88.4 | 89.7 | 89.6 | 92.6 | 92.6 | 57.0 |
|
43 |
-
| **XLMR-MaCoCu-tr** | 89.1 | 89.4 | 90.7 | 90.5 | 94.4 | 94.4 | 60.7 |
|
44 |
|
45 |
# Acknowledgements
|
46 |
|
|
|
30 |
|
31 |
# Benchmark performance
|
32 |
|
33 |
+
We tested the performance of **XLMR-MaCoCu-tr** on benchmarks of XPOS, UPOS and NER from the [Universal Dependencies](https://universaldependencies.org/) project. For COPA, we train on a machine translated (MT) set of the data (for details see our [Github repo](https://github.com/RikVN/COPA)), and evaluate on a similar MT set, but also on the human-translated (HT) test set from the [XCOPA](https://github.com/cambridgeltl/xcopa) project. We compare performance to the strong multi-lingual models XLMR-base and XLMR-large, but also to the monolingual [BERTurk](https://huggingface.co/dbmdz/bert-base-turkish-cased) model. For details regarding the fine-tuning procedure you can checkout our [Github](https://github.com/macocu/LanguageModels).
|
34 |
|
35 |
Scores are averages of three runs, except for COPA, for which we use 10 runs. We use the same hyperparameter settings for all models for POS/NER, for COPA we optimized each learning rate on the dev set.
|
36 |
|
37 |
+
| | **UPOS** | **UPOS** | **XPOS** | **XPOS** | **NER** | **NER** | **COPA** | **COPA** |
|
38 |
+
|--------------------|:--------:|:--------:|:--------:|:--------:|---------|----------| ----------| ----------|
|
39 |
+
| | **Dev** | **Test** | **Dev** | **Test** | **Dev** | **Test** | **Test (MT)** | **Test (HT)** |
|
40 |
+
| **XLM-R-base** | 89.0 | 89.0 | 90.4 | 90.6 | 92.8 | 92.6 | 56.0 | 56.4 |
|
41 |
+
| **XLM-R-large** | 89.4 | 89.3 | 90.8 | 90.7 | 94.1 | 94.1 | 52.1 | 53.2 |
|
42 |
+
| **BERTurk** | 88.2 | 88.4 | 89.7 | 89.6 | 92.6 | 92.6 | 57.0 | |
|
43 |
+
| **XLMR-MaCoCu-tr** | 89.1 | 89.4 | 90.7 | 90.5 | 94.4 | 94.4 | 60.7 | |
|
44 |
|
45 |
# Acknowledgements
|
46 |
|