de-francophones
commited on
Commit
•
398bba1
1
Parent(s):
9ea2122
77d50b733a6e7b3a3c6cc62c3261e6d8994fbfea4918b7b2b59565c148e9a6e9
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +1 -0
- sysoutputs/orange_deskin/test_02/pertreebank/ta_ttb-ud-test.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test-sys.conllu +0 -0
- sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/pl.conllu +0 -0
- sysoutputs/orange_deskin/test_02/pl.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/ru.conllu +3 -0
- sysoutputs/orange_deskin/test_02/ru.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/sk.conllu +0 -0
- sysoutputs/orange_deskin/test_02/sk.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/sv.conllu +0 -0
- sysoutputs/orange_deskin/test_02/sv.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/ta.conllu +0 -0
- sysoutputs/orange_deskin/test_02/ta.eval.log +17 -0
- sysoutputs/orange_deskin/test_02/uk.conllu +0 -0
- sysoutputs/orange_deskin/test_02/uk.eval.log +17 -0
- sysoutputs/robertnlp/mainsubmission/en.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/en.eval.log +17 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/ar_padt-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/bg_btb-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/cs_cac-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/cs_fictree-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pdt-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pud-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test.eval.log +17 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test.eval.log +17 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test.eval.log +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test-sys.conllu +0 -0
- sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test.eval.log +0 -0
.gitattributes
CHANGED
@@ -95,3 +95,4 @@ sysoutputs/koebsala/udpipe_test_20200425_134739/ru.conllu filter=lfs diff=lfs me
|
|
95 |
sysoutputs/orange_deskin/test_02/cs.conllu filter=lfs diff=lfs merge=lfs -text
|
96 |
sysoutputs/orange_deskin/test_02/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
|
97 |
sysoutputs/orange_deskin/test_02/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
|
|
|
|
95 |
sysoutputs/orange_deskin/test_02/cs.conllu filter=lfs diff=lfs merge=lfs -text
|
96 |
sysoutputs/orange_deskin/test_02/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
|
97 |
sysoutputs/orange_deskin/test_02/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
|
98 |
+
sysoutputs/orange_deskin/test_02/ru.conllu filter=lfs diff=lfs merge=lfs -text
|
sysoutputs/orange_deskin/test_02/pertreebank/ta_ttb-ud-test.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.38 | 99.60 | 99.49 |
|
4 |
+
Sentences | 92.68 | 95.00 | 93.83 |
|
5 |
+
Words | 95.71 | 94.22 | 94.96 |
|
6 |
+
UPOS | 88.10 | 86.73 | 87.41 | 92.05
|
7 |
+
XPOS | 81.00 | 79.74 | 80.36 | 84.63
|
8 |
+
UFeats | 84.22 | 82.91 | 83.56 | 87.99
|
9 |
+
AllTags | 79.42 | 78.18 | 78.79 | 82.98
|
10 |
+
Lemmas | 90.81 | 89.39 | 90.09 | 94.88
|
11 |
+
UAS | 72.42 | 71.29 | 71.85 | 75.67
|
12 |
+
LAS | 66.75 | 65.71 | 66.23 | 69.74
|
13 |
+
ELAS | 65.46 | 63.14 | 64.28 | 68.46
|
14 |
+
EULAS | 66.94 | 64.57 | 65.73 | 70.01
|
15 |
+
CLAS | 64.02 | 63.41 | 63.71 | 67.58
|
16 |
+
MLAS | 54.63 | 54.11 | 54.37 | 57.67
|
17 |
+
BLEX | 60.04 | 59.47 | 59.75 | 63.39
|
sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test-sys.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.79 | 99.79 | 99.79 |
|
4 |
+
Sentences | 94.89 | 95.74 | 95.31 |
|
5 |
+
Words | 99.75 | 99.76 | 99.75 |
|
6 |
+
UPOS | 98.02 | 98.03 | 98.03 | 98.27
|
7 |
+
XPOS | 94.68 | 94.70 | 94.69 | 94.92
|
8 |
+
UFeats | 94.73 | 94.74 | 94.74 | 94.97
|
9 |
+
AllTags | 93.70 | 93.71 | 93.70 | 93.93
|
10 |
+
Lemmas | 97.46 | 97.48 | 97.47 | 97.71
|
11 |
+
UAS | 90.71 | 90.72 | 90.72 | 90.94
|
12 |
+
LAS | 88.59 | 88.60 | 88.59 | 88.81
|
13 |
+
ELAS | 87.34 | 82.09 | 84.64 | 89.10
|
14 |
+
EULAS | 88.20 | 82.90 | 85.46 | 89.97
|
15 |
+
CLAS | 86.14 | 85.82 | 85.98 | 86.00
|
16 |
+
MLAS | 79.41 | 79.12 | 79.26 | 79.28
|
17 |
+
BLEX | 83.53 | 83.22 | 83.38 | 83.40
|
sysoutputs/orange_deskin/test_02/pl.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/orange_deskin/test_02/pl.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.30 | 99.36 | 99.33 |
|
4 |
+
Sentences | 96.80 | 96.66 | 96.73 |
|
5 |
+
Words | 99.74 | 99.83 | 99.79 |
|
6 |
+
UPOS | 98.53 | 98.63 | 98.58 | 98.79
|
7 |
+
XPOS | 94.31 | 94.40 | 94.36 | 94.56
|
8 |
+
UFeats | 92.63 | 92.73 | 92.68 | 92.88
|
9 |
+
AllTags | 90.11 | 90.20 | 90.16 | 90.35
|
10 |
+
Lemmas | 97.96 | 98.06 | 98.01 | 98.22
|
11 |
+
UAS | 94.21 | 94.30 | 94.26 | 94.46
|
12 |
+
LAS | 92.06 | 92.15 | 92.11 | 92.31
|
13 |
+
ELAS | 82.07 | 78.78 | 80.39 | 84.22
|
14 |
+
EULAS | 91.02 | 87.37 | 89.15 | 93.40
|
15 |
+
CLAS | 90.57 | 90.58 | 90.58 | 90.76
|
16 |
+
MLAS | 80.97 | 80.98 | 80.97 | 81.14
|
17 |
+
BLEX | 88.48 | 88.49 | 88.49 | 88.67
|
sysoutputs/orange_deskin/test_02/ru.conllu
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02dc2bbabd187989b6cfe22c675fb105c185c5d4e692035f38d4690d28ab888f
|
3 |
+
size 11100718
|
sysoutputs/orange_deskin/test_02/ru.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.37 | 99.58 | 99.48 |
|
4 |
+
Sentences | 97.85 | 98.01 | 97.93 |
|
5 |
+
Words | 99.37 | 99.58 | 99.48 |
|
6 |
+
UPOS | 98.69 | 98.90 | 98.80 | 99.31
|
7 |
+
XPOS | 99.37 | 99.58 | 99.48 | 100.00
|
8 |
+
UFeats | 97.65 | 97.86 | 97.76 | 98.27
|
9 |
+
AllTags | 97.47 | 97.68 | 97.58 | 98.09
|
10 |
+
Lemmas | 98.15 | 98.36 | 98.26 | 98.77
|
11 |
+
UAS | 94.31 | 94.51 | 94.41 | 94.90
|
12 |
+
LAS | 93.21 | 93.41 | 93.31 | 93.80
|
13 |
+
ELAS | 91.16 | 88.56 | 89.84 | 94.11
|
14 |
+
EULAS | 92.30 | 89.68 | 90.97 | 95.29
|
15 |
+
CLAS | 92.16 | 92.31 | 92.24 | 92.81
|
16 |
+
MLAS | 89.85 | 90.00 | 89.93 | 90.49
|
17 |
+
BLEX | 90.67 | 90.82 | 90.75 | 91.31
|
sysoutputs/orange_deskin/test_02/sk.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/orange_deskin/test_02/sk.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.99 | 99.98 | 99.99 |
|
4 |
+
Sentences | 84.62 | 83.51 | 84.06 |
|
5 |
+
Words | 99.99 | 99.98 | 99.99 |
|
6 |
+
UPOS | 97.17 | 97.16 | 97.16 | 97.17
|
7 |
+
XPOS | 88.14 | 88.13 | 88.14 | 88.15
|
8 |
+
UFeats | 92.89 | 92.88 | 92.89 | 92.90
|
9 |
+
AllTags | 87.02 | 87.01 | 87.02 | 87.03
|
10 |
+
Lemmas | 96.63 | 96.62 | 96.63 | 96.64
|
11 |
+
UAS | 91.20 | 91.20 | 91.20 | 91.21
|
12 |
+
LAS | 89.06 | 89.05 | 89.06 | 89.07
|
13 |
+
ELAS | 85.89 | 82.87 | 84.36 | 88.35
|
14 |
+
EULAS | 87.75 | 84.66 | 86.17 | 90.26
|
15 |
+
CLAS | 87.35 | 87.38 | 87.37 | 87.39
|
16 |
+
MLAS | 78.01 | 78.04 | 78.03 | 78.05
|
17 |
+
BLEX | 83.72 | 83.75 | 83.73 | 83.76
|
sysoutputs/orange_deskin/test_02/sv.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/orange_deskin/test_02/sv.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.69 | 99.76 | 99.72 |
|
4 |
+
Sentences | 85.20 | 91.03 | 88.02 |
|
5 |
+
Words | 99.69 | 99.76 | 99.72 |
|
6 |
+
UPOS | 97.48 | 97.54 | 97.51 | 97.78
|
7 |
+
XPOS | 95.61 | 95.68 | 95.64 | 95.91
|
8 |
+
UFeats | 88.98 | 89.04 | 89.01 | 89.26
|
9 |
+
AllTags | 87.66 | 87.72 | 87.69 | 87.94
|
10 |
+
Lemmas | 93.36 | 93.42 | 93.39 | 93.65
|
11 |
+
UAS | 88.62 | 88.68 | 88.65 | 88.89
|
12 |
+
LAS | 85.88 | 85.94 | 85.91 | 86.15
|
13 |
+
ELAS | 83.98 | 82.56 | 83.27 | 87.66
|
14 |
+
EULAS | 85.09 | 83.65 | 84.36 | 88.82
|
15 |
+
CLAS | 84.30 | 84.26 | 84.28 | 84.51
|
16 |
+
MLAS | 69.68 | 69.65 | 69.67 | 69.86
|
17 |
+
BLEX | 77.73 | 77.70 | 77.72 | 77.93
|
sysoutputs/orange_deskin/test_02/ta.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/orange_deskin/test_02/ta.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.38 | 99.60 | 99.49 |
|
4 |
+
Sentences | 92.68 | 95.00 | 93.83 |
|
5 |
+
Words | 95.71 | 94.22 | 94.96 |
|
6 |
+
UPOS | 88.10 | 86.73 | 87.41 | 92.05
|
7 |
+
XPOS | 81.00 | 79.74 | 80.36 | 84.63
|
8 |
+
UFeats | 84.22 | 82.91 | 83.56 | 87.99
|
9 |
+
AllTags | 79.42 | 78.18 | 78.79 | 82.98
|
10 |
+
Lemmas | 90.81 | 89.39 | 90.09 | 94.88
|
11 |
+
UAS | 72.42 | 71.29 | 71.85 | 75.67
|
12 |
+
LAS | 66.75 | 65.71 | 66.23 | 69.74
|
13 |
+
ELAS | 65.46 | 63.05 | 64.23 | 68.46
|
14 |
+
EULAS | 66.94 | 64.47 | 65.68 | 70.01
|
15 |
+
CLAS | 64.02 | 63.41 | 63.71 | 67.58
|
16 |
+
MLAS | 54.63 | 54.11 | 54.37 | 57.67
|
17 |
+
BLEX | 60.04 | 59.47 | 59.75 | 63.39
|
sysoutputs/orange_deskin/test_02/uk.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/orange_deskin/test_02/uk.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.79 | 99.79 | 99.79 |
|
4 |
+
Sentences | 94.89 | 95.74 | 95.31 |
|
5 |
+
Words | 99.75 | 99.76 | 99.75 |
|
6 |
+
UPOS | 98.02 | 98.03 | 98.03 | 98.27
|
7 |
+
XPOS | 94.68 | 94.70 | 94.69 | 94.92
|
8 |
+
UFeats | 94.73 | 94.74 | 94.74 | 94.97
|
9 |
+
AllTags | 93.70 | 93.71 | 93.70 | 93.93
|
10 |
+
Lemmas | 97.46 | 97.48 | 97.47 | 97.71
|
11 |
+
UAS | 90.71 | 90.72 | 90.72 | 90.94
|
12 |
+
LAS | 88.59 | 88.60 | 88.59 | 88.81
|
13 |
+
ELAS | 87.34 | 82.09 | 84.64 | 89.10
|
14 |
+
EULAS | 88.20 | 82.90 | 85.46 | 89.97
|
15 |
+
CLAS | 86.14 | 85.82 | 85.98 | 86.00
|
16 |
+
MLAS | 79.41 | 79.12 | 79.26 | 79.28
|
17 |
+
BLEX | 83.53 | 83.22 | 83.38 | 83.40
|
sysoutputs/robertnlp/mainsubmission/en.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/robertnlp/mainsubmission/en.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.52 | 99.43 | 99.48 |
|
4 |
+
Sentences | 87.70 | 84.79 | 86.22 |
|
5 |
+
Words | 99.52 | 99.43 | 99.48 |
|
6 |
+
UPOS | 95.79 | 95.70 | 95.75 | 96.25
|
7 |
+
XPOS | 0.00 | 0.00 | 0.00 | 0.00
|
8 |
+
UFeats | 94.75 | 94.66 | 94.71 | 95.21
|
9 |
+
AllTags | 0.00 | 0.00 | 0.00 | 0.00
|
10 |
+
Lemmas | 95.67 | 95.57 | 95.62 | 96.12
|
11 |
+
UAS | 89.48 | 89.39 | 89.43 | 89.90
|
12 |
+
LAS | 86.91 | 86.83 | 86.87 | 87.33
|
13 |
+
ELAS | 88.55 | 89.33 | 88.94 | 93.80
|
14 |
+
EULAS | 88.95 | 89.74 | 89.35 | 94.23
|
15 |
+
CLAS | 85.05 | 84.91 | 84.98 | 85.45
|
16 |
+
MLAS | 76.66 | 76.54 | 76.60 | 77.02
|
17 |
+
BLEX | 80.55 | 80.42 | 80.48 | 80.93
|
sysoutputs/robertnlp/mainsubmission/pertreebank/ar_padt-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/bg_btb-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_cac-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_fictree-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pdt-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pud-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test-sys.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.23 | 99.07 | 99.15 |
|
4 |
+
Sentences | 83.80 | 78.72 | 81.18 |
|
5 |
+
Words | 99.23 | 99.07 | 99.15 |
|
6 |
+
UPOS | 95.53 | 95.37 | 95.45 | 96.26
|
7 |
+
XPOS | 0.00 | 0.00 | 0.00 | 0.00
|
8 |
+
UFeats | 95.67 | 95.51 | 95.59 | 96.40
|
9 |
+
AllTags | 0.00 | 0.00 | 0.00 | 0.00
|
10 |
+
Lemmas | 96.78 | 96.62 | 96.70 | 97.53
|
11 |
+
UAS | 88.13 | 87.99 | 88.06 | 88.81
|
12 |
+
LAS | 85.71 | 85.56 | 85.64 | 86.37
|
13 |
+
ELAS | 87.61 | 88.51 | 88.06 | 93.03
|
14 |
+
EULAS | 87.96 | 88.87 | 88.41 | 93.40
|
15 |
+
CLAS | 83.02 | 82.96 | 82.99 | 83.79
|
16 |
+
MLAS | 77.33 | 77.27 | 77.30 | 78.04
|
17 |
+
BLEX | 80.57 | 80.52 | 80.55 | 81.32
|
sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test-sys.conllu
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test.eval.log
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Metric | Precision | Recall | F1 Score | AligndAcc
|
2 |
+
-----------+-----------+-----------+-----------+-----------
|
3 |
+
Tokens | 99.87 | 99.85 | 99.86 |
|
4 |
+
Sentences | 95.12 | 97.40 | 96.25 |
|
5 |
+
Words | 99.87 | 99.85 | 99.86 |
|
6 |
+
UPOS | 96.11 | 96.09 | 96.10 | 96.24
|
7 |
+
XPOS | 0.00 | 0.00 | 0.00 | 0.00
|
8 |
+
UFeats | 93.67 | 93.66 | 93.66 | 93.80
|
9 |
+
AllTags | 0.00 | 0.00 | 0.00 | 0.00
|
10 |
+
Lemmas | 94.35 | 94.33 | 94.34 | 94.47
|
11 |
+
UAS | 91.06 | 91.05 | 91.06 | 91.18
|
12 |
+
LAS | 88.34 | 88.33 | 88.33 | 88.46
|
13 |
+
ELAS | 89.65 | 90.30 | 89.97 | 94.72
|
14 |
+
EULAS | 90.12 | 90.78 | 90.44 | 95.21
|
15 |
+
CLAS | 87.60 | 87.36 | 87.48 | 87.52
|
16 |
+
MLAS | 75.83 | 75.61 | 75.72 | 75.75
|
17 |
+
BLEX | 80.52 | 80.29 | 80.40 | 80.44
|
sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test.eval.log
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test-sys.conllu
ADDED
File without changes
|
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test.eval.log
ADDED
File without changes
|