de-francophones commited on
Commit
398bba1
1 Parent(s): 9ea2122

77d50b733a6e7b3a3c6cc62c3261e6d8994fbfea4918b7b2b59565c148e9a6e9

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. sysoutputs/orange_deskin/test_02/pertreebank/ta_ttb-ud-test.eval.log +17 -0
  3. sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test-sys.conllu +0 -0
  4. sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test.eval.log +17 -0
  5. sysoutputs/orange_deskin/test_02/pl.conllu +0 -0
  6. sysoutputs/orange_deskin/test_02/pl.eval.log +17 -0
  7. sysoutputs/orange_deskin/test_02/ru.conllu +3 -0
  8. sysoutputs/orange_deskin/test_02/ru.eval.log +17 -0
  9. sysoutputs/orange_deskin/test_02/sk.conllu +0 -0
  10. sysoutputs/orange_deskin/test_02/sk.eval.log +17 -0
  11. sysoutputs/orange_deskin/test_02/sv.conllu +0 -0
  12. sysoutputs/orange_deskin/test_02/sv.eval.log +17 -0
  13. sysoutputs/orange_deskin/test_02/ta.conllu +0 -0
  14. sysoutputs/orange_deskin/test_02/ta.eval.log +17 -0
  15. sysoutputs/orange_deskin/test_02/uk.conllu +0 -0
  16. sysoutputs/orange_deskin/test_02/uk.eval.log +17 -0
  17. sysoutputs/robertnlp/mainsubmission/en.conllu +0 -0
  18. sysoutputs/robertnlp/mainsubmission/en.eval.log +17 -0
  19. sysoutputs/robertnlp/mainsubmission/pertreebank/ar_padt-ud-test.eval.log +0 -0
  20. sysoutputs/robertnlp/mainsubmission/pertreebank/bg_btb-ud-test.eval.log +0 -0
  21. sysoutputs/robertnlp/mainsubmission/pertreebank/cs_cac-ud-test.eval.log +0 -0
  22. sysoutputs/robertnlp/mainsubmission/pertreebank/cs_fictree-ud-test.eval.log +0 -0
  23. sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pdt-ud-test.eval.log +0 -0
  24. sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pud-ud-test.eval.log +0 -0
  25. sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test-sys.conllu +0 -0
  26. sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test.eval.log +17 -0
  27. sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test-sys.conllu +0 -0
  28. sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test.eval.log +17 -0
  29. sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test-sys.conllu +0 -0
  30. sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test.eval.log +0 -0
  31. sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test-sys.conllu +0 -0
  32. sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test.eval.log +0 -0
  33. sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test-sys.conllu +0 -0
  34. sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test.eval.log +0 -0
  35. sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test-sys.conllu +0 -0
  36. sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test.eval.log +0 -0
  37. sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test-sys.conllu +0 -0
  38. sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test.eval.log +0 -0
  39. sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test-sys.conllu +0 -0
  40. sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test.eval.log +0 -0
  41. sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test-sys.conllu +0 -0
  42. sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test.eval.log +0 -0
  43. sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test-sys.conllu +0 -0
  44. sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test.eval.log +0 -0
  45. sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test-sys.conllu +0 -0
  46. sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test.eval.log +0 -0
  47. sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test-sys.conllu +0 -0
  48. sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test.eval.log +0 -0
  49. sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test-sys.conllu +0 -0
  50. sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test.eval.log +0 -0
.gitattributes CHANGED
@@ -95,3 +95,4 @@ sysoutputs/koebsala/udpipe_test_20200425_134739/ru.conllu filter=lfs diff=lfs me
95
  sysoutputs/orange_deskin/test_02/cs.conllu filter=lfs diff=lfs merge=lfs -text
96
  sysoutputs/orange_deskin/test_02/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
97
  sysoutputs/orange_deskin/test_02/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
 
 
95
  sysoutputs/orange_deskin/test_02/cs.conllu filter=lfs diff=lfs merge=lfs -text
96
  sysoutputs/orange_deskin/test_02/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
97
  sysoutputs/orange_deskin/test_02/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
98
+ sysoutputs/orange_deskin/test_02/ru.conllu filter=lfs diff=lfs merge=lfs -text
sysoutputs/orange_deskin/test_02/pertreebank/ta_ttb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.38 | 99.60 | 99.49 |
4
+ Sentences | 92.68 | 95.00 | 93.83 |
5
+ Words | 95.71 | 94.22 | 94.96 |
6
+ UPOS | 88.10 | 86.73 | 87.41 | 92.05
7
+ XPOS | 81.00 | 79.74 | 80.36 | 84.63
8
+ UFeats | 84.22 | 82.91 | 83.56 | 87.99
9
+ AllTags | 79.42 | 78.18 | 78.79 | 82.98
10
+ Lemmas | 90.81 | 89.39 | 90.09 | 94.88
11
+ UAS | 72.42 | 71.29 | 71.85 | 75.67
12
+ LAS | 66.75 | 65.71 | 66.23 | 69.74
13
+ ELAS | 65.46 | 63.14 | 64.28 | 68.46
14
+ EULAS | 66.94 | 64.57 | 65.73 | 70.01
15
+ CLAS | 64.02 | 63.41 | 63.71 | 67.58
16
+ MLAS | 54.63 | 54.11 | 54.37 | 57.67
17
+ BLEX | 60.04 | 59.47 | 59.75 | 63.39
sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/orange_deskin/test_02/pertreebank/uk_iu-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.79 | 99.79 | 99.79 |
4
+ Sentences | 94.89 | 95.74 | 95.31 |
5
+ Words | 99.75 | 99.76 | 99.75 |
6
+ UPOS | 98.02 | 98.03 | 98.03 | 98.27
7
+ XPOS | 94.68 | 94.70 | 94.69 | 94.92
8
+ UFeats | 94.73 | 94.74 | 94.74 | 94.97
9
+ AllTags | 93.70 | 93.71 | 93.70 | 93.93
10
+ Lemmas | 97.46 | 97.48 | 97.47 | 97.71
11
+ UAS | 90.71 | 90.72 | 90.72 | 90.94
12
+ LAS | 88.59 | 88.60 | 88.59 | 88.81
13
+ ELAS | 87.34 | 82.09 | 84.64 | 89.10
14
+ EULAS | 88.20 | 82.90 | 85.46 | 89.97
15
+ CLAS | 86.14 | 85.82 | 85.98 | 86.00
16
+ MLAS | 79.41 | 79.12 | 79.26 | 79.28
17
+ BLEX | 83.53 | 83.22 | 83.38 | 83.40
sysoutputs/orange_deskin/test_02/pl.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/orange_deskin/test_02/pl.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.30 | 99.36 | 99.33 |
4
+ Sentences | 96.80 | 96.66 | 96.73 |
5
+ Words | 99.74 | 99.83 | 99.79 |
6
+ UPOS | 98.53 | 98.63 | 98.58 | 98.79
7
+ XPOS | 94.31 | 94.40 | 94.36 | 94.56
8
+ UFeats | 92.63 | 92.73 | 92.68 | 92.88
9
+ AllTags | 90.11 | 90.20 | 90.16 | 90.35
10
+ Lemmas | 97.96 | 98.06 | 98.01 | 98.22
11
+ UAS | 94.21 | 94.30 | 94.26 | 94.46
12
+ LAS | 92.06 | 92.15 | 92.11 | 92.31
13
+ ELAS | 82.07 | 78.78 | 80.39 | 84.22
14
+ EULAS | 91.02 | 87.37 | 89.15 | 93.40
15
+ CLAS | 90.57 | 90.58 | 90.58 | 90.76
16
+ MLAS | 80.97 | 80.98 | 80.97 | 81.14
17
+ BLEX | 88.48 | 88.49 | 88.49 | 88.67
sysoutputs/orange_deskin/test_02/ru.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02dc2bbabd187989b6cfe22c675fb105c185c5d4e692035f38d4690d28ab888f
3
+ size 11100718
sysoutputs/orange_deskin/test_02/ru.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.37 | 99.58 | 99.48 |
4
+ Sentences | 97.85 | 98.01 | 97.93 |
5
+ Words | 99.37 | 99.58 | 99.48 |
6
+ UPOS | 98.69 | 98.90 | 98.80 | 99.31
7
+ XPOS | 99.37 | 99.58 | 99.48 | 100.00
8
+ UFeats | 97.65 | 97.86 | 97.76 | 98.27
9
+ AllTags | 97.47 | 97.68 | 97.58 | 98.09
10
+ Lemmas | 98.15 | 98.36 | 98.26 | 98.77
11
+ UAS | 94.31 | 94.51 | 94.41 | 94.90
12
+ LAS | 93.21 | 93.41 | 93.31 | 93.80
13
+ ELAS | 91.16 | 88.56 | 89.84 | 94.11
14
+ EULAS | 92.30 | 89.68 | 90.97 | 95.29
15
+ CLAS | 92.16 | 92.31 | 92.24 | 92.81
16
+ MLAS | 89.85 | 90.00 | 89.93 | 90.49
17
+ BLEX | 90.67 | 90.82 | 90.75 | 91.31
sysoutputs/orange_deskin/test_02/sk.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/orange_deskin/test_02/sk.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.99 | 99.98 | 99.99 |
4
+ Sentences | 84.62 | 83.51 | 84.06 |
5
+ Words | 99.99 | 99.98 | 99.99 |
6
+ UPOS | 97.17 | 97.16 | 97.16 | 97.17
7
+ XPOS | 88.14 | 88.13 | 88.14 | 88.15
8
+ UFeats | 92.89 | 92.88 | 92.89 | 92.90
9
+ AllTags | 87.02 | 87.01 | 87.02 | 87.03
10
+ Lemmas | 96.63 | 96.62 | 96.63 | 96.64
11
+ UAS | 91.20 | 91.20 | 91.20 | 91.21
12
+ LAS | 89.06 | 89.05 | 89.06 | 89.07
13
+ ELAS | 85.89 | 82.87 | 84.36 | 88.35
14
+ EULAS | 87.75 | 84.66 | 86.17 | 90.26
15
+ CLAS | 87.35 | 87.38 | 87.37 | 87.39
16
+ MLAS | 78.01 | 78.04 | 78.03 | 78.05
17
+ BLEX | 83.72 | 83.75 | 83.73 | 83.76
sysoutputs/orange_deskin/test_02/sv.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/orange_deskin/test_02/sv.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.69 | 99.76 | 99.72 |
4
+ Sentences | 85.20 | 91.03 | 88.02 |
5
+ Words | 99.69 | 99.76 | 99.72 |
6
+ UPOS | 97.48 | 97.54 | 97.51 | 97.78
7
+ XPOS | 95.61 | 95.68 | 95.64 | 95.91
8
+ UFeats | 88.98 | 89.04 | 89.01 | 89.26
9
+ AllTags | 87.66 | 87.72 | 87.69 | 87.94
10
+ Lemmas | 93.36 | 93.42 | 93.39 | 93.65
11
+ UAS | 88.62 | 88.68 | 88.65 | 88.89
12
+ LAS | 85.88 | 85.94 | 85.91 | 86.15
13
+ ELAS | 83.98 | 82.56 | 83.27 | 87.66
14
+ EULAS | 85.09 | 83.65 | 84.36 | 88.82
15
+ CLAS | 84.30 | 84.26 | 84.28 | 84.51
16
+ MLAS | 69.68 | 69.65 | 69.67 | 69.86
17
+ BLEX | 77.73 | 77.70 | 77.72 | 77.93
sysoutputs/orange_deskin/test_02/ta.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/orange_deskin/test_02/ta.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.38 | 99.60 | 99.49 |
4
+ Sentences | 92.68 | 95.00 | 93.83 |
5
+ Words | 95.71 | 94.22 | 94.96 |
6
+ UPOS | 88.10 | 86.73 | 87.41 | 92.05
7
+ XPOS | 81.00 | 79.74 | 80.36 | 84.63
8
+ UFeats | 84.22 | 82.91 | 83.56 | 87.99
9
+ AllTags | 79.42 | 78.18 | 78.79 | 82.98
10
+ Lemmas | 90.81 | 89.39 | 90.09 | 94.88
11
+ UAS | 72.42 | 71.29 | 71.85 | 75.67
12
+ LAS | 66.75 | 65.71 | 66.23 | 69.74
13
+ ELAS | 65.46 | 63.05 | 64.23 | 68.46
14
+ EULAS | 66.94 | 64.47 | 65.68 | 70.01
15
+ CLAS | 64.02 | 63.41 | 63.71 | 67.58
16
+ MLAS | 54.63 | 54.11 | 54.37 | 57.67
17
+ BLEX | 60.04 | 59.47 | 59.75 | 63.39
sysoutputs/orange_deskin/test_02/uk.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/orange_deskin/test_02/uk.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.79 | 99.79 | 99.79 |
4
+ Sentences | 94.89 | 95.74 | 95.31 |
5
+ Words | 99.75 | 99.76 | 99.75 |
6
+ UPOS | 98.02 | 98.03 | 98.03 | 98.27
7
+ XPOS | 94.68 | 94.70 | 94.69 | 94.92
8
+ UFeats | 94.73 | 94.74 | 94.74 | 94.97
9
+ AllTags | 93.70 | 93.71 | 93.70 | 93.93
10
+ Lemmas | 97.46 | 97.48 | 97.47 | 97.71
11
+ UAS | 90.71 | 90.72 | 90.72 | 90.94
12
+ LAS | 88.59 | 88.60 | 88.59 | 88.81
13
+ ELAS | 87.34 | 82.09 | 84.64 | 89.10
14
+ EULAS | 88.20 | 82.90 | 85.46 | 89.97
15
+ CLAS | 86.14 | 85.82 | 85.98 | 86.00
16
+ MLAS | 79.41 | 79.12 | 79.26 | 79.28
17
+ BLEX | 83.53 | 83.22 | 83.38 | 83.40
sysoutputs/robertnlp/mainsubmission/en.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/robertnlp/mainsubmission/en.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.52 | 99.43 | 99.48 |
4
+ Sentences | 87.70 | 84.79 | 86.22 |
5
+ Words | 99.52 | 99.43 | 99.48 |
6
+ UPOS | 95.79 | 95.70 | 95.75 | 96.25
7
+ XPOS | 0.00 | 0.00 | 0.00 | 0.00
8
+ UFeats | 94.75 | 94.66 | 94.71 | 95.21
9
+ AllTags | 0.00 | 0.00 | 0.00 | 0.00
10
+ Lemmas | 95.67 | 95.57 | 95.62 | 96.12
11
+ UAS | 89.48 | 89.39 | 89.43 | 89.90
12
+ LAS | 86.91 | 86.83 | 86.87 | 87.33
13
+ ELAS | 88.55 | 89.33 | 88.94 | 93.80
14
+ EULAS | 88.95 | 89.74 | 89.35 | 94.23
15
+ CLAS | 85.05 | 84.91 | 84.98 | 85.45
16
+ MLAS | 76.66 | 76.54 | 76.60 | 77.02
17
+ BLEX | 80.55 | 80.42 | 80.48 | 80.93
sysoutputs/robertnlp/mainsubmission/pertreebank/ar_padt-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/bg_btb-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_cac-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_fictree-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pdt-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/cs_pud-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/robertnlp/mainsubmission/pertreebank/en_ewt-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.23 | 99.07 | 99.15 |
4
+ Sentences | 83.80 | 78.72 | 81.18 |
5
+ Words | 99.23 | 99.07 | 99.15 |
6
+ UPOS | 95.53 | 95.37 | 95.45 | 96.26
7
+ XPOS | 0.00 | 0.00 | 0.00 | 0.00
8
+ UFeats | 95.67 | 95.51 | 95.59 | 96.40
9
+ AllTags | 0.00 | 0.00 | 0.00 | 0.00
10
+ Lemmas | 96.78 | 96.62 | 96.70 | 97.53
11
+ UAS | 88.13 | 87.99 | 88.06 | 88.81
12
+ LAS | 85.71 | 85.56 | 85.64 | 86.37
13
+ ELAS | 87.61 | 88.51 | 88.06 | 93.03
14
+ EULAS | 87.96 | 88.87 | 88.41 | 93.40
15
+ CLAS | 83.02 | 82.96 | 82.99 | 83.79
16
+ MLAS | 77.33 | 77.27 | 77.30 | 78.04
17
+ BLEX | 80.57 | 80.52 | 80.55 | 81.32
sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/robertnlp/mainsubmission/pertreebank/en_pud-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.87 | 99.85 | 99.86 |
4
+ Sentences | 95.12 | 97.40 | 96.25 |
5
+ Words | 99.87 | 99.85 | 99.86 |
6
+ UPOS | 96.11 | 96.09 | 96.10 | 96.24
7
+ XPOS | 0.00 | 0.00 | 0.00 | 0.00
8
+ UFeats | 93.67 | 93.66 | 93.66 | 93.80
9
+ AllTags | 0.00 | 0.00 | 0.00 | 0.00
10
+ Lemmas | 94.35 | 94.33 | 94.34 | 94.47
11
+ UAS | 91.06 | 91.05 | 91.06 | 91.18
12
+ LAS | 88.34 | 88.33 | 88.33 | 88.46
13
+ ELAS | 89.65 | 90.30 | 89.97 | 94.72
14
+ EULAS | 90.12 | 90.78 | 90.44 | 95.21
15
+ CLAS | 87.60 | 87.36 | 87.48 | 87.52
16
+ MLAS | 75.83 | 75.61 | 75.72 | 75.75
17
+ BLEX | 80.52 | 80.29 | 80.40 | 80.44
sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/et_edt-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/et_ewt-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_pud-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fi_tdt-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_fqb-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/fr_sequoia-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/it_isdt-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/lt_alksnis-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/lv_lvtb-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_alpino-ud-test.eval.log ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test-sys.conllu ADDED
File without changes
sysoutputs/robertnlp/mainsubmission/pertreebank/nl_lassysmall-ud-test.eval.log ADDED
File without changes