de-francophones commited on
Commit
90e05ae
1 Parent(s): 087a448

9e44136cd7f90058ca44f3ea215da769cdf745c435fc1b4599f34a68b569d82c

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +3 -0
  2. sysoutputs/adapt/test/pertreebank/fr_fqb-ud-test.eval.log +17 -0
  3. sysoutputs/adapt/test/pertreebank/fr_sequoia-ud-test-sys.conllu +0 -0
  4. sysoutputs/adapt/test/pertreebank/fr_sequoia-ud-test.eval.log +17 -0
  5. sysoutputs/adapt/test/pertreebank/it_isdt-ud-test-sys.conllu +0 -0
  6. sysoutputs/adapt/test/pertreebank/it_isdt-ud-test.eval.log +17 -0
  7. sysoutputs/adapt/test/pertreebank/lt_alksnis-ud-test-sys.conllu +0 -0
  8. sysoutputs/adapt/test/pertreebank/lt_alksnis-ud-test.eval.log +17 -0
  9. sysoutputs/adapt/test/pertreebank/lv_lvtb-ud-test-sys.conllu +0 -0
  10. sysoutputs/adapt/test/pertreebank/lv_lvtb-ud-test.eval.log +17 -0
  11. sysoutputs/adapt/test/pertreebank/nl_alpino-ud-test-sys.conllu +0 -0
  12. sysoutputs/adapt/test/pertreebank/nl_alpino-ud-test.eval.log +17 -0
  13. sysoutputs/adapt/test/pertreebank/nl_lassysmall-ud-test-sys.conllu +0 -0
  14. sysoutputs/adapt/test/pertreebank/nl_lassysmall-ud-test.eval.log +17 -0
  15. sysoutputs/adapt/test/pertreebank/pl_lfg-ud-test-sys.conllu +0 -0
  16. sysoutputs/adapt/test/pertreebank/pl_lfg-ud-test.eval.log +17 -0
  17. sysoutputs/adapt/test/pertreebank/pl_pdb-ud-test-sys.conllu +0 -0
  18. sysoutputs/adapt/test/pertreebank/pl_pdb-ud-test.eval.log +17 -0
  19. sysoutputs/adapt/test/pertreebank/pl_pud-ud-test-sys.conllu +0 -0
  20. sysoutputs/adapt/test/pertreebank/pl_pud-ud-test.eval.log +17 -0
  21. sysoutputs/adapt/test/pertreebank/ru_syntagrus-ud-test-sys.conllu +3 -0
  22. sysoutputs/adapt/test/pertreebank/ru_syntagrus-ud-test.eval.log +17 -0
  23. sysoutputs/adapt/test/pertreebank/sk_snk-ud-test-sys.conllu +0 -0
  24. sysoutputs/adapt/test/pertreebank/sk_snk-ud-test.eval.log +17 -0
  25. sysoutputs/adapt/test/pertreebank/sv_pud-ud-test-sys.conllu +0 -0
  26. sysoutputs/adapt/test/pertreebank/sv_pud-ud-test.eval.log +17 -0
  27. sysoutputs/adapt/test/pertreebank/sv_talbanken-ud-test-sys.conllu +0 -0
  28. sysoutputs/adapt/test/pertreebank/sv_talbanken-ud-test.eval.log +17 -0
  29. sysoutputs/adapt/test/pertreebank/ta_ttb-ud-test-sys.conllu +0 -0
  30. sysoutputs/adapt/test/pertreebank/ta_ttb-ud-test.eval.log +17 -0
  31. sysoutputs/adapt/test/pertreebank/uk_iu-ud-test-sys.conllu +0 -0
  32. sysoutputs/adapt/test/pertreebank/uk_iu-ud-test.eval.log +17 -0
  33. sysoutputs/adapt/test/pl.conllu +0 -0
  34. sysoutputs/adapt/test/pl.eval.log +17 -0
  35. sysoutputs/adapt/test/ru.conllu +3 -0
  36. sysoutputs/adapt/test/ru.eval.log +17 -0
  37. sysoutputs/adapt/test/sk.conllu +0 -0
  38. sysoutputs/adapt/test/sk.eval.log +17 -0
  39. sysoutputs/adapt/test/sv.conllu +0 -0
  40. sysoutputs/adapt/test/sv.eval.log +17 -0
  41. sysoutputs/adapt/test/ta.conllu +0 -0
  42. sysoutputs/adapt/test/ta.eval.log +17 -0
  43. sysoutputs/adapt/test/uk.conllu +0 -0
  44. sysoutputs/adapt/test/uk.eval.log +17 -0
  45. sysoutputs/baseline1/test/ar.conllu +0 -0
  46. sysoutputs/baseline1/test/ar.eval.log +17 -0
  47. sysoutputs/baseline1/test/bg.conllu +0 -0
  48. sysoutputs/baseline1/test/bg.eval.log +17 -0
  49. sysoutputs/baseline1/test/cs.conllu +3 -0
  50. sysoutputs/baseline1/test/cs.eval.log +17 -0
.gitattributes CHANGED
@@ -62,3 +62,6 @@ dev-gold/ru.conllu filter=lfs diff=lfs merge=lfs -text
62
  dev-gold/ru.nen.conllu filter=lfs diff=lfs merge=lfs -text
63
  sysoutputs/adapt/test/cs.conllu filter=lfs diff=lfs merge=lfs -text
64
  sysoutputs/adapt/test/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
 
 
 
 
62
  dev-gold/ru.nen.conllu filter=lfs diff=lfs merge=lfs -text
63
  sysoutputs/adapt/test/cs.conllu filter=lfs diff=lfs merge=lfs -text
64
  sysoutputs/adapt/test/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
65
+ sysoutputs/adapt/test/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
66
+ sysoutputs/adapt/test/ru.conllu filter=lfs diff=lfs merge=lfs -text
67
+ sysoutputs/baseline1/test/cs.conllu filter=lfs diff=lfs merge=lfs -text
sysoutputs/adapt/test/pertreebank/fr_fqb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.58 | 99.61 | 99.59 |
4
+ Sentences | 94.35 | 97.03 | 95.67 |
5
+ Words | 99.18 | 99.41 | 99.30 |
6
+ UPOS | 94.34 | 94.56 | 94.45 | 95.12
7
+ XPOS | 74.88 | 75.05 | 74.97 | 75.50
8
+ UFeats | 75.11 | 75.28 | 75.19 | 75.73
9
+ AllTags | 54.29 | 54.42 | 54.36 | 54.74
10
+ Lemmas | 96.32 | 96.54 | 96.43 | 97.11
11
+ UAS | 92.09 | 92.30 | 92.19 | 92.85
12
+ LAS | 86.93 | 87.13 | 87.03 | 87.65
13
+ ELAS | 67.18 | 83.10 | 74.30 | 84.66
14
+ EULAS | 67.78 | 83.84 | 74.96 | 85.41
15
+ CLAS | 82.72 | 80.88 | 81.79 | 81.39
16
+ MLAS | 47.79 | 46.72 | 47.25 | 47.02
17
+ BLEX | 79.07 | 77.31 | 78.18 | 77.80
sysoutputs/adapt/test/pertreebank/fr_sequoia-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/fr_sequoia-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.76 | 99.83 | 99.79 |
4
+ Sentences | 89.09 | 85.96 | 87.50 |
5
+ Words | 98.83 | 99.34 | 99.09 |
6
+ UPOS | 96.69 | 97.19 | 96.94 | 97.84
7
+ XPOS | 73.21 | 73.59 | 73.40 | 74.07
8
+ UFeats | 80.84 | 81.26 | 81.05 | 81.80
9
+ AllTags | 57.93 | 58.23 | 58.08 | 58.62
10
+ Lemmas | 96.72 | 97.22 | 96.97 | 97.87
11
+ UAS | 87.91 | 88.37 | 88.14 | 88.95
12
+ LAS | 85.19 | 85.63 | 85.41 | 86.20
13
+ ELAS | 69.19 | 87.91 | 77.44 | 91.97
14
+ EULAS | 69.60 | 88.43 | 77.89 | 92.51
15
+ CLAS | 83.84 | 83.84 | 83.84 | 84.08
16
+ MLAS | 57.17 | 57.17 | 57.17 | 57.34
17
+ BLEX | 81.31 | 81.31 | 81.31 | 81.54
sysoutputs/adapt/test/pertreebank/it_isdt-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/it_isdt-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.91 | 99.95 | 99.93 |
4
+ Sentences | 98.76 | 98.76 | 98.76 |
5
+ Words | 99.82 | 99.87 | 99.84 |
6
+ UPOS | 98.43 | 98.47 | 98.45 | 98.61
7
+ XPOS | 98.29 | 98.34 | 98.32 | 98.47
8
+ UFeats | 98.25 | 98.30 | 98.28 | 98.43
9
+ AllTags | 97.62 | 97.67 | 97.64 | 97.80
10
+ Lemmas | 98.66 | 98.70 | 98.68 | 98.84
11
+ UAS | 93.88 | 93.92 | 93.90 | 94.05
12
+ LAS | 92.18 | 92.22 | 92.20 | 92.35
13
+ ELAS | 59.79 | 90.42 | 71.98 | 94.27
14
+ EULAS | 60.47 | 91.44 | 72.80 | 95.34
15
+ CLAS | 88.67 | 88.39 | 88.53 | 88.54
16
+ MLAS | 86.07 | 85.80 | 85.93 | 85.95
17
+ BLEX | 87.22 | 86.95 | 87.08 | 87.10
sysoutputs/adapt/test/pertreebank/lt_alksnis-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/lt_alksnis-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.91 | 99.91 | 99.91 |
4
+ Sentences | 90.03 | 85.82 | 87.87 |
5
+ Words | 99.91 | 99.91 | 99.91 |
6
+ UPOS | 93.92 | 93.92 | 93.92 | 94.01
7
+ XPOS | 85.97 | 85.97 | 85.97 | 86.05
8
+ UFeats | 87.01 | 87.01 | 87.01 | 87.09
9
+ AllTags | 84.74 | 84.74 | 84.74 | 84.82
10
+ Lemmas | 92.26 | 92.26 | 92.26 | 92.35
11
+ UAS | 76.72 | 76.72 | 76.72 | 76.79
12
+ LAS | 72.10 | 72.10 | 72.10 | 72.17
13
+ ELAS | 50.64 | 68.84 | 58.36 | 76.16
14
+ EULAS | 52.39 | 71.21 | 60.37 | 78.78
15
+ CLAS | 70.06 | 69.68 | 69.87 | 69.73
16
+ MLAS | 58.55 | 58.23 | 58.39 | 58.27
17
+ BLEX | 64.49 | 64.14 | 64.32 | 64.19
sysoutputs/adapt/test/pertreebank/lv_lvtb-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/lv_lvtb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.40 | 99.26 | 99.33 |
4
+ Sentences | 98.42 | 99.07 | 98.74 |
5
+ Words | 99.40 | 99.26 | 99.33 |
6
+ UPOS | 96.57 | 96.44 | 96.50 | 97.16
7
+ XPOS | 90.09 | 89.96 | 90.03 | 90.63
8
+ UFeats | 94.25 | 94.12 | 94.18 | 94.82
9
+ AllTags | 89.51 | 89.38 | 89.45 | 90.05
10
+ Lemmas | 96.19 | 96.06 | 96.13 | 96.77
11
+ UAS | 89.39 | 89.26 | 89.32 | 89.93
12
+ LAS | 86.51 | 86.39 | 86.45 | 87.03
13
+ ELAS | 65.46 | 81.03 | 72.41 | 88.07
14
+ EULAS | 66.01 | 81.72 | 73.03 | 88.82
15
+ CLAS | 84.54 | 84.26 | 84.40 | 85.15
16
+ MLAS | 77.03 | 76.78 | 76.90 | 77.59
17
+ BLEX | 81.27 | 81.00 | 81.13 | 81.85
sysoutputs/adapt/test/pertreebank/nl_alpino-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/nl_alpino-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.86 | 99.81 | 99.83 |
4
+ Sentences | 89.97 | 87.25 | 88.59 |
5
+ Words | 99.86 | 99.81 | 99.83 |
6
+ UPOS | 97.45 | 97.40 | 97.42 | 97.59
7
+ XPOS | 96.41 | 96.37 | 96.39 | 96.55
8
+ UFeats | 97.51 | 97.47 | 97.49 | 97.65
9
+ AllTags | 95.89 | 95.84 | 95.87 | 96.03
10
+ Lemmas | 97.46 | 97.42 | 97.44 | 97.61
11
+ UAS | 91.55 | 91.51 | 91.53 | 91.68
12
+ LAS | 89.12 | 89.08 | 89.10 | 89.25
13
+ ELAS | 60.36 | 87.49 | 71.44 | 91.66
14
+ EULAS | 60.94 | 88.33 | 72.12 | 92.54
15
+ CLAS | 84.74 | 84.76 | 84.75 | 84.94
16
+ MLAS | 81.03 | 81.05 | 81.04 | 81.22
17
+ BLEX | 81.76 | 81.78 | 81.77 | 81.96
sysoutputs/adapt/test/pertreebank/nl_lassysmall-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/nl_lassysmall-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.66 | 99.65 | 99.65 |
4
+ Sentences | 67.53 | 44.46 | 53.62 |
5
+ Words | 99.66 | 99.65 | 99.65 |
6
+ UPOS | 96.55 | 96.53 | 96.54 | 96.87
7
+ XPOS | 95.26 | 95.24 | 95.25 | 95.58
8
+ UFeats | 96.30 | 96.28 | 96.29 | 96.62
9
+ AllTags | 94.63 | 94.62 | 94.63 | 94.95
10
+ Lemmas | 97.17 | 97.15 | 97.16 | 97.50
11
+ UAS | 84.94 | 84.93 | 84.93 | 85.23
12
+ LAS | 82.45 | 82.43 | 82.44 | 82.73
13
+ ELAS | 54.50 | 77.59 | 64.03 | 81.47
14
+ EULAS | 55.09 | 78.44 | 64.72 | 82.35
15
+ CLAS | 76.25 | 75.65 | 75.95 | 75.91
16
+ MLAS | 71.83 | 71.26 | 71.55 | 71.51
17
+ BLEX | 74.09 | 73.50 | 73.79 | 73.75
sysoutputs/adapt/test/pertreebank/pl_lfg-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/pl_lfg-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.82 | 99.88 | 99.85 |
4
+ Sentences | 99.54 | 99.77 | 99.65 |
5
+ Words | 99.82 | 99.88 | 99.85 |
6
+ UPOS | 98.55 | 98.60 | 98.58 | 98.72
7
+ XPOS | 90.06 | 90.11 | 90.08 | 90.22
8
+ UFeats | 84.62 | 84.66 | 84.64 | 84.77
9
+ AllTags | 78.23 | 78.27 | 78.25 | 78.37
10
+ Lemmas | 98.13 | 98.18 | 98.16 | 98.30
11
+ UAS | 95.05 | 95.10 | 95.08 | 95.22
12
+ LAS | 91.89 | 91.94 | 91.91 | 92.05
13
+ ELAS | 53.53 | 71.52 | 61.23 | 73.61
14
+ EULAS | 65.17 | 87.08 | 74.55 | 89.63
15
+ CLAS | 90.30 | 90.48 | 90.39 | 90.61
16
+ MLAS | 70.63 | 70.77 | 70.70 | 70.87
17
+ BLEX | 88.39 | 88.56 | 88.48 | 88.69
sysoutputs/adapt/test/pertreebank/pl_pdb-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/pl_pdb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.05 | 98.72 | 98.38 |
4
+ Sentences | 75.06 | 85.60 | 79.98 |
5
+ Words | 99.53 | 99.43 | 99.48 |
6
+ UPOS | 97.69 | 97.59 | 97.64 | 98.15
7
+ XPOS | 89.46 | 89.37 | 89.41 | 89.88
8
+ UFeats | 83.36 | 83.27 | 83.32 | 83.75
9
+ AllTags | 77.96 | 77.88 | 77.92 | 78.33
10
+ Lemmas | 97.63 | 97.53 | 97.58 | 98.09
11
+ UAS | 88.58 | 88.48 | 88.53 | 88.99
12
+ LAS | 84.97 | 84.88 | 84.92 | 85.37
13
+ ELAS | 58.46 | 80.34 | 67.68 | 87.16
14
+ EULAS | 61.03 | 83.87 | 70.65 | 90.99
15
+ CLAS | 83.48 | 83.44 | 83.46 | 83.92
16
+ MLAS | 64.00 | 63.97 | 63.99 | 64.34
17
+ BLEX | 81.57 | 81.53 | 81.55 | 82.01
sysoutputs/adapt/test/pertreebank/pl_pud-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/pl_pud-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.33 | 98.06 | 98.20 |
4
+ Sentences | 73.21 | 85.00 | 78.67 |
5
+ Words | 98.86 | 98.32 | 98.59 |
6
+ UPOS | 96.27 | 95.74 | 96.00 | 97.38
7
+ XPOS | 87.57 | 87.10 | 87.33 | 88.58
8
+ UFeats | 80.33 | 79.90 | 80.11 | 81.26
9
+ AllTags | 74.92 | 74.52 | 74.72 | 75.79
10
+ Lemmas | 96.18 | 95.66 | 95.92 | 97.29
11
+ UAS | 87.65 | 87.17 | 87.41 | 88.66
12
+ LAS | 83.63 | 83.17 | 83.40 | 84.60
13
+ ELAS | 56.03 | 79.22 | 65.64 | 86.59
14
+ EULAS | 58.69 | 82.99 | 68.76 | 90.71
15
+ CLAS | 81.75 | 81.78 | 81.76 | 83.02
16
+ MLAS | 60.57 | 60.59 | 60.58 | 61.51
17
+ BLEX | 78.85 | 78.88 | 78.86 | 80.07
sysoutputs/adapt/test/pertreebank/ru_syntagrus-ud-test-sys.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7594e6c4efee0f9b3c6d9562522986e24448c8c336f44b157315ba602ed0d5f
3
+ size 11455172
sysoutputs/adapt/test/pertreebank/ru_syntagrus-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.55 | 99.64 | 99.60 |
4
+ Sentences | 98.77 | 98.83 | 98.80 |
5
+ Words | 99.55 | 99.64 | 99.60 |
6
+ UPOS | 98.72 | 98.81 | 98.76 | 99.16
7
+ XPOS | 99.55 | 99.64 | 99.60 | 100.00
8
+ UFeats | 97.37 | 97.46 | 97.42 | 97.81
9
+ AllTags | 97.15 | 97.23 | 97.19 | 97.59
10
+ Lemmas | 98.18 | 98.27 | 98.23 | 98.62
11
+ UAS | 93.46 | 93.55 | 93.50 | 93.88
12
+ LAS | 92.12 | 92.20 | 92.16 | 92.54
13
+ ELAS | 65.00 | 89.39 | 75.27 | 94.94
14
+ EULAS | 65.65 | 90.28 | 76.02 | 95.88
15
+ CLAS | 90.84 | 90.86 | 90.85 | 91.26
16
+ MLAS | 88.14 | 88.16 | 88.15 | 88.54
17
+ BLEX | 89.25 | 89.27 | 89.26 | 89.66
sysoutputs/adapt/test/pertreebank/sk_snk-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/sk_snk-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 85.65 | 84.92 | 85.28 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 97.37 | 97.37 | 97.37 | 97.37
7
+ XPOS | 88.06 | 88.06 | 88.06 | 88.06
8
+ UFeats | 92.81 | 92.81 | 92.81 | 92.81
9
+ AllTags | 87.07 | 87.07 | 87.07 | 87.07
10
+ Lemmas | 96.88 | 96.88 | 96.88 | 96.88
11
+ UAS | 90.90 | 90.90 | 90.90 | 90.90
12
+ LAS | 88.84 | 88.84 | 88.84 | 88.84
13
+ ELAS | 58.96 | 81.51 | 68.43 | 86.89
14
+ EULAS | 61.20 | 84.61 | 71.02 | 90.19
15
+ CLAS | 87.28 | 87.38 | 87.33 | 87.38
16
+ MLAS | 77.85 | 77.94 | 77.89 | 77.94
17
+ BLEX | 83.79 | 83.88 | 83.84 | 83.88
sysoutputs/adapt/test/pertreebank/sv_pud-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/sv_pud-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.81 | 98.01 | 98.41 |
4
+ Sentences | 88.44 | 92.60 | 90.47 |
5
+ Words | 98.81 | 98.01 | 98.41 |
6
+ UPOS | 96.06 | 95.28 | 95.67 | 97.21
7
+ XPOS | 94.36 | 93.59 | 93.98 | 95.50
8
+ UFeats | 80.21 | 79.56 | 79.88 | 81.17
9
+ AllTags | 78.85 | 78.21 | 78.53 | 79.80
10
+ Lemmas | 88.48 | 87.76 | 88.12 | 89.54
11
+ UAS | 85.32 | 84.63 | 84.98 | 86.35
12
+ LAS | 82.05 | 81.38 | 81.72 | 83.04
13
+ ELAS | 54.65 | 79.26 | 64.70 | 85.30
14
+ EULAS | 55.55 | 80.56 | 65.76 | 86.70
15
+ CLAS | 80.06 | 80.43 | 80.25 | 81.64
16
+ MLAS | 55.49 | 55.74 | 55.61 | 56.58
17
+ BLEX | 69.89 | 70.21 | 70.05 | 71.26
sysoutputs/adapt/test/pertreebank/sv_talbanken-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/sv_talbanken-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.87 | 99.92 | 99.89 |
4
+ Sentences | 96.45 | 95.82 | 96.13 |
5
+ Words | 99.87 | 99.92 | 99.89 |
6
+ UPOS | 98.39 | 98.44 | 98.42 | 98.52
7
+ XPOS | 97.42 | 97.47 | 97.45 | 97.55
8
+ UFeats | 97.54 | 97.59 | 97.56 | 97.67
9
+ AllTags | 96.75 | 96.80 | 96.77 | 96.88
10
+ Lemmas | 98.12 | 98.17 | 98.14 | 98.25
11
+ UAS | 91.01 | 91.06 | 91.04 | 91.14
12
+ LAS | 88.65 | 88.69 | 88.67 | 88.76
13
+ ELAS | 61.24 | 86.94 | 71.86 | 92.52
14
+ EULAS | 62.10 | 88.16 | 72.87 | 93.82
15
+ CLAS | 87.13 | 86.80 | 86.96 | 86.91
16
+ MLAS | 83.38 | 83.06 | 83.22 | 83.16
17
+ BLEX | 85.05 | 84.72 | 84.89 | 84.83
sysoutputs/adapt/test/pertreebank/ta_ttb-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/ta_ttb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.88 | 99.44 | 99.16 |
4
+ Sentences | 96.72 | 98.33 | 97.52 |
5
+ Words | 95.11 | 93.92 | 94.51 |
6
+ UPOS | 83.20 | 82.15 | 82.67 | 87.47
7
+ XPOS | 73.52 | 72.60 | 73.06 | 77.30
8
+ UFeats | 77.85 | 76.87 | 77.36 | 81.85
9
+ AllTags | 72.51 | 71.59 | 72.05 | 76.23
10
+ Lemmas | 88.80 | 87.68 | 88.24 | 93.36
11
+ UAS | 67.52 | 66.67 | 67.09 | 70.99
12
+ LAS | 59.62 | 58.87 | 59.25 | 62.69
13
+ ELAS | 41.77 | 57.73 | 48.47 | 62.79
14
+ EULAS | 43.20 | 59.69 | 50.12 | 64.94
15
+ CLAS | 55.09 | 55.13 | 55.11 | 59.01
16
+ MLAS | 44.23 | 44.26 | 44.25 | 47.38
17
+ BLEX | 51.83 | 51.87 | 51.85 | 55.52
sysoutputs/adapt/test/pertreebank/uk_iu-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pertreebank/uk_iu-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.84 | 99.85 | 99.85 |
4
+ Sentences | 95.91 | 97.31 | 96.61 |
5
+ Words | 99.80 | 99.82 | 99.81 |
6
+ UPOS | 97.86 | 97.87 | 97.86 | 98.05
7
+ XPOS | 94.23 | 94.25 | 94.24 | 94.42
8
+ UFeats | 94.19 | 94.20 | 94.20 | 94.38
9
+ AllTags | 93.25 | 93.26 | 93.26 | 93.43
10
+ Lemmas | 97.63 | 97.65 | 97.64 | 97.82
11
+ UAS | 89.50 | 89.51 | 89.50 | 89.68
12
+ LAS | 87.07 | 87.09 | 87.08 | 87.25
13
+ ELAS | 56.58 | 80.44 | 66.43 | 87.26
14
+ EULAS | 57.54 | 81.80 | 67.55 | 88.73
15
+ CLAS | 84.45 | 84.07 | 84.26 | 84.23
16
+ MLAS | 78.00 | 77.64 | 77.82 | 77.80
17
+ BLEX | 82.14 | 81.77 | 81.96 | 81.93
sysoutputs/adapt/test/pl.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/pl.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.49 | 98.77 | 98.63 |
4
+ Sentences | 82.48 | 90.43 | 86.27 |
5
+ Words | 99.40 | 99.20 | 99.30 |
6
+ UPOS | 97.46 | 97.27 | 97.37 | 98.05
7
+ XPOS | 89.05 | 88.87 | 88.96 | 89.59
8
+ UFeats | 82.76 | 82.60 | 82.68 | 83.26
9
+ AllTags | 77.16 | 77.01 | 77.08 | 77.63
10
+ Lemmas | 97.32 | 97.13 | 97.23 | 97.91
11
+ UAS | 89.62 | 89.45 | 89.53 | 90.16
12
+ LAS | 85.99 | 85.82 | 85.90 | 86.51
13
+ ELAS | 56.82 | 78.31 | 65.86 | 84.26
14
+ EULAS | 61.13 | 84.24 | 70.85 | 90.64
15
+ CLAS | 84.33 | 84.35 | 84.34 | 85.00
16
+ MLAS | 64.32 | 64.34 | 64.33 | 64.83
17
+ BLEX | 82.13 | 82.15 | 82.14 | 82.78
sysoutputs/adapt/test/ru.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7594e6c4efee0f9b3c6d9562522986e24448c8c336f44b157315ba602ed0d5f
3
+ size 11455172
sysoutputs/adapt/test/ru.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.55 | 99.64 | 99.60 |
4
+ Sentences | 98.77 | 98.83 | 98.80 |
5
+ Words | 99.55 | 99.64 | 99.60 |
6
+ UPOS | 98.72 | 98.81 | 98.76 | 99.16
7
+ XPOS | 99.55 | 99.64 | 99.60 | 100.00
8
+ UFeats | 97.37 | 97.46 | 97.42 | 97.81
9
+ AllTags | 97.15 | 97.23 | 97.19 | 97.59
10
+ Lemmas | 98.18 | 98.27 | 98.23 | 98.62
11
+ UAS | 93.46 | 93.55 | 93.50 | 93.88
12
+ LAS | 92.12 | 92.20 | 92.16 | 92.54
13
+ ELAS | 65.00 | 89.39 | 75.27 | 94.94
14
+ EULAS | 65.65 | 90.28 | 76.02 | 95.88
15
+ CLAS | 90.84 | 90.86 | 90.85 | 91.26
16
+ MLAS | 88.14 | 88.16 | 88.15 | 88.54
17
+ BLEX | 89.25 | 89.27 | 89.26 | 89.66
sysoutputs/adapt/test/sk.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/sk.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 85.65 | 84.92 | 85.28 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 97.37 | 97.37 | 97.37 | 97.37
7
+ XPOS | 88.06 | 88.06 | 88.06 | 88.06
8
+ UFeats | 92.81 | 92.81 | 92.81 | 92.81
9
+ AllTags | 87.07 | 87.07 | 87.07 | 87.07
10
+ Lemmas | 96.88 | 96.88 | 96.88 | 96.88
11
+ UAS | 90.90 | 90.90 | 90.90 | 90.90
12
+ LAS | 88.84 | 88.84 | 88.84 | 88.84
13
+ ELAS | 58.96 | 81.51 | 68.43 | 86.89
14
+ EULAS | 61.20 | 84.61 | 71.02 | 90.19
15
+ CLAS | 87.28 | 87.38 | 87.33 | 87.38
16
+ MLAS | 77.85 | 77.94 | 77.89 | 77.94
17
+ BLEX | 83.79 | 83.88 | 83.84 | 83.88
sysoutputs/adapt/test/sv.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/sv.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.36 | 99.00 | 99.18 |
4
+ Sentences | 92.74 | 94.37 | 93.54 |
5
+ Words | 99.36 | 99.00 | 99.18 |
6
+ UPOS | 97.27 | 96.91 | 97.09 | 97.90
7
+ XPOS | 95.95 | 95.60 | 95.77 | 96.57
8
+ UFeats | 89.20 | 88.87 | 89.03 | 89.77
9
+ AllTags | 88.14 | 87.81 | 87.97 | 88.70
10
+ Lemmas | 93.48 | 93.14 | 93.31 | 94.08
11
+ UAS | 88.27 | 87.95 | 88.11 | 88.84
12
+ LAS | 85.47 | 85.16 | 85.32 | 86.02
13
+ ELAS | 58.03 | 83.24 | 68.39 | 89.07
14
+ EULAS | 58.91 | 84.50 | 69.42 | 90.41
15
+ CLAS | 83.71 | 83.73 | 83.72 | 84.38
16
+ MLAS | 69.87 | 69.88 | 69.88 | 70.43
17
+ BLEX | 77.71 | 77.72 | 77.72 | 78.33
sysoutputs/adapt/test/ta.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/ta.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.88 | 99.44 | 99.16 |
4
+ Sentences | 96.72 | 98.33 | 97.52 |
5
+ Words | 95.11 | 93.92 | 94.51 |
6
+ UPOS | 83.20 | 82.15 | 82.67 | 87.47
7
+ XPOS | 73.52 | 72.60 | 73.06 | 77.30
8
+ UFeats | 77.85 | 76.87 | 77.36 | 81.85
9
+ AllTags | 72.51 | 71.59 | 72.05 | 76.23
10
+ Lemmas | 88.80 | 87.68 | 88.24 | 93.36
11
+ UAS | 67.52 | 66.67 | 67.09 | 70.99
12
+ LAS | 59.62 | 58.87 | 59.25 | 62.69
13
+ ELAS | 41.79 | 57.69 | 48.47 | 62.85
14
+ EULAS | 43.22 | 59.66 | 50.12 | 64.99
15
+ CLAS | 55.09 | 55.13 | 55.11 | 59.01
16
+ MLAS | 44.23 | 44.26 | 44.25 | 47.38
17
+ BLEX | 51.83 | 51.87 | 51.85 | 55.52
sysoutputs/adapt/test/uk.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/adapt/test/uk.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.84 | 99.85 | 99.85 |
4
+ Sentences | 95.91 | 97.31 | 96.61 |
5
+ Words | 99.80 | 99.82 | 99.81 |
6
+ UPOS | 97.86 | 97.87 | 97.86 | 98.05
7
+ XPOS | 94.23 | 94.25 | 94.24 | 94.42
8
+ UFeats | 94.19 | 94.20 | 94.20 | 94.38
9
+ AllTags | 93.25 | 93.26 | 93.26 | 93.43
10
+ Lemmas | 97.63 | 97.65 | 97.64 | 97.82
11
+ UAS | 89.50 | 89.51 | 89.50 | 89.68
12
+ LAS | 87.07 | 87.09 | 87.08 | 87.25
13
+ ELAS | 56.58 | 80.44 | 66.43 | 87.26
14
+ EULAS | 57.54 | 81.80 | 67.55 | 88.73
15
+ CLAS | 84.45 | 84.07 | 84.26 | 84.23
16
+ MLAS | 78.00 | 77.64 | 77.82 | 77.80
17
+ BLEX | 82.14 | 81.77 | 81.96 | 81.93
sysoutputs/baseline1/test/ar.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/baseline1/test/ar.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 100.00 | 100.00 | 100.00 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 100.00 | 100.00 | 100.00 | 100.00
7
+ XPOS | 100.00 | 100.00 | 100.00 | 100.00
8
+ UFeats | 100.00 | 100.00 | 100.00 | 100.00
9
+ AllTags | 100.00 | 100.00 | 100.00 | 100.00
10
+ Lemmas | 100.00 | 100.00 | 100.00 | 100.00
11
+ UAS | 100.00 | 100.00 | 100.00 | 100.00
12
+ LAS | 100.00 | 100.00 | 100.00 | 100.00
13
+ ELAS | 69.45 | 65.37 | 67.35 | 69.45
14
+ EULAS | 99.13 | 93.30 | 96.12 | 99.13
15
+ CLAS | 100.00 | 100.00 | 100.00 | 100.00
16
+ MLAS | 100.00 | 100.00 | 100.00 | 100.00
17
+ BLEX | 100.00 | 100.00 | 100.00 | 100.00
sysoutputs/baseline1/test/bg.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/baseline1/test/bg.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 100.00 | 100.00 | 100.00 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 100.00 | 100.00 | 100.00 | 100.00
7
+ XPOS | 100.00 | 100.00 | 100.00 | 100.00
8
+ UFeats | 100.00 | 100.00 | 100.00 | 100.00
9
+ AllTags | 100.00 | 100.00 | 100.00 | 100.00
10
+ Lemmas | 100.00 | 100.00 | 100.00 | 100.00
11
+ UAS | 100.00 | 100.00 | 100.00 | 100.00
12
+ LAS | 100.00 | 100.00 | 100.00 | 100.00
13
+ ELAS | 87.05 | 84.63 | 85.82 | 87.05
14
+ EULAS | 99.03 | 96.28 | 97.64 | 99.03
15
+ CLAS | 100.00 | 100.00 | 100.00 | 100.00
16
+ MLAS | 100.00 | 100.00 | 100.00 | 100.00
17
+ BLEX | 100.00 | 100.00 | 100.00 | 100.00
sysoutputs/baseline1/test/cs.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cab977995b12d9e6b6c028cd29d9e17924946179f92bf239aa4ef5858cbedda3
3
+ size 24636791
sysoutputs/baseline1/test/cs.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 100.00 | 100.00 | 100.00 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 100.00 | 100.00 | 100.00 | 100.00
7
+ XPOS | 100.00 | 100.00 | 100.00 | 100.00
8
+ UFeats | 100.00 | 100.00 | 100.00 | 100.00
9
+ AllTags | 100.00 | 100.00 | 100.00 | 100.00
10
+ Lemmas | 100.00 | 100.00 | 100.00 | 100.00
11
+ UAS | 100.00 | 100.00 | 100.00 | 100.00
12
+ LAS | 100.00 | 100.00 | 100.00 | 100.00
13
+ ELAS | 81.47 | 75.62 | 78.44 | 81.47
14
+ EULAS | 98.30 | 91.23 | 94.63 | 98.30
15
+ CLAS | 100.00 | 100.00 | 100.00 | 100.00
16
+ MLAS | 100.00 | 100.00 | 100.00 | 100.00
17
+ BLEX | 100.00 | 100.00 | 100.00 | 100.00