de-francophones commited on
Commit
a1f0762
1 Parent(s): 124ad77

c3586fe46e3c61e510f60d0bceb573cedc6cf25c5da56db4ebf4ed641b717dd7

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +3 -0
  2. sysoutputs/fastparse/v2/pertreebank/lv_lvtb-ud-test.eval.log +17 -0
  3. sysoutputs/fastparse/v2/pertreebank/nl_alpino-ud-test-sys.conllu +0 -0
  4. sysoutputs/fastparse/v2/pertreebank/nl_alpino-ud-test.eval.log +17 -0
  5. sysoutputs/fastparse/v2/pertreebank/nl_lassysmall-ud-test-sys.conllu +0 -0
  6. sysoutputs/fastparse/v2/pertreebank/nl_lassysmall-ud-test.eval.log +17 -0
  7. sysoutputs/fastparse/v2/pertreebank/pl_lfg-ud-test-sys.conllu +0 -0
  8. sysoutputs/fastparse/v2/pertreebank/pl_lfg-ud-test.eval.log +17 -0
  9. sysoutputs/fastparse/v2/pertreebank/pl_pdb-ud-test-sys.conllu +0 -0
  10. sysoutputs/fastparse/v2/pertreebank/pl_pdb-ud-test.eval.log +17 -0
  11. sysoutputs/fastparse/v2/pertreebank/pl_pud-ud-test-sys.conllu +0 -0
  12. sysoutputs/fastparse/v2/pertreebank/pl_pud-ud-test.eval.log +17 -0
  13. sysoutputs/fastparse/v2/pertreebank/ru_syntagrus-ud-test-sys.conllu +3 -0
  14. sysoutputs/fastparse/v2/pertreebank/ru_syntagrus-ud-test.eval.log +17 -0
  15. sysoutputs/fastparse/v2/pertreebank/sk_snk-ud-test-sys.conllu +0 -0
  16. sysoutputs/fastparse/v2/pertreebank/sk_snk-ud-test.eval.log +17 -0
  17. sysoutputs/fastparse/v2/pertreebank/sv_pud-ud-test-sys.conllu +0 -0
  18. sysoutputs/fastparse/v2/pertreebank/sv_pud-ud-test.eval.log +17 -0
  19. sysoutputs/fastparse/v2/pertreebank/sv_talbanken-ud-test-sys.conllu +0 -0
  20. sysoutputs/fastparse/v2/pertreebank/sv_talbanken-ud-test.eval.log +17 -0
  21. sysoutputs/fastparse/v2/pertreebank/ta_ttb-ud-test-sys.conllu +0 -0
  22. sysoutputs/fastparse/v2/pertreebank/ta_ttb-ud-test.eval.log +17 -0
  23. sysoutputs/fastparse/v2/pertreebank/uk_iu-ud-test-sys.conllu +0 -0
  24. sysoutputs/fastparse/v2/pertreebank/uk_iu-ud-test.eval.log +17 -0
  25. sysoutputs/fastparse/v2/pl.conllu +0 -0
  26. sysoutputs/fastparse/v2/pl.eval.log +17 -0
  27. sysoutputs/fastparse/v2/ru.conllu +3 -0
  28. sysoutputs/fastparse/v2/ru.eval.log +17 -0
  29. sysoutputs/fastparse/v2/sk.conllu +0 -0
  30. sysoutputs/fastparse/v2/sk.eval.log +17 -0
  31. sysoutputs/fastparse/v2/sv.conllu +0 -0
  32. sysoutputs/fastparse/v2/sv.eval.log +17 -0
  33. sysoutputs/fastparse/v2/ta.conllu +0 -0
  34. sysoutputs/fastparse/v2/ta.eval.log +17 -0
  35. sysoutputs/fastparse/v2/uk.conllu +0 -0
  36. sysoutputs/fastparse/v2/uk.eval.log +17 -0
  37. sysoutputs/koebsala/udpipe_test_20200425_134739/ar.conllu +0 -0
  38. sysoutputs/koebsala/udpipe_test_20200425_134739/ar.eval.log +17 -0
  39. sysoutputs/koebsala/udpipe_test_20200425_134739/bg.conllu +0 -0
  40. sysoutputs/koebsala/udpipe_test_20200425_134739/bg.eval.log +17 -0
  41. sysoutputs/koebsala/udpipe_test_20200425_134739/cs.conllu +3 -0
  42. sysoutputs/koebsala/udpipe_test_20200425_134739/cs.eval.log +17 -0
  43. sysoutputs/koebsala/udpipe_test_20200425_134739/en.conllu +0 -0
  44. sysoutputs/koebsala/udpipe_test_20200425_134739/en.eval.log +17 -0
  45. sysoutputs/koebsala/udpipe_test_20200425_134739/et.conllu +0 -0
  46. sysoutputs/koebsala/udpipe_test_20200425_134739/et.eval.log +17 -0
  47. sysoutputs/koebsala/udpipe_test_20200425_134739/fi.conllu +0 -0
  48. sysoutputs/koebsala/udpipe_test_20200425_134739/fi.eval.log +17 -0
  49. sysoutputs/koebsala/udpipe_test_20200425_134739/fr.conllu +0 -0
  50. sysoutputs/koebsala/udpipe_test_20200425_134739/fr.eval.log +17 -0
.gitattributes CHANGED
@@ -86,3 +86,6 @@ sysoutputs/emorynlp/final2/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lf
86
  sysoutputs/emorynlp/final2/ru.conllu filter=lfs diff=lfs merge=lfs -text
87
  sysoutputs/fastparse/v2/cs.conllu filter=lfs diff=lfs merge=lfs -text
88
  sysoutputs/fastparse/v2/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
 
 
 
 
86
  sysoutputs/emorynlp/final2/ru.conllu filter=lfs diff=lfs merge=lfs -text
87
  sysoutputs/fastparse/v2/cs.conllu filter=lfs diff=lfs merge=lfs -text
88
  sysoutputs/fastparse/v2/pertreebank/cs_pdt-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
89
+ sysoutputs/fastparse/v2/pertreebank/ru_syntagrus-ud-test-sys.conllu filter=lfs diff=lfs merge=lfs -text
90
+ sysoutputs/fastparse/v2/ru.conllu filter=lfs diff=lfs merge=lfs -text
91
+ sysoutputs/koebsala/udpipe_test_20200425_134739/cs.conllu filter=lfs diff=lfs merge=lfs -text
sysoutputs/fastparse/v2/pertreebank/lv_lvtb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.40 | 99.26 | 99.33 |
4
+ Sentences | 98.42 | 99.07 | 98.74 |
5
+ Words | 99.40 | 99.26 | 99.33 |
6
+ UPOS | 93.54 | 93.42 | 93.48 | 94.11
7
+ XPOS | 84.35 | 84.23 | 84.29 | 84.86
8
+ UFeats | 89.61 | 89.49 | 89.55 | 90.16
9
+ AllTags | 83.98 | 83.87 | 83.93 | 84.49
10
+ Lemmas | 92.80 | 92.67 | 92.73 | 93.36
11
+ UAS | 84.47 | 84.35 | 84.41 | 84.98
12
+ LAS | 79.74 | 79.63 | 79.69 | 80.23
13
+ ELAS | 76.05 | 75.10 | 75.57 | 81.63
14
+ EULAS | 77.63 | 76.67 | 77.15 | 83.33
15
+ CLAS | 76.15 | 76.01 | 76.08 | 76.81
16
+ MLAS | 63.68 | 63.57 | 63.63 | 64.24
17
+ BLEX | 70.05 | 69.92 | 69.98 | 70.66
sysoutputs/fastparse/v2/pertreebank/nl_alpino-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/nl_alpino-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.86 | 99.81 | 99.83 |
4
+ Sentences | 89.97 | 87.25 | 88.59 |
5
+ Words | 99.86 | 99.81 | 99.83 |
6
+ UPOS | 94.13 | 94.09 | 94.11 | 94.27
7
+ XPOS | 91.49 | 91.44 | 91.47 | 91.62
8
+ UFeats | 93.33 | 93.29 | 93.31 | 93.47
9
+ AllTags | 90.72 | 90.68 | 90.70 | 90.85
10
+ Lemmas | 95.13 | 95.08 | 95.11 | 95.27
11
+ UAS | 87.03 | 86.99 | 87.01 | 87.16
12
+ LAS | 82.84 | 82.80 | 82.82 | 82.96
13
+ ELAS | 81.06 | 80.67 | 80.86 | 84.51
14
+ EULAS | 81.77 | 81.38 | 81.57 | 85.25
15
+ CLAS | 76.67 | 75.70 | 76.18 | 75.86
16
+ MLAS | 66.83 | 65.98 | 66.40 | 66.12
17
+ BLEX | 71.21 | 70.31 | 70.76 | 70.46
sysoutputs/fastparse/v2/pertreebank/nl_lassysmall-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/nl_lassysmall-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.66 | 99.65 | 99.65 |
4
+ Sentences | 67.53 | 44.46 | 53.62 |
5
+ Words | 99.66 | 99.65 | 99.65 |
6
+ UPOS | 91.17 | 91.16 | 91.16 | 91.48
7
+ XPOS | 88.38 | 88.36 | 88.37 | 88.68
8
+ UFeats | 90.78 | 90.77 | 90.77 | 91.09
9
+ AllTags | 87.38 | 87.37 | 87.37 | 87.68
10
+ Lemmas | 93.75 | 93.73 | 93.74 | 94.07
11
+ UAS | 79.89 | 79.88 | 79.89 | 80.17
12
+ LAS | 76.21 | 76.19 | 76.20 | 76.47
13
+ ELAS | 73.98 | 74.12 | 74.05 | 77.82
14
+ EULAS | 74.51 | 74.65 | 74.58 | 78.38
15
+ CLAS | 68.35 | 67.65 | 68.00 | 67.88
16
+ MLAS | 58.07 | 57.47 | 57.77 | 57.67
17
+ BLEX | 62.76 | 62.11 | 62.43 | 62.32
sysoutputs/fastparse/v2/pertreebank/pl_lfg-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/pl_lfg-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.40 | 96.94 | 97.66 |
4
+ Sentences | 98.83 | 97.68 | 98.25 |
5
+ Words | 99.83 | 99.91 | 99.87 |
6
+ UPOS | 96.31 | 96.38 | 96.35 | 96.47
7
+ XPOS | 77.81 | 77.87 | 77.84 | 77.94
8
+ UFeats | 70.05 | 70.10 | 70.08 | 70.17
9
+ AllTags | 58.89 | 58.94 | 58.92 | 58.99
10
+ Lemmas | 95.73 | 95.81 | 95.77 | 95.89
11
+ UAS | 92.33 | 92.40 | 92.36 | 92.48
12
+ LAS | 87.36 | 87.42 | 87.39 | 87.50
13
+ ELAS | 74.12 | 75.23 | 74.67 | 77.40
14
+ EULAS | 84.88 | 86.14 | 85.50 | 88.63
15
+ CLAS | 84.02 | 84.10 | 84.06 | 84.21
16
+ MLAS | 46.82 | 46.86 | 46.84 | 46.92
17
+ BLEX | 79.28 | 79.36 | 79.32 | 79.45
sysoutputs/fastparse/v2/pertreebank/pl_pdb-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/pl_pdb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.82 | 99.90 | 99.86 |
4
+ Sentences | 97.82 | 97.20 | 97.51 |
5
+ Words | 99.81 | 99.88 | 99.84 |
6
+ UPOS | 97.04 | 97.11 | 97.07 | 97.22
7
+ XPOS | 88.07 | 88.14 | 88.11 | 88.24
8
+ UFeats | 88.34 | 88.40 | 88.37 | 88.51
9
+ AllTags | 87.26 | 87.33 | 87.30 | 87.43
10
+ Lemmas | 95.90 | 95.97 | 95.94 | 96.09
11
+ UAS | 87.42 | 87.49 | 87.46 | 87.59
12
+ LAS | 82.41 | 82.47 | 82.44 | 82.57
13
+ ELAS | 74.92 | 74.78 | 74.85 | 80.75
14
+ EULAS | 80.55 | 80.40 | 80.48 | 86.82
15
+ CLAS | 78.53 | 78.26 | 78.40 | 78.38
16
+ MLAS | 65.28 | 65.06 | 65.17 | 65.15
17
+ BLEX | 74.52 | 74.26 | 74.39 | 74.37
sysoutputs/fastparse/v2/pertreebank/pl_pud-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/pl_pud-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.71 | 99.87 | 99.79 |
4
+ Sentences | 95.57 | 97.00 | 96.28 |
5
+ Words | 99.67 | 99.86 | 99.77 |
6
+ UPOS | 95.22 | 95.40 | 95.31 | 95.53
7
+ XPOS | 83.87 | 84.03 | 83.95 | 84.15
8
+ UFeats | 84.50 | 84.65 | 84.58 | 84.77
9
+ AllTags | 82.87 | 83.03 | 82.95 | 83.15
10
+ Lemmas | 94.78 | 94.95 | 94.87 | 95.09
11
+ UAS | 87.07 | 87.23 | 87.15 | 87.35
12
+ LAS | 82.15 | 82.30 | 82.23 | 82.42
13
+ ELAS | 73.79 | 73.96 | 73.87 | 79.59
14
+ EULAS | 80.15 | 80.33 | 80.24 | 86.45
15
+ CLAS | 78.16 | 77.86 | 78.01 | 78.02
16
+ MLAS | 60.24 | 60.00 | 60.12 | 60.13
17
+ BLEX | 73.15 | 72.87 | 73.01 | 73.02
sysoutputs/fastparse/v2/pertreebank/ru_syntagrus-ud-test-sys.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90167b135a77f64e0153bd3bfa058a33218418a74fc32653d5e4803b5145483b
3
+ size 10975966
sysoutputs/fastparse/v2/pertreebank/ru_syntagrus-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.55 | 99.64 | 99.60 |
4
+ Sentences | 98.77 | 98.83 | 98.80 |
5
+ Words | 99.55 | 99.64 | 99.60 |
6
+ UPOS | 97.73 | 97.82 | 97.78 | 98.17
7
+ XPOS | 99.55 | 99.64 | 99.60 | 100.00
8
+ UFeats | 85.30 | 85.38 | 85.34 | 85.68
9
+ AllTags | 84.93 | 85.00 | 84.97 | 85.31
10
+ Lemmas | 96.51 | 96.59 | 96.55 | 96.94
11
+ UAS | 86.77 | 86.84 | 86.81 | 87.16
12
+ LAS | 83.20 | 83.27 | 83.24 | 83.57
13
+ ELAS | 80.42 | 80.27 | 80.35 | 85.25
14
+ EULAS | 81.75 | 81.60 | 81.68 | 86.66
15
+ CLAS | 80.11 | 79.90 | 80.01 | 80.25
16
+ MLAS | 62.90 | 62.73 | 62.81 | 63.01
17
+ BLEX | 76.83 | 76.63 | 76.73 | 76.96
sysoutputs/fastparse/v2/pertreebank/sk_snk-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/sk_snk-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 85.65 | 84.92 | 85.28 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 92.93 | 92.93 | 92.93 | 92.93
7
+ XPOS | 77.06 | 77.06 | 77.06 | 77.06
8
+ UFeats | 80.34 | 80.34 | 80.34 | 80.34
9
+ AllTags | 76.71 | 76.71 | 76.71 | 76.71
10
+ Lemmas | 86.56 | 86.56 | 86.56 | 86.56
11
+ UAS | 83.24 | 83.24 | 83.24 | 83.24
12
+ LAS | 78.26 | 78.26 | 78.26 | 78.26
13
+ ELAS | 73.49 | 73.44 | 73.46 | 78.29
14
+ EULAS | 76.50 | 76.45 | 76.47 | 81.49
15
+ CLAS | 73.72 | 74.02 | 73.87 | 74.02
16
+ MLAS | 53.72 | 53.94 | 53.83 | 53.94
17
+ BLEX | 61.43 | 61.68 | 61.56 | 61.68
sysoutputs/fastparse/v2/pertreebank/sv_pud-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/sv_pud-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.81 | 98.01 | 98.41 |
4
+ Sentences | 88.44 | 92.60 | 90.47 |
5
+ Words | 98.81 | 98.01 | 98.41 |
6
+ UPOS | 91.31 | 90.57 | 90.94 | 92.41
7
+ XPOS | 88.16 | 87.44 | 87.80 | 89.22
8
+ UFeats | 74.89 | 74.28 | 74.58 | 75.79
9
+ AllTags | 73.03 | 72.44 | 72.73 | 73.91
10
+ Lemmas | 84.53 | 83.84 | 84.19 | 85.55
11
+ UAS | 78.82 | 78.18 | 78.50 | 79.77
12
+ LAS | 73.68 | 73.08 | 73.38 | 74.57
13
+ ELAS | 71.02 | 70.81 | 70.91 | 76.21
14
+ EULAS | 72.01 | 71.80 | 71.90 | 77.27
15
+ CLAS | 69.18 | 69.33 | 69.26 | 70.37
16
+ MLAS | 42.37 | 42.46 | 42.41 | 43.10
17
+ BLEX | 56.35 | 56.47 | 56.41 | 57.32
sysoutputs/fastparse/v2/pertreebank/sv_talbanken-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/sv_talbanken-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.87 | 99.92 | 99.89 |
4
+ Sentences | 96.45 | 95.82 | 96.13 |
5
+ Words | 99.87 | 99.92 | 99.89 |
6
+ UPOS | 95.57 | 95.62 | 95.59 | 95.69
7
+ XPOS | 93.87 | 93.92 | 93.90 | 94.00
8
+ UFeats | 94.44 | 94.49 | 94.46 | 94.56
9
+ AllTags | 92.84 | 92.89 | 92.87 | 92.97
10
+ Lemmas | 95.36 | 95.41 | 95.38 | 95.48
11
+ UAS | 85.66 | 85.70 | 85.68 | 85.77
12
+ LAS | 81.53 | 81.58 | 81.56 | 81.64
13
+ ELAS | 79.21 | 79.27 | 79.24 | 84.36
14
+ EULAS | 80.07 | 80.13 | 80.10 | 85.27
15
+ CLAS | 78.14 | 77.62 | 77.88 | 77.72
16
+ MLAS | 70.56 | 70.10 | 70.33 | 70.19
17
+ BLEX | 73.08 | 72.60 | 72.84 | 72.70
sysoutputs/fastparse/v2/pertreebank/ta_ttb-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/ta_ttb-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.88 | 99.44 | 99.16 |
4
+ Sentences | 96.72 | 98.33 | 97.52 |
5
+ Words | 95.11 | 93.92 | 94.51 |
6
+ UPOS | 81.82 | 80.79 | 81.31 | 86.03
7
+ XPOS | 76.83 | 75.87 | 76.35 | 80.78
8
+ UFeats | 80.96 | 79.94 | 80.45 | 85.12
9
+ AllTags | 76.07 | 75.11 | 75.59 | 79.98
10
+ Lemmas | 84.67 | 83.61 | 84.14 | 89.03
11
+ UAS | 60.03 | 59.28 | 59.65 | 63.12
12
+ LAS | 49.13 | 48.52 | 48.82 | 51.66
13
+ ELAS | 47.00 | 47.05 | 47.02 | 51.18
14
+ EULAS | 48.08 | 48.13 | 48.11 | 52.36
15
+ CLAS | 43.48 | 43.72 | 43.60 | 46.80
16
+ MLAS | 35.38 | 35.57 | 35.48 | 38.08
17
+ BLEX | 39.50 | 39.71 | 39.61 | 42.51
sysoutputs/fastparse/v2/pertreebank/uk_iu-ud-test-sys.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pertreebank/uk_iu-ud-test.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.84 | 99.85 | 99.85 |
4
+ Sentences | 95.91 | 97.31 | 96.61 |
5
+ Words | 99.80 | 99.82 | 99.81 |
6
+ UPOS | 94.91 | 94.92 | 94.91 | 95.10
7
+ XPOS | 84.02 | 84.03 | 84.03 | 84.18
8
+ UFeats | 84.27 | 84.29 | 84.28 | 84.44
9
+ AllTags | 83.32 | 83.33 | 83.32 | 83.48
10
+ Lemmas | 93.55 | 93.57 | 93.56 | 93.74
11
+ UAS | 83.39 | 83.40 | 83.39 | 83.55
12
+ LAS | 78.73 | 78.74 | 78.73 | 78.88
13
+ ELAS | 74.61 | 73.44 | 74.02 | 79.67
14
+ EULAS | 76.83 | 75.63 | 76.23 | 82.04
15
+ CLAS | 73.67 | 73.45 | 73.56 | 73.60
16
+ MLAS | 57.83 | 57.67 | 57.75 | 57.78
17
+ BLEX | 67.49 | 67.29 | 67.39 | 67.42
sysoutputs/fastparse/v2/pl.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/pl.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.51 | 99.29 | 99.40 |
4
+ Sentences | 97.70 | 97.33 | 97.52 |
5
+ Words | 99.77 | 99.88 | 99.83 |
6
+ UPOS | 96.38 | 96.48 | 96.43 | 96.60
7
+ XPOS | 84.82 | 84.91 | 84.87 | 85.01
8
+ UFeats | 83.57 | 83.66 | 83.62 | 83.76
9
+ AllTags | 80.31 | 80.40 | 80.35 | 80.49
10
+ Lemmas | 95.55 | 95.65 | 95.60 | 95.77
11
+ UAS | 88.31 | 88.40 | 88.36 | 88.51
12
+ LAS | 83.33 | 83.42 | 83.38 | 83.52
13
+ ELAS | 74.44 | 74.63 | 74.54 | 79.75
14
+ EULAS | 81.29 | 81.49 | 81.39 | 87.08
15
+ CLAS | 79.52 | 79.30 | 79.41 | 79.43
16
+ MLAS | 60.16 | 60.00 | 60.08 | 60.09
17
+ BLEX | 75.07 | 74.87 | 74.97 | 74.99
sysoutputs/fastparse/v2/ru.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90167b135a77f64e0153bd3bfa058a33218418a74fc32653d5e4803b5145483b
3
+ size 10975966
sysoutputs/fastparse/v2/ru.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.55 | 99.64 | 99.60 |
4
+ Sentences | 98.77 | 98.83 | 98.80 |
5
+ Words | 99.55 | 99.64 | 99.60 |
6
+ UPOS | 97.73 | 97.82 | 97.78 | 98.17
7
+ XPOS | 99.55 | 99.64 | 99.60 | 100.00
8
+ UFeats | 85.30 | 85.38 | 85.34 | 85.68
9
+ AllTags | 84.93 | 85.00 | 84.97 | 85.31
10
+ Lemmas | 96.51 | 96.59 | 96.55 | 96.94
11
+ UAS | 86.77 | 86.84 | 86.81 | 87.16
12
+ LAS | 83.20 | 83.27 | 83.24 | 83.57
13
+ ELAS | 80.42 | 80.27 | 80.35 | 85.25
14
+ EULAS | 81.75 | 81.60 | 81.68 | 86.66
15
+ CLAS | 80.11 | 79.90 | 80.01 | 80.25
16
+ MLAS | 62.90 | 62.73 | 62.81 | 63.01
17
+ BLEX | 76.83 | 76.63 | 76.73 | 76.96
sysoutputs/fastparse/v2/sk.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/sk.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 100.00 | 100.00 | 100.00 |
4
+ Sentences | 85.65 | 84.92 | 85.28 |
5
+ Words | 100.00 | 100.00 | 100.00 |
6
+ UPOS | 92.93 | 92.93 | 92.93 | 92.93
7
+ XPOS | 77.06 | 77.06 | 77.06 | 77.06
8
+ UFeats | 80.34 | 80.34 | 80.34 | 80.34
9
+ AllTags | 76.71 | 76.71 | 76.71 | 76.71
10
+ Lemmas | 86.56 | 86.56 | 86.56 | 86.56
11
+ UAS | 83.24 | 83.24 | 83.24 | 83.24
12
+ LAS | 78.26 | 78.26 | 78.26 | 78.26
13
+ ELAS | 73.49 | 73.44 | 73.46 | 78.29
14
+ EULAS | 76.50 | 76.45 | 76.47 | 81.49
15
+ CLAS | 73.72 | 74.02 | 73.87 | 74.02
16
+ MLAS | 53.72 | 53.94 | 53.83 | 53.94
17
+ BLEX | 61.43 | 61.68 | 61.56 | 61.68
sysoutputs/fastparse/v2/sv.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/sv.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.36 | 99.00 | 99.18 |
4
+ Sentences | 92.74 | 94.37 | 93.54 |
5
+ Words | 99.36 | 99.00 | 99.18 |
6
+ UPOS | 93.52 | 93.18 | 93.35 | 94.12
7
+ XPOS | 91.12 | 90.79 | 90.96 | 91.71
8
+ UFeats | 85.03 | 84.72 | 84.87 | 85.58
9
+ AllTags | 83.31 | 83.00 | 83.15 | 83.84
10
+ Lemmas | 90.14 | 89.82 | 89.98 | 90.73
11
+ UAS | 82.37 | 82.07 | 82.22 | 82.90
12
+ LAS | 77.75 | 77.47 | 77.61 | 78.25
13
+ ELAS | 75.27 | 75.20 | 75.24 | 80.46
14
+ EULAS | 76.20 | 76.12 | 76.16 | 81.44
15
+ CLAS | 73.80 | 73.63 | 73.71 | 74.20
16
+ MLAS | 56.90 | 56.77 | 56.84 | 57.22
17
+ BLEX | 64.98 | 64.82 | 64.90 | 65.33
sysoutputs/fastparse/v2/ta.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/ta.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 98.88 | 99.44 | 99.16 |
4
+ Sentences | 96.72 | 98.33 | 97.52 |
5
+ Words | 95.11 | 93.92 | 94.51 |
6
+ UPOS | 81.82 | 80.79 | 81.31 | 86.03
7
+ XPOS | 76.83 | 75.87 | 76.35 | 80.78
8
+ UFeats | 80.96 | 79.94 | 80.45 | 85.12
9
+ AllTags | 76.07 | 75.11 | 75.59 | 79.98
10
+ Lemmas | 84.67 | 83.61 | 84.14 | 89.03
11
+ UAS | 60.03 | 59.28 | 59.65 | 63.12
12
+ LAS | 49.13 | 48.52 | 48.82 | 51.66
13
+ ELAS | 47.00 | 46.98 | 46.99 | 51.18
14
+ EULAS | 48.08 | 48.06 | 48.07 | 52.36
15
+ CLAS | 43.48 | 43.72 | 43.60 | 46.80
16
+ MLAS | 35.38 | 35.57 | 35.48 | 38.08
17
+ BLEX | 39.50 | 39.71 | 39.61 | 42.51
sysoutputs/fastparse/v2/uk.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/fastparse/v2/uk.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.84 | 99.85 | 99.85 |
4
+ Sentences | 95.91 | 97.31 | 96.61 |
5
+ Words | 99.80 | 99.82 | 99.81 |
6
+ UPOS | 94.91 | 94.92 | 94.91 | 95.10
7
+ XPOS | 84.02 | 84.03 | 84.03 | 84.18
8
+ UFeats | 84.27 | 84.29 | 84.28 | 84.44
9
+ AllTags | 83.32 | 83.33 | 83.32 | 83.48
10
+ Lemmas | 93.55 | 93.57 | 93.56 | 93.74
11
+ UAS | 83.39 | 83.40 | 83.39 | 83.55
12
+ LAS | 78.73 | 78.74 | 78.73 | 78.88
13
+ ELAS | 74.61 | 73.44 | 74.02 | 79.67
14
+ EULAS | 76.83 | 75.63 | 76.23 | 82.04
15
+ CLAS | 73.67 | 73.45 | 73.56 | 73.60
16
+ MLAS | 57.83 | 57.67 | 57.75 | 57.78
17
+ BLEX | 67.49 | 67.29 | 67.39 | 67.42
sysoutputs/koebsala/udpipe_test_20200425_134739/ar.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/koebsala/udpipe_test_20200425_134739/ar.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.98 | 99.98 | 99.98 |
4
+ Sentences | 76.20 | 88.97 | 82.09 |
5
+ Words | 95.76 | 93.42 | 94.58 |
6
+ UPOS | 89.64 | 87.44 | 88.53 | 93.60
7
+ XPOS | 85.05 | 82.97 | 84.00 | 88.81
8
+ UFeats | 85.21 | 83.13 | 84.16 | 88.98
9
+ AllTags | 83.00 | 80.97 | 81.97 | 86.67
10
+ Lemmas | 89.57 | 87.37 | 88.46 | 93.53
11
+ UAS | 73.63 | 71.82 | 72.71 | 76.88
12
+ LAS | 68.62 | 66.94 | 67.77 | 71.66
13
+ ELAS | 55.30 | 67.61 | 60.84 | 76.89
14
+ EULAS | 56.78 | 69.41 | 62.46 | 78.94
15
+ CLAS | 62.94 | 63.06 | 63.00 | 66.83
16
+ MLAS | 54.55 | 54.66 | 54.60 | 57.92
17
+ BLEX | 58.65 | 58.76 | 58.70 | 62.27
sysoutputs/koebsala/udpipe_test_20200425_134739/bg.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/koebsala/udpipe_test_20200425_134739/bg.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.94 | 99.88 | 99.91 |
4
+ Sentences | 94.33 | 94.00 | 94.17 |
5
+ Words | 99.94 | 99.88 | 99.91 |
6
+ UPOS | 97.65 | 97.59 | 97.62 | 97.71
7
+ XPOS | 94.37 | 94.31 | 94.34 | 94.43
8
+ UFeats | 95.42 | 95.36 | 95.39 | 95.48
9
+ AllTags | 93.80 | 93.75 | 93.78 | 93.86
10
+ Lemmas | 94.63 | 94.58 | 94.60 | 94.69
11
+ UAS | 89.13 | 89.07 | 89.10 | 89.18
12
+ LAS | 85.09 | 85.04 | 85.07 | 85.14
13
+ ELAS | 58.13 | 84.52 | 68.88 | 87.04
14
+ EULAS | 59.13 | 85.98 | 70.07 | 88.54
15
+ CLAS | 80.50 | 80.28 | 80.39 | 80.34
16
+ MLAS | 75.46 | 75.26 | 75.36 | 75.31
17
+ BLEX | 74.11 | 73.91 | 74.01 | 73.97
sysoutputs/koebsala/udpipe_test_20200425_134739/cs.conllu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7658779fba6cc1cbe4155b81b94ea090bd8697009dfcd8bd0852e8ca8a437719
3
+ size 24354175
sysoutputs/koebsala/udpipe_test_20200425_134739/cs.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.90 | 99.86 | 99.88 |
4
+ Sentences | 94.48 | 91.92 | 93.18 |
5
+ Words | 99.90 | 99.86 | 99.88 |
6
+ UPOS | 97.85 | 97.81 | 97.83 | 97.95
7
+ XPOS | 90.89 | 90.86 | 90.88 | 90.99
8
+ UFeats | 90.79 | 90.75 | 90.77 | 90.88
9
+ AllTags | 89.73 | 89.70 | 89.71 | 89.82
10
+ Lemmas | 97.44 | 97.40 | 97.42 | 97.54
11
+ UAS | 86.67 | 86.63 | 86.65 | 86.76
12
+ LAS | 82.71 | 82.68 | 82.70 | 82.80
13
+ ELAS | 52.26 | 73.58 | 61.11 | 79.39
14
+ EULAS | 55.25 | 77.80 | 64.61 | 83.94
15
+ CLAS | 79.70 | 79.61 | 79.66 | 79.73
16
+ MLAS | 71.00 | 70.92 | 70.96 | 71.02
17
+ BLEX | 77.49 | 77.40 | 77.45 | 77.52
sysoutputs/koebsala/udpipe_test_20200425_134739/en.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/koebsala/udpipe_test_20200425_134739/en.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.23 | 99.20 | 99.22 |
4
+ Sentences | 88.00 | 80.11 | 83.87 |
5
+ Words | 99.23 | 99.20 | 99.22 |
6
+ UPOS | 93.64 | 93.62 | 93.63 | 94.37
7
+ XPOS | 92.79 | 92.76 | 92.77 | 93.51
8
+ UFeats | 94.10 | 94.08 | 94.09 | 94.83
9
+ AllTags | 90.71 | 90.68 | 90.70 | 91.41
10
+ Lemmas | 95.43 | 95.40 | 95.41 | 96.17
11
+ UAS | 81.34 | 81.32 | 81.33 | 81.97
12
+ LAS | 77.94 | 77.92 | 77.93 | 78.54
13
+ ELAS | 54.81 | 80.97 | 65.37 | 85.21
14
+ EULAS | 55.83 | 82.48 | 66.58 | 86.80
15
+ CLAS | 73.77 | 73.31 | 73.54 | 74.01
16
+ MLAS | 67.58 | 67.16 | 67.37 | 67.80
17
+ BLEX | 70.13 | 69.70 | 69.91 | 70.36
sysoutputs/koebsala/udpipe_test_20200425_134739/et.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/koebsala/udpipe_test_20200425_134739/et.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.61 | 99.80 | 99.70 |
4
+ Sentences | 92.14 | 87.93 | 89.98 |
5
+ Words | 99.61 | 99.80 | 99.70 |
6
+ UPOS | 94.90 | 95.08 | 94.99 | 95.27
7
+ XPOS | 96.14 | 96.33 | 96.23 | 96.52
8
+ UFeats | 92.71 | 92.89 | 92.80 | 93.08
9
+ AllTags | 90.88 | 91.05 | 90.97 | 91.24
10
+ Lemmas | 90.28 | 90.45 | 90.37 | 90.63
11
+ UAS | 78.05 | 78.20 | 78.12 | 78.36
12
+ LAS | 74.06 | 74.20 | 74.13 | 74.35
13
+ ELAS | 50.02 | 72.11 | 59.07 | 73.47
14
+ EULAS | 51.25 | 73.88 | 60.52 | 75.27
15
+ CLAS | 71.89 | 71.95 | 71.92 | 72.12
16
+ MLAS | 66.33 | 66.38 | 66.36 | 66.54
17
+ BLEX | 63.21 | 63.25 | 63.23 | 63.40
sysoutputs/koebsala/udpipe_test_20200425_134739/fi.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/koebsala/udpipe_test_20200425_134739/fi.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.75 | 99.64 | 99.70 |
4
+ Sentences | 87.61 | 89.71 | 88.65 |
5
+ Words | 99.74 | 99.63 | 99.68 |
6
+ UPOS | 94.89 | 94.78 | 94.83 | 95.13
7
+ XPOS | 54.55 | 54.49 | 54.52 | 54.69
8
+ UFeats | 93.07 | 92.96 | 93.02 | 93.31
9
+ AllTags | 51.85 | 51.79 | 51.82 | 51.98
10
+ Lemmas | 87.14 | 87.03 | 87.09 | 87.36
11
+ UAS | 81.39 | 81.29 | 81.34 | 81.60
12
+ LAS | 77.95 | 77.86 | 77.90 | 78.15
13
+ ELAS | 57.32 | 82.17 | 67.54 | 87.44
14
+ EULAS | 58.47 | 83.82 | 68.89 | 89.19
15
+ CLAS | 75.40 | 75.22 | 75.31 | 75.51
16
+ MLAS | 70.21 | 70.04 | 70.13 | 70.31
17
+ BLEX | 63.85 | 63.69 | 63.77 | 63.94
sysoutputs/koebsala/udpipe_test_20200425_134739/fr.conllu ADDED
The diff for this file is too large to render. See raw diff
 
sysoutputs/koebsala/udpipe_test_20200425_134739/fr.eval.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metric | Precision | Recall | F1 Score | AligndAcc
2
+ -----------+-----------+-----------+-----------+-----------
3
+ Tokens | 99.75 | 99.81 | 99.78 |
4
+ Sentences | 96.64 | 94.32 | 95.46 |
5
+ Words | 99.20 | 99.53 | 99.36 |
6
+ UPOS | 93.78 | 94.10 | 93.94 | 94.54
7
+ XPOS | 99.20 | 99.53 | 99.36 | 100.00
8
+ UFeats | 75.89 | 76.15 | 76.02 | 76.51
9
+ AllTags | 73.16 | 73.41 | 73.29 | 73.76
10
+ Lemmas | 95.91 | 96.23 | 96.07 | 96.69
11
+ UAS | 81.68 | 81.95 | 81.82 | 82.34
12
+ LAS | 75.27 | 75.52 | 75.40 | 75.88
13
+ ELAS | 57.63 | 82.71 | 67.93 | 85.12
14
+ EULAS | 58.71 | 84.26 | 69.21 | 86.71
15
+ CLAS | 70.38 | 69.47 | 69.92 | 69.67
16
+ MLAS | 40.87 | 40.34 | 40.60 | 40.46
17
+ BLEX | 66.98 | 66.11 | 66.55 | 66.31