Datasets:
wmt
/

Modalities:
Text
Formats:
parquet
Libraries:
Datasets
Dask
License:

Convert dataset to Parquet

#8
by albertvillanova HF staff - opened
Files changed (49) hide show
  1. README.md +67 -26
  2. cs-en/test-00000-of-00001.parquet +3 -0
  3. cs-en/train-00000-of-00001.parquet +3 -0
  4. cs-en/validation-00000-of-00001.parquet +3 -0
  5. de-en/test-00000-of-00001.parquet +3 -0
  6. de-en/train-00000-of-00003.parquet +3 -0
  7. de-en/train-00001-of-00003.parquet +3 -0
  8. de-en/train-00002-of-00003.parquet +3 -0
  9. de-en/validation-00000-of-00001.parquet +3 -0
  10. fr-en/test-00000-of-00001.parquet +3 -0
  11. fr-en/train-00000-of-00030.parquet +3 -0
  12. fr-en/train-00001-of-00030.parquet +3 -0
  13. fr-en/train-00002-of-00030.parquet +3 -0
  14. fr-en/train-00003-of-00030.parquet +3 -0
  15. fr-en/train-00004-of-00030.parquet +3 -0
  16. fr-en/train-00005-of-00030.parquet +3 -0
  17. fr-en/train-00006-of-00030.parquet +3 -0
  18. fr-en/train-00007-of-00030.parquet +3 -0
  19. fr-en/train-00008-of-00030.parquet +3 -0
  20. fr-en/train-00009-of-00030.parquet +3 -0
  21. fr-en/train-00010-of-00030.parquet +3 -0
  22. fr-en/train-00011-of-00030.parquet +3 -0
  23. fr-en/train-00012-of-00030.parquet +3 -0
  24. fr-en/train-00013-of-00030.parquet +3 -0
  25. fr-en/train-00014-of-00030.parquet +3 -0
  26. fr-en/train-00015-of-00030.parquet +3 -0
  27. fr-en/train-00016-of-00030.parquet +3 -0
  28. fr-en/train-00017-of-00030.parquet +3 -0
  29. fr-en/train-00018-of-00030.parquet +3 -0
  30. fr-en/train-00019-of-00030.parquet +3 -0
  31. fr-en/train-00020-of-00030.parquet +3 -0
  32. fr-en/train-00021-of-00030.parquet +3 -0
  33. fr-en/train-00022-of-00030.parquet +3 -0
  34. fr-en/train-00023-of-00030.parquet +3 -0
  35. fr-en/train-00024-of-00030.parquet +3 -0
  36. fr-en/train-00025-of-00030.parquet +3 -0
  37. fr-en/train-00026-of-00030.parquet +3 -0
  38. fr-en/train-00027-of-00030.parquet +3 -0
  39. fr-en/train-00028-of-00030.parquet +3 -0
  40. fr-en/train-00029-of-00030.parquet +3 -0
  41. fr-en/validation-00000-of-00001.parquet +3 -0
  42. hi-en/test-00000-of-00001.parquet +3 -0
  43. hi-en/train-00000-of-00001.parquet +3 -0
  44. hi-en/validation-00000-of-00001.parquet +3 -0
  45. ru-en/test-00000-of-00001.parquet +3 -0
  46. ru-en/train-00000-of-00001.parquet +3 -0
  47. ru-en/validation-00000-of-00001.parquet +3 -0
  48. wmt14.py +0 -81
  49. wmt_utils.py +0 -1022
README.md CHANGED
@@ -25,8 +25,8 @@ source_datasets:
25
  task_categories:
26
  - translation
27
  task_ids: []
28
- pretty_name: WMT14
29
  paperswithcode_id: wmt-2014
 
30
  dataset_info:
31
  - config_name: cs-en
32
  features:
@@ -38,16 +38,16 @@ dataset_info:
38
  - en
39
  splits:
40
  - name: train
41
- num_bytes: 280992794
42
  num_examples: 953621
43
  - name: validation
44
- num_bytes: 702473
45
  num_examples: 3000
46
  - name: test
47
- num_bytes: 757817
48
  num_examples: 3003
49
- download_size: 1696003559
50
- dataset_size: 282453084
51
  - config_name: de-en
52
  features:
53
  - name: translation
@@ -58,16 +58,16 @@ dataset_info:
58
  - en
59
  splits:
60
  - name: train
61
- num_bytes: 1358410408
62
  num_examples: 4508785
63
  - name: validation
64
- num_bytes: 736415
65
  num_examples: 3000
66
  - name: test
67
- num_bytes: 777334
68
  num_examples: 3003
69
- download_size: 1696003559
70
- dataset_size: 1359924157
71
  - config_name: fr-en
72
  features:
73
  - name: translation
@@ -78,16 +78,16 @@ dataset_info:
78
  - en
79
  splits:
80
  - name: train
81
- num_bytes: 14752554924
82
  num_examples: 40836715
83
  - name: validation
84
- num_bytes: 744447
85
  num_examples: 3000
86
  - name: test
87
- num_bytes: 838857
88
  num_examples: 3003
89
- download_size: 6658118909
90
- dataset_size: 14754138228
91
  - config_name: hi-en
92
  features:
93
  - name: translation
@@ -98,16 +98,16 @@ dataset_info:
98
  - en
99
  splits:
100
  - name: train
101
- num_bytes: 1936035
102
  num_examples: 32863
103
  - name: validation
104
- num_bytes: 181465
105
  num_examples: 520
106
  - name: test
107
- num_bytes: 1075016
108
  num_examples: 2507
109
- download_size: 46879684
110
- dataset_size: 3192516
111
  - config_name: ru-en
112
  features:
113
  - name: translation
@@ -118,16 +118,57 @@ dataset_info:
118
  - en
119
  splits:
120
  - name: train
121
- num_bytes: 433210270
122
  num_examples: 1486965
123
  - name: validation
124
- num_bytes: 977946
125
  num_examples: 3000
126
  - name: test
127
- num_bytes: 1087746
128
  num_examples: 3003
129
- download_size: 1047396736
130
- dataset_size: 435275962
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  ---
132
 
133
  # Dataset Card for "wmt14"
 
25
  task_categories:
26
  - translation
27
  task_ids: []
 
28
  paperswithcode_id: wmt-2014
29
+ pretty_name: WMT14
30
  dataset_info:
31
  - config_name: cs-en
32
  features:
 
38
  - en
39
  splits:
40
  - name: train
41
+ num_bytes: 280992026
42
  num_examples: 953621
43
  - name: validation
44
+ num_bytes: 702465
45
  num_examples: 3000
46
  - name: test
47
+ num_bytes: 757809
48
  num_examples: 3003
49
+ download_size: 168878237
50
+ dataset_size: 282452300
51
  - config_name: de-en
52
  features:
53
  - name: translation
 
58
  - en
59
  splits:
60
  - name: train
61
+ num_bytes: 1358406800
62
  num_examples: 4508785
63
  - name: validation
64
+ num_bytes: 736407
65
  num_examples: 3000
66
  - name: test
67
+ num_bytes: 777326
68
  num_examples: 3003
69
+ download_size: 818467512
70
+ dataset_size: 1359920533
71
  - config_name: fr-en
72
  features:
73
  - name: translation
 
78
  - en
79
  splits:
80
  - name: train
81
+ num_bytes: 14752522252
82
  num_examples: 40836715
83
  - name: validation
84
+ num_bytes: 744439
85
  num_examples: 3000
86
  - name: test
87
+ num_bytes: 838849
88
  num_examples: 3003
89
+ download_size: 7777527744
90
+ dataset_size: 14754105540
91
  - config_name: hi-en
92
  features:
93
  - name: translation
 
98
  - en
99
  splits:
100
  - name: train
101
+ num_bytes: 1936003
102
  num_examples: 32863
103
  - name: validation
104
+ num_bytes: 181457
105
  num_examples: 520
106
  - name: test
107
+ num_bytes: 1075008
108
  num_examples: 2507
109
+ download_size: 1583004
110
+ dataset_size: 3192468
111
  - config_name: ru-en
112
  features:
113
  - name: translation
 
118
  - en
119
  splits:
120
  - name: train
121
+ num_bytes: 433209078
122
  num_examples: 1486965
123
  - name: validation
124
+ num_bytes: 977938
125
  num_examples: 3000
126
  - name: test
127
+ num_bytes: 1087738
128
  num_examples: 3003
129
+ download_size: 223537244
130
+ dataset_size: 435274754
131
+ configs:
132
+ - config_name: cs-en
133
+ data_files:
134
+ - split: train
135
+ path: cs-en/train-*
136
+ - split: validation
137
+ path: cs-en/validation-*
138
+ - split: test
139
+ path: cs-en/test-*
140
+ - config_name: de-en
141
+ data_files:
142
+ - split: train
143
+ path: de-en/train-*
144
+ - split: validation
145
+ path: de-en/validation-*
146
+ - split: test
147
+ path: de-en/test-*
148
+ - config_name: fr-en
149
+ data_files:
150
+ - split: train
151
+ path: fr-en/train-*
152
+ - split: validation
153
+ path: fr-en/validation-*
154
+ - split: test
155
+ path: fr-en/test-*
156
+ - config_name: hi-en
157
+ data_files:
158
+ - split: train
159
+ path: hi-en/train-*
160
+ - split: validation
161
+ path: hi-en/validation-*
162
+ - split: test
163
+ path: hi-en/test-*
164
+ - config_name: ru-en
165
+ data_files:
166
+ - split: train
167
+ path: ru-en/train-*
168
+ - split: validation
169
+ path: ru-en/validation-*
170
+ - split: test
171
+ path: ru-en/test-*
172
  ---
173
 
174
  # Dataset Card for "wmt14"
cs-en/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864967b2cb62dc7479aaa77d2a9acca602535af67a784b6779cea10b5cda0f72
3
+ size 506590
cs-en/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4b78556c53efa95e42cadda2202e6e7184f525062766e89e668397bcfcf2d03
3
+ size 167898713
cs-en/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a5da5893513e1018f58ec005be686640994ec26d693aa9d7186c24518f87590
3
+ size 472934
de-en/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3baab716f14f8abfa63e54995e31f035bb78b20c7943de72f3e1110b7ff35bb
3
+ size 508753
de-en/train-00000-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bbe0f0c7d53a71dc9e22b8f5406b058da3bc0f587ebd1b04b0675020ec14dcb
3
+ size 279527247
de-en/train-00001-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94d02ee1d27bd31654e28676a0fb9c0beb485180edecc4a93bb8403919de70df
3
+ size 264970514
de-en/train-00002-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5192553dd3f9ec0ccb73387784ce2a3be0c407f539d61a30abecd692433dfed0
3
+ size 272987200
de-en/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4a2a0e3b3e472f97f4eceb6d06724571f5390ac705768a02c1a9ad5c3d675ea
3
+ size 473798
fr-en/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b9f62543c776b5e992d15524f76f9fe589d838c1dcb04a8738c21e8c42269f1
3
+ size 535966
fr-en/train-00000-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09ba03062833aeaf53092cc6519d73250418c5508c0697243cb5d5b064ff1dc4
3
+ size 252421180
fr-en/train-00001-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f111d50d7b4b41a62c6a09f6a80ed726e810b357f2c7d78178cd34d86a023cb9
3
+ size 240967470
fr-en/train-00002-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3cda45ef2fa974670f7d98d7e2a46a603f6078b0347c9981dbe05f50573f9bd
3
+ size 243304904
fr-en/train-00003-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:777941659d92eacb8d04fd7aaf0fb22b3174254b9daa853ae2197cf9d4ac7524
3
+ size 247396927
fr-en/train-00004-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:072549b25e83f34d5ad0896033b5fddd83abab185a19ae9a94d0c75be958f75a
3
+ size 241928194
fr-en/train-00005-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a67ffb2a896e513996fdbc4288b8075ab6a9c44200259500d2d7bee620638b36
3
+ size 238463332
fr-en/train-00006-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:808ada8287dd08c2c3b2507b3a604098d287e56c9d0dd59141a15c9b8462d200
3
+ size 239654632
fr-en/train-00007-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81df9362da642784f22fd57a1427d86546c9b7dd1c9fbc670a659374abab6122
3
+ size 240573755
fr-en/train-00008-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ced722fbde2e60cd609fb520e36c883a0b653b02899945ce550250cf12e1ff67
3
+ size 242283578
fr-en/train-00009-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06f53a80a25cb2ab1e0a46a676e16b9eb9d35918c14b8757f26d8cd4526ead18
3
+ size 238671372
fr-en/train-00010-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b7fa7d18d71a069df7f87390a3a910d1c4ada5efebe19b626f5cd3174692589
3
+ size 239240023
fr-en/train-00011-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f47f540c00cb5d78e003b243c8a0025bdf7a2c720c0b5ee63f04d5c049c134ef
3
+ size 241314516
fr-en/train-00012-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06c75aaea487fc5a622b77fcd2374570833a4bd4c9bad605fe6fba88361ab9bb
3
+ size 241458478
fr-en/train-00013-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b35ea29ec9e70af1f7ba256be1fd4520355603f65b2d7e10c3cf1d004adb8bb
3
+ size 229680498
fr-en/train-00014-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f15766ca53465d1448e03b6a2378c6a46ab99c2c2cd0fc22bab5cbb025ea0180
3
+ size 214351652
fr-en/train-00015-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44bf4fea33c4378c4ca43536f698e1c624ac2d8a8d09d0c5ee1f8f110c31f5a4
3
+ size 230720193
fr-en/train-00016-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e4e8001eda1da3068acba4f61d9fa643bb27302efdc34ef45f143639d489003
3
+ size 227173824
fr-en/train-00017-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73579c11775d6b8a2e10178f8ceb6df4a9a40bb9631275219c7f46e57026035b
3
+ size 226218538
fr-en/train-00018-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864f70a4fa132d4a49f0878371aa943429dc379e10522696305b14200ecf882a
3
+ size 260645288
fr-en/train-00019-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec063f71f0dd1102a441b4f35ed52b9fe0d7d7974a098747bb6351d0690b7c60
3
+ size 259017811
fr-en/train-00020-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f967ae536f78fd88b1af9288e75605fc8b5f36693626157402725488dbde60e1
3
+ size 260668330
fr-en/train-00021-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90e16f724e932744649233782fc6395090b4efc617a469655f20921583f3bee3
3
+ size 263558920
fr-en/train-00022-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cba68f677b05f094a28c3307b703f03c32f17b3fb4fe12cf454e0b0af8efdb16
3
+ size 267480152
fr-en/train-00023-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8092df8b3298d94b92fac524a573ede8fef55b461ce54751bb181ea3d250e220
3
+ size 270115875
fr-en/train-00024-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0ca647e8e29f9de7fa89f48acaf6451419d894810825ba7bc41070b497d8364
3
+ size 274199364
fr-en/train-00025-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faae6bba253e87a7eae24759a732154e60b57d14e4d691f630c182e00816b772
3
+ size 277510150
fr-en/train-00026-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4ed5eb02970e253dbc50162dde9dbcf5ee78c1453f627acd614006780931e17
3
+ size 364632797
fr-en/train-00027-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caf2a5d41d1d29160033b5fb7002eda6e2a6ec437d262e1609d7ca00129e7ead
3
+ size 321612859
fr-en/train-00028-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ede56cc034202b070e3635b0adfbd4cf67c413d41af6f5e9732b8255a388f599
3
+ size 370163256
fr-en/train-00029-of-00030.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d7f9dc9c2022f2deec6532b322935d55877a888ebfc48598b9fbb06ef10a0b5
3
+ size 311089369
fr-en/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:113e01f028c35bc5ac4286ce95ac9dcfb70badc3e3c71c91ab7d74fe6357fb5e
3
+ size 474541
hi-en/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c98e7f381a37db0195bba7dd8cee0e064dfd9b625d26618c9554e616c987543
3
+ size 505642
hi-en/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c7c61422711780fd6124d9652d91c84f5ea7fb2f4717ae60524611030ec9780
3
+ size 991521
hi-en/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:177a8b958eacd92e01f6583e378b70f3cd9197db2d25a1549552eee0d287204e
3
+ size 85841
ru-en/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fecb7d7b1c42bc162a2b530735dbf1eaa70f8504903426f6d43199019c0af55e
3
+ size 607619
ru-en/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:068a64892f3ef0a1b80a0a515c86d7fcf91745a93407fbb32eadf97e89699721
3
+ size 222375954
ru-en/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8817ce4db1e5d0285601d34afb8e451b90b8597502b7a7005a8da229ee363d9
3
+ size 553671
wmt14.py DELETED
@@ -1,81 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- # Lint as: python3
17
- """WMT14: Translate dataset."""
18
-
19
- import datasets
20
-
21
- from .wmt_utils import Wmt, WmtConfig
22
-
23
-
24
- _URL = "http://www.statmt.org/wmt14/translation-task.html"
25
- _CITATION = """
26
- @InProceedings{bojar-EtAl:2014:W14-33,
27
- author = {Bojar, Ondrej and Buck, Christian and Federmann, Christian and Haddow, Barry and Koehn, Philipp and Leveling, Johannes and Monz, Christof and Pecina, Pavel and Post, Matt and Saint-Amand, Herve and Soricut, Radu and Specia, Lucia and Tamchyna, Ale\v{s}},
28
- title = {Findings of the 2014 Workshop on Statistical Machine Translation},
29
- booktitle = {Proceedings of the Ninth Workshop on Statistical Machine Translation},
30
- month = {June},
31
- year = {2014},
32
- address = {Baltimore, Maryland, USA},
33
- publisher = {Association for Computational Linguistics},
34
- pages = {12--58},
35
- url = {http://www.aclweb.org/anthology/W/W14/W14-3302}
36
- }
37
- """
38
-
39
- _LANGUAGE_PAIRS = [(lang, "en") for lang in ["cs", "de", "fr", "hi", "ru"]]
40
-
41
-
42
- class Wmt14(Wmt):
43
- """WMT 14 translation datasets for all {xx, "en"} language pairs."""
44
-
45
- # Version history:
46
- # 1.0.0: S3 (new shuffling, sharding and slicing mechanism).
47
- BUILDER_CONFIGS = [
48
- WmtConfig( # pylint:disable=g-complex-comprehension
49
- description="WMT 2014 %s-%s translation task dataset." % (l1, l2),
50
- url=_URL,
51
- citation=_CITATION,
52
- language_pair=(l1, l2),
53
- version=datasets.Version("1.0.0"),
54
- )
55
- for l1, l2 in _LANGUAGE_PAIRS
56
- ]
57
-
58
- @property
59
- def manual_download_instructions(self):
60
- if self.config.language_pair[1] in ["cs", "hi", "ru"]:
61
- return "Please download the data manually as explained. TODO(PVP)"
62
- return None
63
-
64
- @property
65
- def _subsets(self):
66
- return {
67
- datasets.Split.TRAIN: [
68
- "europarl_v7",
69
- "commoncrawl",
70
- "multiun",
71
- "newscommentary_v9",
72
- "gigafren",
73
- "czeng_10",
74
- "yandexcorpus",
75
- "wikiheadlines_hi",
76
- "wikiheadlines_ru",
77
- "hindencorp_01",
78
- ],
79
- datasets.Split.VALIDATION: ["newsdev2014", "newstest2013"],
80
- datasets.Split.TEST: ["newstest2014"],
81
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wmt_utils.py DELETED
@@ -1,1022 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- # Lint as: python3
17
- """WMT: Translate dataset."""
18
-
19
-
20
- import codecs
21
- import functools
22
- import glob
23
- import gzip
24
- import itertools
25
- import os
26
- import re
27
- import xml.etree.cElementTree as ElementTree
28
-
29
- import datasets
30
-
31
-
32
- logger = datasets.logging.get_logger(__name__)
33
-
34
-
35
- _DESCRIPTION = """\
36
- Translation dataset based on the data from statmt.org.
37
-
38
- Versions exist for different years using a combination of data
39
- sources. The base `wmt` allows you to create a custom dataset by choosing
40
- your own data/language pair. This can be done as follows:
41
-
42
- ```python
43
- from datasets import inspect_dataset, load_dataset_builder
44
-
45
- inspect_dataset("wmt14", "path/to/scripts")
46
- builder = load_dataset_builder(
47
- "path/to/scripts/wmt_utils.py",
48
- language_pair=("fr", "de"),
49
- subsets={
50
- datasets.Split.TRAIN: ["commoncrawl_frde"],
51
- datasets.Split.VALIDATION: ["euelections_dev2019"],
52
- },
53
- )
54
-
55
- # Standard version
56
- builder.download_and_prepare()
57
- ds = builder.as_dataset()
58
-
59
- # Streamable version
60
- ds = builder.as_streaming_dataset()
61
- ```
62
-
63
- """
64
-
65
-
66
- CWMT_SUBSET_NAMES = ["casia2015", "casict2011", "casict2015", "datum2015", "datum2017", "neu2017"]
67
-
68
-
69
- class SubDataset:
70
- """Class to keep track of information on a sub-dataset of WMT."""
71
-
72
- def __init__(self, name, target, sources, url, path, manual_dl_files=None):
73
- """Sub-dataset of WMT.
74
-
75
- Args:
76
- name: `string`, a unique dataset identifier.
77
- target: `string`, the target language code.
78
- sources: `set<string>`, the set of source language codes.
79
- url: `string` or `(string, string)`, URL(s) or URL template(s) specifying
80
- where to download the raw data from. If two strings are provided, the
81
- first is used for the source language and the second for the target.
82
- Template strings can either contain '{src}' placeholders that will be
83
- filled in with the source language code, '{0}' and '{1}' placeholders
84
- that will be filled in with the source and target language codes in
85
- alphabetical order, or all 3.
86
- path: `string` or `(string, string)`, path(s) or path template(s)
87
- specifing the path to the raw data relative to the root of the
88
- downloaded archive. If two strings are provided, the dataset is assumed
89
- to be made up of parallel text files, the first being the source and the
90
- second the target. If one string is provided, both languages are assumed
91
- to be stored within the same file and the extension is used to determine
92
- how to parse it. Template strings should be formatted the same as in
93
- `url`.
94
- manual_dl_files: `<list>(string)` (optional), the list of files that must
95
- be manually downloaded to the data directory.
96
- """
97
- self._paths = (path,) if isinstance(path, str) else path
98
- self._urls = (url,) if isinstance(url, str) else url
99
- self._manual_dl_files = manual_dl_files if manual_dl_files else []
100
- self.name = name
101
- self.target = target
102
- self.sources = set(sources)
103
-
104
- def _inject_language(self, src, strings):
105
- """Injects languages into (potentially) template strings."""
106
- if src not in self.sources:
107
- raise ValueError(f"Invalid source for '{self.name}': {src}")
108
-
109
- def _format_string(s):
110
- if "{0}" in s and "{1}" and "{src}" in s:
111
- return s.format(*sorted([src, self.target]), src=src)
112
- elif "{0}" in s and "{1}" in s:
113
- return s.format(*sorted([src, self.target]))
114
- elif "{src}" in s:
115
- return s.format(src=src)
116
- else:
117
- return s
118
-
119
- return [_format_string(s) for s in strings]
120
-
121
- def get_url(self, src):
122
- return self._inject_language(src, self._urls)
123
-
124
- def get_manual_dl_files(self, src):
125
- return self._inject_language(src, self._manual_dl_files)
126
-
127
- def get_path(self, src):
128
- return self._inject_language(src, self._paths)
129
-
130
-
131
- # Subsets used in the training sets for various years of WMT.
132
- _TRAIN_SUBSETS = [
133
- # pylint:disable=line-too-long
134
- SubDataset(
135
- name="commoncrawl",
136
- target="en", # fr-de pair in commoncrawl_frde
137
- sources={"cs", "de", "es", "fr", "ru"},
138
- url="https://huggingface.co/datasets/wmt/wmt13/resolve/main-zip/training-parallel-commoncrawl.zip",
139
- path=("commoncrawl.{src}-en.{src}", "commoncrawl.{src}-en.en"),
140
- ),
141
- SubDataset(
142
- name="commoncrawl_frde",
143
- target="de",
144
- sources={"fr"},
145
- url=(
146
- "https://huggingface.co/datasets/wmt/wmt19/resolve/main/translation-task/fr-de/bitexts/commoncrawl.fr.gz",
147
- "https://huggingface.co/datasets/wmt/wmt19/resolve/main/translation-task/fr-de/bitexts/commoncrawl.de.gz",
148
- ),
149
- path=("", ""),
150
- ),
151
- SubDataset(
152
- name="czeng_10",
153
- target="en",
154
- sources={"cs"},
155
- url="http://ufal.mff.cuni.cz/czeng/czeng10",
156
- manual_dl_files=["data-plaintext-format.%d.tar" % i for i in range(10)],
157
- # Each tar contains multiple files, which we process specially in
158
- # _parse_czeng.
159
- path=("data.plaintext-format/??train.gz",) * 10,
160
- ),
161
- SubDataset(
162
- name="czeng_16pre",
163
- target="en",
164
- sources={"cs"},
165
- url="http://ufal.mff.cuni.cz/czeng/czeng16pre",
166
- manual_dl_files=["czeng16pre.deduped-ignoring-sections.txt.gz"],
167
- path="",
168
- ),
169
- SubDataset(
170
- name="czeng_16",
171
- target="en",
172
- sources={"cs"},
173
- url="http://ufal.mff.cuni.cz/czeng",
174
- manual_dl_files=["data-plaintext-format.%d.tar" % i for i in range(10)],
175
- # Each tar contains multiple files, which we process specially in
176
- # _parse_czeng.
177
- path=("data.plaintext-format/??train.gz",) * 10,
178
- ),
179
- SubDataset(
180
- # This dataset differs from the above in the filtering that is applied
181
- # during parsing.
182
- name="czeng_17",
183
- target="en",
184
- sources={"cs"},
185
- url="http://ufal.mff.cuni.cz/czeng",
186
- manual_dl_files=["data-plaintext-format.%d.tar" % i for i in range(10)],
187
- # Each tar contains multiple files, which we process specially in
188
- # _parse_czeng.
189
- path=("data.plaintext-format/??train.gz",) * 10,
190
- ),
191
- SubDataset(
192
- name="dcep_v1",
193
- target="en",
194
- sources={"lv"},
195
- url="https://huggingface.co/datasets/wmt/wmt17/resolve/main-zip/translation-task/dcep.lv-en.v1.zip",
196
- path=("dcep.en-lv/dcep.lv", "dcep.en-lv/dcep.en"),
197
- ),
198
- SubDataset(
199
- name="europarl_v7",
200
- target="en",
201
- sources={"cs", "de", "es", "fr"},
202
- url="https://huggingface.co/datasets/wmt/wmt13/resolve/main-zip/training-parallel-europarl-v7.zip",
203
- path=("training/europarl-v7.{src}-en.{src}", "training/europarl-v7.{src}-en.en"),
204
- ),
205
- SubDataset(
206
- name="europarl_v7_frde",
207
- target="de",
208
- sources={"fr"},
209
- url=(
210
- "https://huggingface.co/datasets/wmt/wmt19/resolve/main/translation-task/fr-de/bitexts/europarl-v7.fr.gz",
211
- "https://huggingface.co/datasets/wmt/wmt19/resolve/main/translation-task/fr-de/bitexts/europarl-v7.de.gz",
212
- ),
213
- path=("", ""),
214
- ),
215
- SubDataset(
216
- name="europarl_v8_18",
217
- target="en",
218
- sources={"et", "fi"},
219
- url="https://huggingface.co/datasets/wmt/wmt18/resolve/main-zip/translation-task/training-parallel-ep-v8.zip",
220
- path=("training/europarl-v8.{src}-en.{src}", "training/europarl-v8.{src}-en.en"),
221
- ),
222
- SubDataset(
223
- name="europarl_v8_16",
224
- target="en",
225
- sources={"fi", "ro"},
226
- url="https://huggingface.co/datasets/wmt/wmt16/resolve/main-zip/translation-task/training-parallel-ep-v8.zip",
227
- path=("training-parallel-ep-v8/europarl-v8.{src}-en.{src}", "training-parallel-ep-v8/europarl-v8.{src}-en.en"),
228
- ),
229
- SubDataset(
230
- name="europarl_v9",
231
- target="en",
232
- sources={"cs", "de", "fi", "lt"},
233
- url="https://huggingface.co/datasets/wmt/europarl/resolve/main/v9/training/europarl-v9.{src}-en.tsv.gz",
234
- path="",
235
- ),
236
- SubDataset(
237
- name="gigafren",
238
- target="en",
239
- sources={"fr"},
240
- url="https://huggingface.co/datasets/wmt/wmt10/resolve/main-zip/training-giga-fren.zip",
241
- path=("giga-fren.release2.fixed.fr.gz", "giga-fren.release2.fixed.en.gz"),
242
- ),
243
- SubDataset(
244
- name="hindencorp_01",
245
- target="en",
246
- sources={"hi"},
247
- url="http://ufallab.ms.mff.cuni.cz/~bojar/hindencorp",
248
- manual_dl_files=["hindencorp0.1.gz"],
249
- path="",
250
- ),
251
- SubDataset(
252
- name="leta_v1",
253
- target="en",
254
- sources={"lv"},
255
- url="https://huggingface.co/datasets/wmt/wmt17/resolve/main-zip/translation-task/leta.v1.zip",
256
- path=("LETA-lv-en/leta.lv", "LETA-lv-en/leta.en"),
257
- ),
258
- SubDataset(
259
- name="multiun",
260
- target="en",
261
- sources={"es", "fr"},
262
- url="https://huggingface.co/datasets/wmt/wmt13/resolve/main-zip/training-parallel-un.zip",
263
- path=("un/undoc.2000.{src}-en.{src}", "un/undoc.2000.{src}-en.en"),
264
- ),
265
- SubDataset(
266
- name="newscommentary_v9",
267
- target="en",
268
- sources={"cs", "de", "fr", "ru"},
269
- url="https://huggingface.co/datasets/wmt/wmt14/resolve/main-zip/training-parallel-nc-v9.zip",
270
- path=("training/news-commentary-v9.{src}-en.{src}", "training/news-commentary-v9.{src}-en.en"),
271
- ),
272
- SubDataset(
273
- name="newscommentary_v10",
274
- target="en",
275
- sources={"cs", "de", "fr", "ru"},
276
- url="https://huggingface.co/datasets/wmt/wmt15/resolve/main-zip/training-parallel-nc-v10.zip",
277
- path=("news-commentary-v10.{src}-en.{src}", "news-commentary-v10.{src}-en.en"),
278
- ),
279
- SubDataset(
280
- name="newscommentary_v11",
281
- target="en",
282
- sources={"cs", "de", "ru"},
283
- url="https://huggingface.co/datasets/wmt/wmt16/resolve/main-zip/translation-task/training-parallel-nc-v11.zip",
284
- path=(
285
- "training-parallel-nc-v11/news-commentary-v11.{src}-en.{src}",
286
- "training-parallel-nc-v11/news-commentary-v11.{src}-en.en",
287
- ),
288
- ),
289
- SubDataset(
290
- name="newscommentary_v12",
291
- target="en",
292
- sources={"cs", "de", "ru", "zh"},
293
- url="https://huggingface.co/datasets/wmt/wmt17/resolve/main-zip/translation-task/training-parallel-nc-v12.zip",
294
- path=("training/news-commentary-v12.{src}-en.{src}", "training/news-commentary-v12.{src}-en.en"),
295
- ),
296
- SubDataset(
297
- name="newscommentary_v13",
298
- target="en",
299
- sources={"cs", "de", "ru", "zh"},
300
- url="https://huggingface.co/datasets/wmt/wmt18/resolve/main-zip/translation-task/training-parallel-nc-v13.zip",
301
- path=(
302
- "training-parallel-nc-v13/news-commentary-v13.{src}-en.{src}",
303
- "training-parallel-nc-v13/news-commentary-v13.{src}-en.en",
304
- ),
305
- ),
306
- SubDataset(
307
- name="newscommentary_v14",
308
- target="en", # fr-de pair in newscommentary_v14_frde
309
- sources={"cs", "de", "kk", "ru", "zh"},
310
- url="http://data.statmt.org/news-commentary/v14/training/news-commentary-v14.{0}-{1}.tsv.gz",
311
- path="",
312
- ),
313
- SubDataset(
314
- name="newscommentary_v14_frde",
315
- target="de",
316
- sources={"fr"},
317
- url="http://data.statmt.org/news-commentary/v14/training/news-commentary-v14.de-fr.tsv.gz",
318
- path="",
319
- ),
320
- SubDataset(
321
- name="onlinebooks_v1",
322
- target="en",
323
- sources={"lv"},
324
- url="https://huggingface.co/datasets/wmt/wmt17/resolve/main-zip/translation-task/books.lv-en.v1.zip",
325
- path=("farewell/farewell.lv", "farewell/farewell.en"),
326
- ),
327
- SubDataset(
328
- name="paracrawl_v1",
329
- target="en",
330
- sources={"cs", "de", "et", "fi", "ru"},
331
- url="https://s3.amazonaws.com/web-language-models/paracrawl/release1/paracrawl-release1.en-{src}.zipporah0-dedup-clean.tgz", # TODO(QL): use gzip for streaming
332
- path=(
333
- "paracrawl-release1.en-{src}.zipporah0-dedup-clean.{src}",
334
- "paracrawl-release1.en-{src}.zipporah0-dedup-clean.en",
335
- ),
336
- ),
337
- SubDataset(
338
- name="paracrawl_v1_ru",
339
- target="en",
340
- sources={"ru"},
341
- url="https://s3.amazonaws.com/web-language-models/paracrawl/release1/paracrawl-release1.en-ru.zipporah0-dedup-clean.tgz", # TODO(QL): use gzip for streaming
342
- path=(
343
- "paracrawl-release1.en-ru.zipporah0-dedup-clean.ru",
344
- "paracrawl-release1.en-ru.zipporah0-dedup-clean.en",
345
- ),
346
- ),
347
- SubDataset(
348
- name="paracrawl_v3",
349
- target="en", # fr-de pair in paracrawl_v3_frde
350
- sources={"cs", "de", "fi", "lt"},
351
- url="https://s3.amazonaws.com/web-language-models/paracrawl/release3/en-{src}.bicleaner07.tmx.gz",
352
- path="",
353
- ),
354
- SubDataset(
355
- name="paracrawl_v3_frde",
356
- target="de",
357
- sources={"fr"},
358
- url=(
359
- "https://huggingface.co/datasets/wmt/wmt19/resolve/main/translation-task/fr-de/bitexts/de-fr.bicleaner07.de.gz",
360
- "https://huggingface.co/datasets/wmt/wmt19/resolve/main/translation-task/fr-de/bitexts/de-fr.bicleaner07.fr.gz",
361
- ),
362
- path=("", ""),
363
- ),
364
- SubDataset(
365
- name="rapid_2016",
366
- target="en",
367
- sources={"de", "et", "fi"},
368
- url="https://huggingface.co/datasets/wmt/wmt18/resolve/main-zip/translation-task/rapid2016.zip",
369
- path=("rapid2016.{0}-{1}.{src}", "rapid2016.{0}-{1}.en"),
370
- ),
371
- SubDataset(
372
- name="rapid_2016_ltfi",
373
- target="en",
374
- sources={"fi", "lt"},
375
- url="https://tilde-model.s3-eu-west-1.amazonaws.com/rapid2016.en-{src}.tmx.zip",
376
- path="rapid2016.en-{src}.tmx",
377
- ),
378
- SubDataset(
379
- name="rapid_2019",
380
- target="en",
381
- sources={"de"},
382
- url="https://s3-eu-west-1.amazonaws.com/tilde-model/rapid2019.de-en.zip",
383
- path=("rapid2019.de-en.de", "rapid2019.de-en.en"),
384
- ),
385
- SubDataset(
386
- name="setimes_2",
387
- target="en",
388
- sources={"ro", "tr"},
389
- url="https://object.pouta.csc.fi/OPUS-SETIMES/v2/tmx/en-{src}.tmx.gz",
390
- path="",
391
- ),
392
- SubDataset(
393
- name="uncorpus_v1",
394
- target="en",
395
- sources={"ru", "zh"},
396
- url="https://huggingface.co/datasets/wmt/uncorpus/resolve/main-zip/UNv1.0.en-{src}.zip",
397
- path=("en-{src}/UNv1.0.en-{src}.{src}", "en-{src}/UNv1.0.en-{src}.en"),
398
- ),
399
- SubDataset(
400
- name="wikiheadlines_fi",
401
- target="en",
402
- sources={"fi"},
403
- url="https://huggingface.co/datasets/wmt/wmt15/resolve/main-zip/wiki-titles.zip",
404
- path="wiki/fi-en/titles.fi-en",
405
- ),
406
- SubDataset(
407
- name="wikiheadlines_hi",
408
- target="en",
409
- sources={"hi"},
410
- url="https://huggingface.co/datasets/wmt/wmt14/resolve/main-zip/wiki-titles.zip",
411
- path="wiki/hi-en/wiki-titles.hi-en",
412
- ),
413
- SubDataset(
414
- # Verified that wmt14 and wmt15 files are identical.
415
- name="wikiheadlines_ru",
416
- target="en",
417
- sources={"ru"},
418
- url="https://huggingface.co/datasets/wmt/wmt15/resolve/main-zip/wiki-titles.zip",
419
- path="wiki/ru-en/wiki.ru-en",
420
- ),
421
- SubDataset(
422
- name="wikititles_v1",
423
- target="en",
424
- sources={"cs", "de", "fi", "gu", "kk", "lt", "ru", "zh"},
425
- url="https://huggingface.co/datasets/wmt/wikititles/resolve/main/v1/wikititles-v1.{src}-en.tsv.gz",
426
- path="",
427
- ),
428
- SubDataset(
429
- name="yandexcorpus",
430
- target="en",
431
- sources={"ru"},
432
- url="https://translate.yandex.ru/corpus?lang=en",
433
- manual_dl_files=["1mcorpus.zip"],
434
- path=("corpus.en_ru.1m.ru", "corpus.en_ru.1m.en"),
435
- ),
436
- # pylint:enable=line-too-long
437
- ] + [
438
- SubDataset( # pylint:disable=g-complex-comprehension
439
- name=ss,
440
- target="en",
441
- sources={"zh"},
442
- url="http://www.hackcha.cn/cwmt_data//%s.zip" % ss,
443
- path=("%s/*_c[hn].txt" % ss, "%s/*_en.txt" % ss),
444
- )
445
- for ss in CWMT_SUBSET_NAMES
446
- ]
447
-
448
- _DEV_SUBSETS = [
449
- SubDataset(
450
- name="euelections_dev2019",
451
- target="de",
452
- sources={"fr"},
453
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
454
- path=("dev/euelections_dev2019.fr-de.src.fr", "dev/euelections_dev2019.fr-de.tgt.de"),
455
- ),
456
- SubDataset(
457
- name="newsdev2014",
458
- target="en",
459
- sources={"hi"},
460
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
461
- path=("dev/newsdev2014.hi", "dev/newsdev2014.en"),
462
- ),
463
- SubDataset(
464
- name="newsdev2015",
465
- target="en",
466
- sources={"fi"},
467
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
468
- path=("dev/newsdev2015-fien-src.{src}.sgm", "dev/newsdev2015-fien-ref.en.sgm"),
469
- ),
470
- SubDataset(
471
- name="newsdiscussdev2015",
472
- target="en",
473
- sources={"ro", "tr"},
474
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
475
- path=("dev/newsdiscussdev2015-{src}en-src.{src}.sgm", "dev/newsdiscussdev2015-{src}en-ref.en.sgm"),
476
- ),
477
- SubDataset(
478
- name="newsdev2016",
479
- target="en",
480
- sources={"ro", "tr"},
481
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
482
- path=("dev/newsdev2016-{src}en-src.{src}.sgm", "dev/newsdev2016-{src}en-ref.en.sgm"),
483
- ),
484
- SubDataset(
485
- name="newsdev2017",
486
- target="en",
487
- sources={"lv", "zh"},
488
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
489
- path=("dev/newsdev2017-{src}en-src.{src}.sgm", "dev/newsdev2017-{src}en-ref.en.sgm"),
490
- ),
491
- SubDataset(
492
- name="newsdev2018",
493
- target="en",
494
- sources={"et"},
495
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
496
- path=("dev/newsdev2018-{src}en-src.{src}.sgm", "dev/newsdev2018-{src}en-ref.en.sgm"),
497
- ),
498
- SubDataset(
499
- name="newsdev2019",
500
- target="en",
501
- sources={"gu", "kk", "lt"},
502
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
503
- path=("dev/newsdev2019-{src}en-src.{src}.sgm", "dev/newsdev2019-{src}en-ref.en.sgm"),
504
- ),
505
- SubDataset(
506
- name="newsdiscussdev2015",
507
- target="en",
508
- sources={"fr"},
509
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
510
- path=("dev/newsdiscussdev2015-{src}en-src.{src}.sgm", "dev/newsdiscussdev2015-{src}en-ref.en.sgm"),
511
- ),
512
- SubDataset(
513
- name="newsdiscusstest2015",
514
- target="en",
515
- sources={"fr"},
516
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
517
- path=("dev/newsdiscusstest2015-{src}en-src.{src}.sgm", "dev/newsdiscusstest2015-{src}en-ref.en.sgm"),
518
- ),
519
- SubDataset(
520
- name="newssyscomb2009",
521
- target="en",
522
- sources={"cs", "de", "es", "fr"},
523
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
524
- path=("dev/newssyscomb2009.{src}", "dev/newssyscomb2009.en"),
525
- ),
526
- SubDataset(
527
- name="newstest2008",
528
- target="en",
529
- sources={"cs", "de", "es", "fr", "hu"},
530
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
531
- path=("dev/news-test2008.{src}", "dev/news-test2008.en"),
532
- ),
533
- SubDataset(
534
- name="newstest2009",
535
- target="en",
536
- sources={"cs", "de", "es", "fr"},
537
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
538
- path=("dev/newstest2009.{src}", "dev/newstest2009.en"),
539
- ),
540
- SubDataset(
541
- name="newstest2010",
542
- target="en",
543
- sources={"cs", "de", "es", "fr"},
544
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
545
- path=("dev/newstest2010.{src}", "dev/newstest2010.en"),
546
- ),
547
- SubDataset(
548
- name="newstest2011",
549
- target="en",
550
- sources={"cs", "de", "es", "fr"},
551
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
552
- path=("dev/newstest2011.{src}", "dev/newstest2011.en"),
553
- ),
554
- SubDataset(
555
- name="newstest2012",
556
- target="en",
557
- sources={"cs", "de", "es", "fr", "ru"},
558
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
559
- path=("dev/newstest2012.{src}", "dev/newstest2012.en"),
560
- ),
561
- SubDataset(
562
- name="newstest2013",
563
- target="en",
564
- sources={"cs", "de", "es", "fr", "ru"},
565
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
566
- path=("dev/newstest2013.{src}", "dev/newstest2013.en"),
567
- ),
568
- SubDataset(
569
- name="newstest2014",
570
- target="en",
571
- sources={"cs", "de", "es", "fr", "hi", "ru"},
572
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
573
- path=("dev/newstest2014-{src}en-src.{src}.sgm", "dev/newstest2014-{src}en-ref.en.sgm"),
574
- ),
575
- SubDataset(
576
- name="newstest2015",
577
- target="en",
578
- sources={"cs", "de", "fi", "ru"},
579
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
580
- path=("dev/newstest2015-{src}en-src.{src}.sgm", "dev/newstest2015-{src}en-ref.en.sgm"),
581
- ),
582
- SubDataset(
583
- name="newsdiscusstest2015",
584
- target="en",
585
- sources={"fr"},
586
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
587
- path=("dev/newsdiscusstest2015-{src}en-src.{src}.sgm", "dev/newsdiscusstest2015-{src}en-ref.en.sgm"),
588
- ),
589
- SubDataset(
590
- name="newstest2016",
591
- target="en",
592
- sources={"cs", "de", "fi", "ro", "ru", "tr"},
593
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
594
- path=("dev/newstest2016-{src}en-src.{src}.sgm", "dev/newstest2016-{src}en-ref.en.sgm"),
595
- ),
596
- SubDataset(
597
- name="newstestB2016",
598
- target="en",
599
- sources={"fi"},
600
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
601
- path=("dev/newstestB2016-enfi-ref.{src}.sgm", "dev/newstestB2016-enfi-src.en.sgm"),
602
- ),
603
- SubDataset(
604
- name="newstest2017",
605
- target="en",
606
- sources={"cs", "de", "fi", "lv", "ru", "tr", "zh"},
607
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
608
- path=("dev/newstest2017-{src}en-src.{src}.sgm", "dev/newstest2017-{src}en-ref.en.sgm"),
609
- ),
610
- SubDataset(
611
- name="newstestB2017",
612
- target="en",
613
- sources={"fi"},
614
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
615
- path=("dev/newstestB2017-fien-src.fi.sgm", "dev/newstestB2017-fien-ref.en.sgm"),
616
- ),
617
- SubDataset(
618
- name="newstest2018",
619
- target="en",
620
- sources={"cs", "de", "et", "fi", "ru", "tr", "zh"},
621
- url="https://huggingface.co/datasets/wmt/wmt19/resolve/main-zip/translation-task/dev.zip",
622
- path=("dev/newstest2018-{src}en-src.{src}.sgm", "dev/newstest2018-{src}en-ref.en.sgm"),
623
- ),
624
- ]
625
-
626
- DATASET_MAP = {dataset.name: dataset for dataset in _TRAIN_SUBSETS + _DEV_SUBSETS}
627
-
628
- _CZENG17_FILTER = SubDataset(
629
- name="czeng17_filter",
630
- target="en",
631
- sources={"cs"},
632
- url="http://ufal.mff.cuni.cz/czeng/download.php?f=convert_czeng16_to_17.pl.zip",
633
- path="convert_czeng16_to_17.pl",
634
- )
635
-
636
-
637
- class WmtConfig(datasets.BuilderConfig):
638
- """BuilderConfig for WMT."""
639
-
640
- def __init__(self, url=None, citation=None, description=None, language_pair=(None, None), subsets=None, **kwargs):
641
- """BuilderConfig for WMT.
642
-
643
- Args:
644
- url: The reference URL for the dataset.
645
- citation: The paper citation for the dataset.
646
- description: The description of the dataset.
647
- language_pair: pair of languages that will be used for translation. Should
648
- contain 2 letter coded strings. For example: ("en", "de").
649
- configuration for the `datasets.features.text.TextEncoder` used for the
650
- `datasets.features.text.Translation` features.
651
- subsets: Dict[split, list[str]]. List of the subset to use for each of the
652
- split. Note that WMT subclasses overwrite this parameter.
653
- **kwargs: keyword arguments forwarded to super.
654
- """
655
- name = "%s-%s" % (language_pair[0], language_pair[1])
656
- if "name" in kwargs: # Add name suffix for custom configs
657
- name += "." + kwargs.pop("name")
658
-
659
- super(WmtConfig, self).__init__(name=name, description=description, **kwargs)
660
-
661
- self.url = url or "http://www.statmt.org"
662
- self.citation = citation
663
- self.language_pair = language_pair
664
- self.subsets = subsets
665
-
666
- # TODO(PVP): remove when manual dir works
667
- # +++++++++++++++++++++
668
- if language_pair[1] in ["cs", "hi", "ru"]:
669
- assert NotImplementedError(f"The dataset for {language_pair[1]}-en is currently not fully supported.")
670
- # +++++++++++++++++++++
671
-
672
-
673
- class Wmt(datasets.GeneratorBasedBuilder):
674
- """WMT translation dataset."""
675
-
676
- BUILDER_CONFIG_CLASS = WmtConfig
677
-
678
- def __init__(self, *args, **kwargs):
679
- super(Wmt, self).__init__(*args, **kwargs)
680
-
681
- @property
682
- def _subsets(self):
683
- """Subsets that make up each split of the dataset."""
684
- raise NotImplementedError("This is a abstract method")
685
-
686
- @property
687
- def subsets(self):
688
- """Subsets that make up each split of the dataset for the language pair."""
689
- source, target = self.config.language_pair
690
- filtered_subsets = {}
691
- subsets = self._subsets if self.config.subsets is None else self.config.subsets
692
- for split, ss_names in subsets.items():
693
- filtered_subsets[split] = []
694
- for ss_name in ss_names:
695
- dataset = DATASET_MAP[ss_name]
696
- if dataset.target != target or source not in dataset.sources:
697
- logger.info("Skipping sub-dataset that does not include language pair: %s", ss_name)
698
- else:
699
- filtered_subsets[split].append(ss_name)
700
- logger.info("Using sub-datasets: %s", filtered_subsets)
701
- return filtered_subsets
702
-
703
- def _info(self):
704
- src, target = self.config.language_pair
705
- return datasets.DatasetInfo(
706
- description=_DESCRIPTION,
707
- features=datasets.Features(
708
- {"translation": datasets.features.Translation(languages=self.config.language_pair)}
709
- ),
710
- supervised_keys=(src, target),
711
- homepage=self.config.url,
712
- citation=self.config.citation,
713
- )
714
-
715
- def _vocab_text_gen(self, split_subsets, extraction_map, language):
716
- for _, ex in self._generate_examples(split_subsets, extraction_map, with_translation=False):
717
- yield ex[language]
718
-
719
- def _split_generators(self, dl_manager):
720
- source, _ = self.config.language_pair
721
- manual_paths_dict = {}
722
- urls_to_download = {}
723
- for ss_name in itertools.chain.from_iterable(self.subsets.values()):
724
- if ss_name == "czeng_17":
725
- # CzEng1.7 is CzEng1.6 with some blocks filtered out. We must download
726
- # the filtering script so we can parse out which blocks need to be
727
- # removed.
728
- urls_to_download[_CZENG17_FILTER.name] = _CZENG17_FILTER.get_url(source)
729
-
730
- # get dataset
731
- dataset = DATASET_MAP[ss_name]
732
- if dataset.get_manual_dl_files(source):
733
- # TODO(PVP): following two lines skip configs that are incomplete for now
734
- # +++++++++++++++++++++
735
- logger.info(f"Skipping {dataset.name} for now. Incomplete dataset for {self.config.name}")
736
- continue
737
- # +++++++++++++++++++++
738
-
739
- manual_dl_files = dataset.get_manual_dl_files(source)
740
- manual_paths = [
741
- os.path.join(os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), fname)
742
- for fname in manual_dl_files
743
- ]
744
- assert all(
745
- os.path.exists(path) for path in manual_paths
746
- ), f"For {dataset.name}, you must manually download the following file(s) from {dataset.get_url(source)} and place them in {dl_manager.manual_dir}: {', '.join(manual_dl_files)}"
747
-
748
- # set manual path for correct subset
749
- manual_paths_dict[ss_name] = manual_paths
750
- else:
751
- urls_to_download[ss_name] = dataset.get_url(source)
752
-
753
- # Download and extract files from URLs.
754
- downloaded_files = dl_manager.download_and_extract(urls_to_download)
755
- # Extract manually downloaded files.
756
- manual_files = dl_manager.extract(manual_paths_dict)
757
- extraction_map = dict(downloaded_files, **manual_files)
758
-
759
- for language in self.config.language_pair:
760
- self._vocab_text_gen(self.subsets[datasets.Split.TRAIN], extraction_map, language)
761
-
762
- return [
763
- datasets.SplitGenerator( # pylint:disable=g-complex-comprehension
764
- name=split, gen_kwargs={"split_subsets": split_subsets, "extraction_map": extraction_map}
765
- )
766
- for split, split_subsets in self.subsets.items()
767
- ]
768
-
769
- def _generate_examples(self, split_subsets, extraction_map, with_translation=True):
770
- """Returns the examples in the raw (text) form."""
771
- source, _ = self.config.language_pair
772
-
773
- def _get_local_paths(dataset, extract_dirs):
774
- rel_paths = dataset.get_path(source)
775
- if len(extract_dirs) == 1:
776
- extract_dirs = extract_dirs * len(rel_paths)
777
- return [
778
- os.path.join(ex_dir, rel_path) if rel_path else ex_dir
779
- for ex_dir, rel_path in zip(extract_dirs, rel_paths)
780
- ]
781
-
782
- def _get_filenames(dataset):
783
- rel_paths = dataset.get_path(source)
784
- urls = dataset.get_url(source)
785
- if len(urls) == 1:
786
- urls = urls * len(rel_paths)
787
- return [rel_path if rel_path else os.path.basename(url) for url, rel_path in zip(urls, rel_paths)]
788
-
789
- for ss_name in split_subsets:
790
- # TODO(PVP) remove following five lines when manual data works
791
- # +++++++++++++++++++++
792
- dataset = DATASET_MAP[ss_name]
793
- source, _ = self.config.language_pair
794
- if dataset.get_manual_dl_files(source):
795
- logger.info(f"Skipping {dataset.name} for now. Incomplete dataset for {self.config.name}")
796
- continue
797
- # +++++++++++++++++++++
798
-
799
- logger.info("Generating examples from: %s", ss_name)
800
- dataset = DATASET_MAP[ss_name]
801
- extract_dirs = extraction_map[ss_name]
802
- files = _get_local_paths(dataset, extract_dirs)
803
- filenames = _get_filenames(dataset)
804
-
805
- sub_generator_args = tuple(files)
806
-
807
- if ss_name.startswith("czeng"):
808
- if ss_name.endswith("16pre"):
809
- sub_generator = functools.partial(_parse_tsv, language_pair=("en", "cs"))
810
- elif ss_name.endswith("17"):
811
- filter_path = _get_local_paths(_CZENG17_FILTER, extraction_map[_CZENG17_FILTER.name])[0]
812
- sub_generator = functools.partial(_parse_czeng, filter_path=filter_path)
813
- else:
814
- sub_generator = _parse_czeng
815
- elif ss_name == "hindencorp_01":
816
- sub_generator = _parse_hindencorp
817
- elif len(files) == 2:
818
- if ss_name.endswith("_frde"):
819
- sub_generator = _parse_frde_bitext
820
- else:
821
- sub_generator = _parse_parallel_sentences
822
- sub_generator_args += tuple(filenames)
823
- elif len(files) == 1:
824
- fname = filenames[0]
825
- # Note: Due to formatting used by `download_manager`, the file
826
- # extension may not be at the end of the file path.
827
- if ".tsv" in fname:
828
- sub_generator = _parse_tsv
829
- elif (
830
- ss_name.startswith("newscommentary_v14")
831
- or ss_name.startswith("europarl_v9")
832
- or ss_name.startswith("wikititles_v1")
833
- ):
834
- sub_generator = functools.partial(_parse_tsv, language_pair=self.config.language_pair)
835
- elif "tmx" in fname or ss_name.startswith("paracrawl_v3"):
836
- sub_generator = _parse_tmx
837
- elif ss_name.startswith("wikiheadlines"):
838
- sub_generator = _parse_wikiheadlines
839
- else:
840
- raise ValueError("Unsupported file format: %s" % fname)
841
- else:
842
- raise ValueError("Invalid number of files: %d" % len(files))
843
-
844
- for sub_key, ex in sub_generator(*sub_generator_args):
845
- if not all(ex.values()):
846
- continue
847
- # TODO(adarob): Add subset feature.
848
- # ex["subset"] = subset
849
- key = f"{ss_name}/{sub_key}"
850
- if with_translation is True:
851
- ex = {"translation": ex}
852
- yield key, ex
853
-
854
-
855
- def _parse_parallel_sentences(f1, f2, filename1, filename2):
856
- """Returns examples from parallel SGML or text files, which may be gzipped."""
857
-
858
- def _parse_text(path, original_filename):
859
- """Returns the sentences from a single text file, which may be gzipped."""
860
- split_path = original_filename.split(".")
861
-
862
- if split_path[-1] == "gz":
863
- lang = split_path[-2]
864
-
865
- def gen():
866
- with open(path, "rb") as f, gzip.GzipFile(fileobj=f) as g:
867
- for line in g:
868
- yield line.decode("utf-8").rstrip()
869
-
870
- return gen(), lang
871
-
872
- if split_path[-1] == "txt":
873
- # CWMT
874
- lang = split_path[-2].split("_")[-1]
875
- lang = "zh" if lang in ("ch", "cn", "c[hn]") else lang
876
- else:
877
- lang = split_path[-1]
878
-
879
- def gen():
880
- with open(path, "rb") as f:
881
- for line in f:
882
- yield line.decode("utf-8").rstrip()
883
-
884
- return gen(), lang
885
-
886
- def _parse_sgm(path, original_filename):
887
- """Returns sentences from a single SGML file."""
888
- lang = original_filename.split(".")[-2]
889
- # Note: We can't use the XML parser since some of the files are badly
890
- # formatted.
891
- seg_re = re.compile(r"<seg id=\"\d+\">(.*)</seg>")
892
-
893
- def gen():
894
- with open(path, encoding="utf-8") as f:
895
- for line in f:
896
- seg_match = re.match(seg_re, line)
897
- if seg_match:
898
- assert len(seg_match.groups()) == 1
899
- yield seg_match.groups()[0]
900
-
901
- return gen(), lang
902
-
903
- parse_file = _parse_sgm if os.path.basename(f1).endswith(".sgm") else _parse_text
904
-
905
- # Some datasets (e.g., CWMT) contain multiple parallel files specified with
906
- # a wildcard. We sort both sets to align them and parse them one by one.
907
- f1_files = sorted(glob.glob(f1))
908
- f2_files = sorted(glob.glob(f2))
909
-
910
- assert f1_files and f2_files, "No matching files found: %s, %s." % (f1, f2)
911
- assert len(f1_files) == len(f2_files), "Number of files do not match: %d vs %d for %s vs %s." % (
912
- len(f1_files),
913
- len(f2_files),
914
- f1,
915
- f2,
916
- )
917
-
918
- for f_id, (f1_i, f2_i) in enumerate(zip(sorted(f1_files), sorted(f2_files))):
919
- l1_sentences, l1 = parse_file(f1_i, filename1)
920
- l2_sentences, l2 = parse_file(f2_i, filename2)
921
-
922
- for line_id, (s1, s2) in enumerate(zip(l1_sentences, l2_sentences)):
923
- key = f"{f_id}/{line_id}"
924
- yield key, {l1: s1, l2: s2}
925
-
926
-
927
- def _parse_frde_bitext(fr_path, de_path):
928
- with open(fr_path, encoding="utf-8") as fr_f:
929
- with open(de_path, encoding="utf-8") as de_f:
930
- for line_id, (s1, s2) in enumerate(zip(fr_f, de_f)):
931
- yield line_id, {"fr": s1.rstrip(), "de": s2.rstrip()}
932
-
933
-
934
- def _parse_tmx(path):
935
- """Generates examples from TMX file."""
936
-
937
- def _get_tuv_lang(tuv):
938
- for k, v in tuv.items():
939
- if k.endswith("}lang"):
940
- return v
941
- raise AssertionError("Language not found in `tuv` attributes.")
942
-
943
- def _get_tuv_seg(tuv):
944
- segs = tuv.findall("seg")
945
- assert len(segs) == 1, "Invalid number of segments: %d" % len(segs)
946
- return segs[0].text
947
-
948
- with open(path, "rb") as f:
949
- # Workaround due to: https://github.com/tensorflow/tensorflow/issues/33563
950
- utf_f = codecs.getreader("utf-8")(f)
951
- for line_id, (_, elem) in enumerate(ElementTree.iterparse(utf_f)):
952
- if elem.tag == "tu":
953
- yield line_id, {_get_tuv_lang(tuv): _get_tuv_seg(tuv) for tuv in elem.iterfind("tuv")}
954
- elem.clear()
955
-
956
-
957
- def _parse_tsv(path, language_pair=None):
958
- """Generates examples from TSV file."""
959
- if language_pair is None:
960
- lang_match = re.match(r".*\.([a-z][a-z])-([a-z][a-z])\.tsv", path)
961
- assert lang_match is not None, "Invalid TSV filename: %s" % path
962
- l1, l2 = lang_match.groups()
963
- else:
964
- l1, l2 = language_pair
965
- with open(path, encoding="utf-8") as f:
966
- for j, line in enumerate(f):
967
- cols = line.split("\t")
968
- if len(cols) != 2:
969
- logger.warning("Skipping line %d in TSV (%s) with %d != 2 columns.", j, path, len(cols))
970
- continue
971
- s1, s2 = cols
972
- yield j, {l1: s1.strip(), l2: s2.strip()}
973
-
974
-
975
- def _parse_wikiheadlines(path):
976
- """Generates examples from Wikiheadlines dataset file."""
977
- lang_match = re.match(r".*\.([a-z][a-z])-([a-z][a-z])$", path)
978
- assert lang_match is not None, "Invalid Wikiheadlines filename: %s" % path
979
- l1, l2 = lang_match.groups()
980
- with open(path, encoding="utf-8") as f:
981
- for line_id, line in enumerate(f):
982
- s1, s2 = line.split("|||")
983
- yield line_id, {l1: s1.strip(), l2: s2.strip()}
984
-
985
-
986
- def _parse_czeng(*paths, **kwargs):
987
- """Generates examples from CzEng v1.6, with optional filtering for v1.7."""
988
- filter_path = kwargs.get("filter_path", None)
989
- if filter_path:
990
- re_block = re.compile(r"^[^-]+-b(\d+)-\d\d[tde]")
991
- with open(filter_path, encoding="utf-8") as f:
992
- bad_blocks = {blk for blk in re.search(r"qw{([\s\d]*)}", f.read()).groups()[0].split()}
993
- logger.info("Loaded %d bad blocks to filter from CzEng v1.6 to make v1.7.", len(bad_blocks))
994
-
995
- for path in paths:
996
- for gz_path in sorted(glob.glob(path)):
997
- with open(gz_path, "rb") as g, gzip.GzipFile(fileobj=g) as f:
998
- filename = os.path.basename(gz_path)
999
- for line_id, line in enumerate(f):
1000
- line = line.decode("utf-8") # required for py3
1001
- if not line.strip():
1002
- continue
1003
- id_, unused_score, cs, en = line.split("\t")
1004
- if filter_path:
1005
- block_match = re.match(re_block, id_)
1006
- if block_match and block_match.groups()[0] in bad_blocks:
1007
- continue
1008
- sub_key = f"{filename}/{line_id}"
1009
- yield sub_key, {
1010
- "cs": cs.strip(),
1011
- "en": en.strip(),
1012
- }
1013
-
1014
-
1015
- def _parse_hindencorp(path):
1016
- with open(path, encoding="utf-8") as f:
1017
- for line_id, line in enumerate(f):
1018
- split_line = line.split("\t")
1019
- if len(split_line) != 5:
1020
- logger.warning("Skipping invalid HindEnCorp line: %s", line)
1021
- continue
1022
- yield line_id, {"translation": {"en": split_line[3].strip(), "hi": split_line[4].strip()}}