Datasets:
Commit
•
2678080
1
Parent(s):
572e4a5
Convert dataset to Parquet (#5)
Browse files- Convert dataset to Parquet (47b5113460d5546d36e624e92d3f05dcfc8d2bad)
- Add java data files (132d471ffed79f85987fbf87d185e4b566e364a4)
- Add javascript data files (f1d1c458f254351b837c542cffdc8d688b5fb438)
- Add php data files (63fb0ee869729c9171f3dd354678d665d32db222)
- Add python data files (86b786f7a6a185f7506801ee727610770c5d08e1)
- Add ruby data files (e805461fb7fb366762f045a20ca5cf68458f9de4)
- Delete loading script (87eae9979d842618e595029c495f1829cace5040)
- Delete loading script auxiliary file (9377f91f35085af8875eb96578768b3de99d28fb)
- Delete loading script auxiliary file (6013ac2dd252b219f767072ff2987d0fc35fb548)
- README.md +86 -37
- code_x_glue_ct_code_to_text.py +0 -155
- common.py +0 -75
- generated_definitions.py +0 -68
- go/test-00000-of-00001.parquet +3 -0
- go/train-00000-of-00001.parquet +3 -0
- go/validation-00000-of-00001.parquet +3 -0
- java/test-00000-of-00001.parquet +3 -0
- java/train-00000-of-00001.parquet +3 -0
- java/validation-00000-of-00001.parquet +3 -0
- javascript/test-00000-of-00001.parquet +3 -0
- javascript/train-00000-of-00001.parquet +3 -0
- javascript/validation-00000-of-00001.parquet +3 -0
- php/test-00000-of-00001.parquet +3 -0
- php/train-00000-of-00002.parquet +3 -0
- php/train-00001-of-00002.parquet +3 -0
- php/validation-00000-of-00001.parquet +3 -0
- python/test-00000-of-00001.parquet +3 -0
- python/train-00000-of-00002.parquet +3 -0
- python/train-00001-of-00002.parquet +3 -0
- python/validation-00000-of-00001.parquet +3 -0
- ruby/test-00000-of-00001.parquet +3 -0
- ruby/train-00000-of-00001.parquet +3 -0
- ruby/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -19,6 +19,13 @@ task_categories:
|
|
19 |
- translation
|
20 |
task_ids: []
|
21 |
pretty_name: CodeXGlueCtCodeToText
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
tags:
|
23 |
- code-to-text
|
24 |
dataset_info:
|
@@ -50,16 +57,16 @@ dataset_info:
|
|
50 |
dtype: string
|
51 |
splits:
|
52 |
- name: train
|
53 |
-
num_bytes:
|
54 |
num_examples: 167288
|
55 |
- name: validation
|
56 |
-
num_bytes:
|
57 |
num_examples: 7325
|
58 |
- name: test
|
59 |
-
num_bytes:
|
60 |
num_examples: 8122
|
61 |
-
download_size:
|
62 |
-
dataset_size:
|
63 |
- config_name: java
|
64 |
features:
|
65 |
- name: id
|
@@ -88,16 +95,16 @@ dataset_info:
|
|
88 |
dtype: string
|
89 |
splits:
|
90 |
- name: train
|
91 |
-
num_bytes:
|
92 |
num_examples: 164923
|
93 |
- name: validation
|
94 |
-
num_bytes:
|
95 |
num_examples: 5183
|
96 |
- name: test
|
97 |
-
num_bytes:
|
98 |
num_examples: 10955
|
99 |
-
download_size:
|
100 |
-
dataset_size:
|
101 |
- config_name: javascript
|
102 |
features:
|
103 |
- name: id
|
@@ -126,16 +133,16 @@ dataset_info:
|
|
126 |
dtype: string
|
127 |
splits:
|
128 |
- name: train
|
129 |
-
num_bytes:
|
130 |
num_examples: 58025
|
131 |
- name: validation
|
132 |
-
num_bytes:
|
133 |
num_examples: 3885
|
134 |
- name: test
|
135 |
-
num_bytes:
|
136 |
num_examples: 3291
|
137 |
-
download_size:
|
138 |
-
dataset_size:
|
139 |
- config_name: php
|
140 |
features:
|
141 |
- name: id
|
@@ -164,16 +171,16 @@ dataset_info:
|
|
164 |
dtype: string
|
165 |
splits:
|
166 |
- name: train
|
167 |
-
num_bytes:
|
168 |
num_examples: 241241
|
169 |
- name: validation
|
170 |
-
num_bytes:
|
171 |
num_examples: 12982
|
172 |
- name: test
|
173 |
-
num_bytes:
|
174 |
num_examples: 14014
|
175 |
-
download_size:
|
176 |
-
dataset_size:
|
177 |
- config_name: python
|
178 |
features:
|
179 |
- name: id
|
@@ -202,16 +209,16 @@ dataset_info:
|
|
202 |
dtype: string
|
203 |
splits:
|
204 |
- name: train
|
205 |
-
num_bytes:
|
206 |
num_examples: 251820
|
207 |
- name: validation
|
208 |
-
num_bytes:
|
209 |
num_examples: 13914
|
210 |
- name: test
|
211 |
-
num_bytes:
|
212 |
num_examples: 14918
|
213 |
-
download_size:
|
214 |
-
dataset_size:
|
215 |
- config_name: ruby
|
216 |
features:
|
217 |
- name: id
|
@@ -240,23 +247,65 @@ dataset_info:
|
|
240 |
dtype: string
|
241 |
splits:
|
242 |
- name: train
|
243 |
-
num_bytes:
|
244 |
num_examples: 24927
|
245 |
- name: validation
|
246 |
-
num_bytes:
|
247 |
num_examples: 1400
|
248 |
- name: test
|
249 |
-
num_bytes:
|
250 |
num_examples: 1261
|
251 |
-
download_size:
|
252 |
-
dataset_size:
|
253 |
-
|
254 |
-
- go
|
255 |
-
|
256 |
-
-
|
257 |
-
|
258 |
-
-
|
259 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
260 |
---
|
261 |
# Dataset Card for "code_x_glue_ct_code_to_text"
|
262 |
|
|
|
19 |
- translation
|
20 |
task_ids: []
|
21 |
pretty_name: CodeXGlueCtCodeToText
|
22 |
+
config_names:
|
23 |
+
- go
|
24 |
+
- java
|
25 |
+
- javascript
|
26 |
+
- php
|
27 |
+
- python
|
28 |
+
- ruby
|
29 |
tags:
|
30 |
- code-to-text
|
31 |
dataset_info:
|
|
|
57 |
dtype: string
|
58 |
splits:
|
59 |
- name: train
|
60 |
+
num_bytes: 342243143
|
61 |
num_examples: 167288
|
62 |
- name: validation
|
63 |
+
num_bytes: 13721860
|
64 |
num_examples: 7325
|
65 |
- name: test
|
66 |
+
num_bytes: 16328406
|
67 |
num_examples: 8122
|
68 |
+
download_size: 121341698
|
69 |
+
dataset_size: 372293409
|
70 |
- config_name: java
|
71 |
features:
|
72 |
- name: id
|
|
|
95 |
dtype: string
|
96 |
splits:
|
97 |
- name: train
|
98 |
+
num_bytes: 452553835
|
99 |
num_examples: 164923
|
100 |
- name: validation
|
101 |
+
num_bytes: 13366344
|
102 |
num_examples: 5183
|
103 |
- name: test
|
104 |
+
num_bytes: 29080753
|
105 |
num_examples: 10955
|
106 |
+
download_size: 154701399
|
107 |
+
dataset_size: 495000932
|
108 |
- config_name: javascript
|
109 |
features:
|
110 |
- name: id
|
|
|
133 |
dtype: string
|
134 |
splits:
|
135 |
- name: train
|
136 |
+
num_bytes: 160860431
|
137 |
num_examples: 58025
|
138 |
- name: validation
|
139 |
+
num_bytes: 10337344
|
140 |
num_examples: 3885
|
141 |
- name: test
|
142 |
+
num_bytes: 10190713
|
143 |
num_examples: 3291
|
144 |
+
download_size: 65788314
|
145 |
+
dataset_size: 181388488
|
146 |
- config_name: php
|
147 |
features:
|
148 |
- name: id
|
|
|
171 |
dtype: string
|
172 |
splits:
|
173 |
- name: train
|
174 |
+
num_bytes: 614654499
|
175 |
num_examples: 241241
|
176 |
- name: validation
|
177 |
+
num_bytes: 33283045
|
178 |
num_examples: 12982
|
179 |
- name: test
|
180 |
+
num_bytes: 35374993
|
181 |
num_examples: 14014
|
182 |
+
download_size: 219692158
|
183 |
+
dataset_size: 683312537
|
184 |
- config_name: python
|
185 |
features:
|
186 |
- name: id
|
|
|
209 |
dtype: string
|
210 |
splits:
|
211 |
- name: train
|
212 |
+
num_bytes: 813663148
|
213 |
num_examples: 251820
|
214 |
- name: validation
|
215 |
+
num_bytes: 46888564
|
216 |
num_examples: 13914
|
217 |
- name: test
|
218 |
+
num_bytes: 50659688
|
219 |
num_examples: 14918
|
220 |
+
download_size: 325551862
|
221 |
+
dataset_size: 911211400
|
222 |
- config_name: ruby
|
223 |
features:
|
224 |
- name: id
|
|
|
247 |
dtype: string
|
248 |
splits:
|
249 |
- name: train
|
250 |
+
num_bytes: 51956439
|
251 |
num_examples: 24927
|
252 |
- name: validation
|
253 |
+
num_bytes: 2821037
|
254 |
num_examples: 1400
|
255 |
- name: test
|
256 |
+
num_bytes: 2671551
|
257 |
num_examples: 1261
|
258 |
+
download_size: 21921316
|
259 |
+
dataset_size: 57449027
|
260 |
+
configs:
|
261 |
+
- config_name: go
|
262 |
+
data_files:
|
263 |
+
- split: train
|
264 |
+
path: go/train-*
|
265 |
+
- split: validation
|
266 |
+
path: go/validation-*
|
267 |
+
- split: test
|
268 |
+
path: go/test-*
|
269 |
+
- config_name: java
|
270 |
+
data_files:
|
271 |
+
- split: train
|
272 |
+
path: java/train-*
|
273 |
+
- split: validation
|
274 |
+
path: java/validation-*
|
275 |
+
- split: test
|
276 |
+
path: java/test-*
|
277 |
+
- config_name: javascript
|
278 |
+
data_files:
|
279 |
+
- split: train
|
280 |
+
path: javascript/train-*
|
281 |
+
- split: validation
|
282 |
+
path: javascript/validation-*
|
283 |
+
- split: test
|
284 |
+
path: javascript/test-*
|
285 |
+
- config_name: php
|
286 |
+
data_files:
|
287 |
+
- split: train
|
288 |
+
path: php/train-*
|
289 |
+
- split: validation
|
290 |
+
path: php/validation-*
|
291 |
+
- split: test
|
292 |
+
path: php/test-*
|
293 |
+
- config_name: python
|
294 |
+
data_files:
|
295 |
+
- split: train
|
296 |
+
path: python/train-*
|
297 |
+
- split: validation
|
298 |
+
path: python/validation-*
|
299 |
+
- split: test
|
300 |
+
path: python/test-*
|
301 |
+
- config_name: ruby
|
302 |
+
data_files:
|
303 |
+
- split: train
|
304 |
+
path: ruby/train-*
|
305 |
+
- split: validation
|
306 |
+
path: ruby/validation-*
|
307 |
+
- split: test
|
308 |
+
path: ruby/test-*
|
309 |
---
|
310 |
# Dataset Card for "code_x_glue_ct_code_to_text"
|
311 |
|
code_x_glue_ct_code_to_text.py
DELETED
@@ -1,155 +0,0 @@
|
|
1 |
-
import json
|
2 |
-
import os
|
3 |
-
import os.path
|
4 |
-
from typing import List
|
5 |
-
|
6 |
-
import datasets
|
7 |
-
|
8 |
-
from .common import TrainValidTestChild
|
9 |
-
from .generated_definitions import DEFINITIONS
|
10 |
-
|
11 |
-
|
12 |
-
_DESCRIPTION = """The dataset we use comes from CodeSearchNet and we filter the dataset as the following:
|
13 |
-
- Remove examples that codes cannot be parsed into an abstract syntax tree.
|
14 |
-
- Remove examples that #tokens of documents is < 3 or >256
|
15 |
-
- Remove examples that documents contain special tokens (e.g. <img ...> or https:...)
|
16 |
-
- Remove examples that documents are not English.
|
17 |
-
"""
|
18 |
-
_CITATION = """@article{husain2019codesearchnet,
|
19 |
-
title={Codesearchnet challenge: Evaluating the state of semantic code search},
|
20 |
-
author={Husain, Hamel and Wu, Ho-Hsiang and Gazit, Tiferet and Allamanis, Miltiadis and Brockschmidt, Marc},
|
21 |
-
journal={arXiv preprint arXiv:1909.09436},
|
22 |
-
year={2019}
|
23 |
-
}"""
|
24 |
-
|
25 |
-
|
26 |
-
class CodeXGlueCtCodeToTextBaseImpl(TrainValidTestChild):
|
27 |
-
_DESCRIPTION = _DESCRIPTION
|
28 |
-
_CITATION = _CITATION
|
29 |
-
|
30 |
-
# For each file, each line in the uncompressed file represents one function.
|
31 |
-
_FEATURES = {
|
32 |
-
"id": datasets.Value("int32"), # Index of the sample
|
33 |
-
"repo": datasets.Value("string"), # repo: the owner/repo
|
34 |
-
"path": datasets.Value("string"), # path: the full path to the original file
|
35 |
-
"func_name": datasets.Value("string"), # func_name: the function or method name
|
36 |
-
"original_string": datasets.Value("string"), # original_string: the raw string before tokenization or parsing
|
37 |
-
"language": datasets.Value("string"), # language: the programming language name
|
38 |
-
"code": datasets.Value("string"), # code/function: the part of the original_string that is code
|
39 |
-
"code_tokens": datasets.features.Sequence(
|
40 |
-
datasets.Value("string")
|
41 |
-
), # code_tokens/function_tokens: tokenized version of code
|
42 |
-
"docstring": datasets.Value(
|
43 |
-
"string"
|
44 |
-
), # docstring: the top-level comment or docstring, if it exists in the original string
|
45 |
-
"docstring_tokens": datasets.features.Sequence(
|
46 |
-
datasets.Value("string")
|
47 |
-
), # docstring_tokens: tokenized version of docstring
|
48 |
-
"sha": datasets.Value("string"), # sha of the file
|
49 |
-
"url": datasets.Value("string"), # url of the file
|
50 |
-
}
|
51 |
-
|
52 |
-
_SUPERVISED_KEYS = ["docstring", "docstring_tokens"]
|
53 |
-
|
54 |
-
def generate_urls(self, split_name, language):
|
55 |
-
yield "language", f"https://huggingface.co/datasets/code_search_net/resolve/main/data/{language}.zip"
|
56 |
-
yield "dataset", "dataset.zip"
|
57 |
-
|
58 |
-
def get_data_files(self, split_name, file_paths, language):
|
59 |
-
language_specific_path = file_paths["language"]
|
60 |
-
final_path = os.path.join(language_specific_path, language, "final")
|
61 |
-
# Make some cleanup to save space
|
62 |
-
for path in os.listdir(final_path):
|
63 |
-
if path.endswith(".pkl"):
|
64 |
-
os.unlink(path)
|
65 |
-
|
66 |
-
data_files = []
|
67 |
-
for root, dirs, files in os.walk(final_path):
|
68 |
-
for file in files:
|
69 |
-
temp = os.path.join(root, file)
|
70 |
-
if ".jsonl" in temp:
|
71 |
-
if split_name in temp:
|
72 |
-
data_files.append(temp)
|
73 |
-
return data_files
|
74 |
-
|
75 |
-
def post_process(self, split_name, language, js):
|
76 |
-
return js
|
77 |
-
|
78 |
-
def _generate_examples(self, split_name, file_paths, language):
|
79 |
-
import gzip
|
80 |
-
|
81 |
-
data_set_path = file_paths["dataset"]
|
82 |
-
|
83 |
-
data_files = self.get_data_files(split_name, file_paths, language)
|
84 |
-
|
85 |
-
urls = {}
|
86 |
-
f1_path_parts = [data_set_path, "dataset", language, f"{split_name}.txt"]
|
87 |
-
if self.SINGLE_LANGUAGE:
|
88 |
-
del f1_path_parts[2]
|
89 |
-
|
90 |
-
f1_path = os.path.join(*f1_path_parts)
|
91 |
-
with open(f1_path, encoding="utf-8") as f1:
|
92 |
-
for line in f1:
|
93 |
-
line = line.strip()
|
94 |
-
urls[line] = True
|
95 |
-
|
96 |
-
idx = 0
|
97 |
-
for file in data_files:
|
98 |
-
if ".gz" in file:
|
99 |
-
f = gzip.open(file)
|
100 |
-
else:
|
101 |
-
f = open(file, encoding="utf-8")
|
102 |
-
|
103 |
-
for line in f:
|
104 |
-
line = line.strip()
|
105 |
-
js = json.loads(line)
|
106 |
-
if js["url"] in urls:
|
107 |
-
js["id"] = idx
|
108 |
-
js = self.post_process(split_name, language, js)
|
109 |
-
if "partition" in js:
|
110 |
-
del js["partition"]
|
111 |
-
yield idx, js
|
112 |
-
idx += 1
|
113 |
-
f.close()
|
114 |
-
|
115 |
-
|
116 |
-
class CodeXGlueCtCodeToTextImpl(CodeXGlueCtCodeToTextBaseImpl):
|
117 |
-
SINGLE_LANGUAGE = False
|
118 |
-
|
119 |
-
def generate_urls(self, split_name):
|
120 |
-
language = self.info["parameters"]["language"]
|
121 |
-
for e in super().generate_urls(split_name, language):
|
122 |
-
yield e
|
123 |
-
|
124 |
-
def _generate_examples(self, split_name, file_paths):
|
125 |
-
language = self.info["parameters"]["language"]
|
126 |
-
for e in super()._generate_examples(split_name, file_paths, language):
|
127 |
-
yield e
|
128 |
-
|
129 |
-
|
130 |
-
CLASS_MAPPING = {
|
131 |
-
"CodeXGlueCtCodeToText": CodeXGlueCtCodeToTextImpl,
|
132 |
-
}
|
133 |
-
|
134 |
-
|
135 |
-
class CodeXGlueCtCodeToText(datasets.GeneratorBasedBuilder):
|
136 |
-
BUILDER_CONFIG_CLASS = datasets.BuilderConfig
|
137 |
-
BUILDER_CONFIGS = [
|
138 |
-
datasets.BuilderConfig(name=name, description=info["description"]) for name, info in DEFINITIONS.items()
|
139 |
-
]
|
140 |
-
|
141 |
-
def _info(self):
|
142 |
-
name = self.config.name
|
143 |
-
info = DEFINITIONS[name]
|
144 |
-
if info["class_name"] in CLASS_MAPPING:
|
145 |
-
self.child = CLASS_MAPPING[info["class_name"]](info)
|
146 |
-
else:
|
147 |
-
raise RuntimeError(f"Unknown python class for dataset configuration {name}")
|
148 |
-
ret = self.child._info()
|
149 |
-
return ret
|
150 |
-
|
151 |
-
def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]:
|
152 |
-
return self.child._split_generators(dl_manager=dl_manager)
|
153 |
-
|
154 |
-
def _generate_examples(self, split_name, file_paths):
|
155 |
-
return self.child._generate_examples(split_name, file_paths)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
common.py
DELETED
@@ -1,75 +0,0 @@
|
|
1 |
-
from typing import List
|
2 |
-
|
3 |
-
import datasets
|
4 |
-
|
5 |
-
|
6 |
-
# Citation, taken from https://github.com/microsoft/CodeXGLUE
|
7 |
-
_DEFAULT_CITATION = """@article{CodeXGLUE,
|
8 |
-
title={CodeXGLUE: A Benchmark Dataset and Open Challenge for Code Intelligence},
|
9 |
-
year={2020},}"""
|
10 |
-
|
11 |
-
|
12 |
-
class Child:
|
13 |
-
_DESCRIPTION = None
|
14 |
-
_FEATURES = None
|
15 |
-
_CITATION = None
|
16 |
-
SPLITS = {"train": datasets.Split.TRAIN}
|
17 |
-
_SUPERVISED_KEYS = None
|
18 |
-
|
19 |
-
def __init__(self, info):
|
20 |
-
self.info = info
|
21 |
-
|
22 |
-
def homepage(self):
|
23 |
-
return self.info["project_url"]
|
24 |
-
|
25 |
-
def _info(self):
|
26 |
-
# This is the description that will appear on the datasets page.
|
27 |
-
return datasets.DatasetInfo(
|
28 |
-
description=self.info["description"] + "\n\n" + self._DESCRIPTION,
|
29 |
-
features=datasets.Features(self._FEATURES),
|
30 |
-
homepage=self.homepage(),
|
31 |
-
citation=self._CITATION or _DEFAULT_CITATION,
|
32 |
-
supervised_keys=self._SUPERVISED_KEYS,
|
33 |
-
)
|
34 |
-
|
35 |
-
def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]:
|
36 |
-
SPLITS = self.SPLITS
|
37 |
-
_URL = self.info["raw_url"]
|
38 |
-
urls_to_download = {}
|
39 |
-
for split in SPLITS:
|
40 |
-
if split not in urls_to_download:
|
41 |
-
urls_to_download[split] = {}
|
42 |
-
|
43 |
-
for key, url in self.generate_urls(split):
|
44 |
-
if not url.startswith("http"):
|
45 |
-
url = _URL + "/" + url
|
46 |
-
urls_to_download[split][key] = url
|
47 |
-
|
48 |
-
downloaded_files = {}
|
49 |
-
for k, v in urls_to_download.items():
|
50 |
-
downloaded_files[k] = dl_manager.download_and_extract(v)
|
51 |
-
|
52 |
-
return [
|
53 |
-
datasets.SplitGenerator(
|
54 |
-
name=SPLITS[k],
|
55 |
-
gen_kwargs={"split_name": k, "file_paths": downloaded_files[k]},
|
56 |
-
)
|
57 |
-
for k in SPLITS
|
58 |
-
]
|
59 |
-
|
60 |
-
def check_empty(self, entries):
|
61 |
-
all_empty = all([v == "" for v in entries.values()])
|
62 |
-
all_non_empty = all([v != "" for v in entries.values()])
|
63 |
-
|
64 |
-
if not all_non_empty and not all_empty:
|
65 |
-
raise RuntimeError("Parallel data files should have the same number of lines.")
|
66 |
-
|
67 |
-
return all_empty
|
68 |
-
|
69 |
-
|
70 |
-
class TrainValidTestChild(Child):
|
71 |
-
SPLITS = {
|
72 |
-
"train": datasets.Split.TRAIN,
|
73 |
-
"valid": datasets.Split.VALIDATION,
|
74 |
-
"test": datasets.Split.TEST,
|
75 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
generated_definitions.py
DELETED
@@ -1,68 +0,0 @@
|
|
1 |
-
DEFINITIONS = {
|
2 |
-
"go": {
|
3 |
-
"class_name": "CodeXGlueCtCodeToText",
|
4 |
-
"dataset_type": "Code-Text",
|
5 |
-
"description": "CodeXGLUE code-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
6 |
-
"dir_name": "code-to-text",
|
7 |
-
"name": "go",
|
8 |
-
"parameters": {"language": "go"},
|
9 |
-
"project_url": "https://github.com/madlag/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
10 |
-
"raw_url": "https://raw.githubusercontent.com/madlag/CodeXGLUE/main/Code-Text/code-to-text",
|
11 |
-
"sizes": {"test": 8122, "train": 167288, "validation": 7325},
|
12 |
-
},
|
13 |
-
"java": {
|
14 |
-
"class_name": "CodeXGlueCtCodeToText",
|
15 |
-
"dataset_type": "Code-Text",
|
16 |
-
"description": "CodeXGLUE code-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
17 |
-
"dir_name": "code-to-text",
|
18 |
-
"name": "java",
|
19 |
-
"parameters": {"language": "java"},
|
20 |
-
"project_url": "https://github.com/madlag/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
21 |
-
"raw_url": "https://raw.githubusercontent.com/madlag/CodeXGLUE/main/Code-Text/code-to-text",
|
22 |
-
"sizes": {"test": 10955, "train": 164923, "validation": 5183},
|
23 |
-
},
|
24 |
-
"javascript": {
|
25 |
-
"class_name": "CodeXGlueCtCodeToText",
|
26 |
-
"dataset_type": "Code-Text",
|
27 |
-
"description": "CodeXGLUE code-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
28 |
-
"dir_name": "code-to-text",
|
29 |
-
"name": "javascript",
|
30 |
-
"parameters": {"language": "javascript"},
|
31 |
-
"project_url": "https://github.com/madlag/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
32 |
-
"raw_url": "https://raw.githubusercontent.com/madlag/CodeXGLUE/main/Code-Text/code-to-text",
|
33 |
-
"sizes": {"test": 3291, "train": 58025, "validation": 3885},
|
34 |
-
},
|
35 |
-
"php": {
|
36 |
-
"class_name": "CodeXGlueCtCodeToText",
|
37 |
-
"dataset_type": "Code-Text",
|
38 |
-
"description": "CodeXGLUE code-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
39 |
-
"dir_name": "code-to-text",
|
40 |
-
"name": "php",
|
41 |
-
"parameters": {"language": "php"},
|
42 |
-
"project_url": "https://github.com/madlag/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
43 |
-
"raw_url": "https://raw.githubusercontent.com/madlag/CodeXGLUE/main/Code-Text/code-to-text",
|
44 |
-
"sizes": {"test": 14014, "train": 241241, "validation": 12982},
|
45 |
-
},
|
46 |
-
"python": {
|
47 |
-
"class_name": "CodeXGlueCtCodeToText",
|
48 |
-
"dataset_type": "Code-Text",
|
49 |
-
"description": "CodeXGLUE code-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
50 |
-
"dir_name": "code-to-text",
|
51 |
-
"name": "python",
|
52 |
-
"parameters": {"language": "python"},
|
53 |
-
"project_url": "https://github.com/madlag/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
54 |
-
"raw_url": "https://raw.githubusercontent.com/madlag/CodeXGLUE/main/Code-Text/code-to-text",
|
55 |
-
"sizes": {"test": 14918, "train": 251820, "validation": 13914},
|
56 |
-
},
|
57 |
-
"ruby": {
|
58 |
-
"class_name": "CodeXGlueCtCodeToText",
|
59 |
-
"dataset_type": "Code-Text",
|
60 |
-
"description": "CodeXGLUE code-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
61 |
-
"dir_name": "code-to-text",
|
62 |
-
"name": "ruby",
|
63 |
-
"parameters": {"language": "ruby"},
|
64 |
-
"project_url": "https://github.com/madlag/CodeXGLUE/tree/main/Code-Text/code-to-text",
|
65 |
-
"raw_url": "https://raw.githubusercontent.com/madlag/CodeXGLUE/main/Code-Text/code-to-text",
|
66 |
-
"sizes": {"test": 1261, "train": 24927, "validation": 1400},
|
67 |
-
},
|
68 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
go/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e4faa7c186c60891fad6dba97adde01e706243e2a10f03e39f457a3db1600cb
|
3 |
+
size 5432557
|
go/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da6535760b76ab1aa1a74f2e57672f7bdcad3d7cbe23e44836125ae6da190a69
|
3 |
+
size 111614231
|
go/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22a813598a30ed4e526873a98d292fc0710caef769748370ca5275bcead95a60
|
3 |
+
size 4294910
|
java/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:38e2806527d689ee75109e035ba3d3031497b1c9c0bea0055795ac72c8557c94
|
3 |
+
size 9378593
|
java/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c1f85dd441e2ea99fd9c9d04e26784d75a0bffc6f1c91b5eb8499ab86fda20b
|
3 |
+
size 141076943
|
java/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a4890bb491a03e3442a15b454b41d202fd95521d4a506a5fc5afddcb581755d
|
3 |
+
size 4245863
|
javascript/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb485b1b996bd78cd5fea0c9105b3c648d72c268b97b63344d09fce7b5ef37d0
|
3 |
+
size 3593407
|
javascript/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8fa783615e9ddbefea9eb56a945c66e303f4d617c79ae7ec936346bbf1236c2d
|
3 |
+
size 58415544
|
javascript/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:967e3a6d2d0f730ca1d194aef0ca64dabec98741b4ca3003dc0c73ad73b67923
|
3 |
+
size 3779363
|
php/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:def714be2c74037aaa08a22167e46952fd1e61d0e4d3457f6f0110c2aebe2b96
|
3 |
+
size 11160469
|
php/train-00000-of-00002.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dee7ca8ecdfe70c3b0213666844aeaffd526173ddafb74b2c31b599379f8c9de
|
3 |
+
size 97692232
|
php/train-00001-of-00002.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7cf5cc5d45cfd1ed4f505e8ce97084e7521d9f8289363896f904e4e5b2a1915
|
3 |
+
size 100304628
|
php/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:12487b8af8760b45ea89c8f3243bb1533c9cb8c0c0ccc77fd33489eb9192ed48
|
3 |
+
size 10534829
|
python/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69acf4b46f35ffdf5e4184615d99feb08b96b388ae6891744c0b73ae3f564e9e
|
3 |
+
size 17972583
|
python/train-00000-of-00002.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9decfb7a7e4eb455fc585815b4fad828830183e3b7be8266638a7abf924a56cd
|
3 |
+
size 144077099
|
python/train-00001-of-00002.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7f108a8a5418f7f8ad267965ab39c81096a1cc8812019f3e240499e8d05364c
|
3 |
+
size 146851339
|
python/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e078006b8829aed4cb087e2003e76d5308b94817740fd9af817ac5cf52a24ce2
|
3 |
+
size 16650841
|
ruby/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39597f63eb27343b21bc68a532fffd739760a38bec6bdba8eeceb4e6ccd04a48
|
3 |
+
size 1030532
|
ruby/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a96dc16c07f1802ddd3f46f82680527e18cabeacf1d5e00954cdb69b5962faf4
|
3 |
+
size 19831678
|
ruby/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef0c3acd6ebd6be0bf36859debbef7695915ed52304d06f66222a9c1c276d86e
|
3 |
+
size 1059106
|