Datasets:

Convert dataset to Parquet

#2
by lawhy - opened
OntoLAMA.py DELETED
@@ -1,213 +0,0 @@
1
- # Copyright 2020 The HuggingFace Datasets Authors.
2
- # Copyright 2023 Yuan He.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- # TODO: Address all TODOs and remove all explanatory comments
16
- """OntoLAMA Dataset Loading Script"""
17
-
18
-
19
- import csv
20
- import json
21
- import os
22
-
23
- import datasets
24
-
25
-
26
- # TODO: Add BibTeX citation
27
- # Find for instance the citation on arxiv or on the dataset repo/website
28
- _CITATION = """\
29
- @inproceedings{he2023language,
30
- title={Language Model Analysis for Ontology Subsumption Inference},
31
- author={He, Yuan and Chen, Jiaoyan and Jimenez-Ruiz, Ernesto and Dong, Hang and Horrocks, Ian},
32
- booktitle={Findings of the Association for Computational Linguistics: ACL 2023},
33
- pages={3439--3453},
34
- year={2023}
35
- }
36
- """
37
-
38
- # TODO: Add description of the dataset here
39
- # You can copy an official description
40
- _DESCRIPTION = """\
41
- OntoLAMA: LAnguage Model Analysis datasets for Ontology Subsumption Inference.
42
- """
43
-
44
- _URL = lambda name: f"https://zenodo.org/record/7700458/files/{name}.zip?download=1"
45
-
46
- # TODO: Add a link to an official homepage for the dataset here
47
- _HOMEPAGE = "https://krr-oxford.github.io/DeepOnto/"
48
-
49
- # TODO: Add the licence for the dataset here if you can find it
50
- _LICENSE = "Apache License, Version 2.0"
51
-
52
-
53
- # TODO: Name of the dataset usually matches the script name with CamelCase instead of snake_case
54
- class OntoLAMA(datasets.GeneratorBasedBuilder):
55
- """TODO: Short description of my dataset."""
56
-
57
- VERSION = datasets.Version("1.0.0")
58
-
59
- # This is an example of a dataset with multiple configurations.
60
- # If you don't want/need to define several sub-sets in your dataset,
61
- # just remove the BUILDER_CONFIG_CLASS and the BUILDER_CONFIGS attributes.
62
-
63
- # If you need to make complex sub-parts in the datasets with configurable options
64
- # You can create your own builder configuration class to store attribute, inheriting from datasets.BuilderConfig
65
- # BUILDER_CONFIG_CLASS = MyBuilderConfig
66
-
67
- # You will be able to load one or the other configurations in the following list with
68
- # data = datasets.load_dataset('my_dataset', 'first_domain')
69
- # data = datasets.load_dataset('my_dataset', 'second_domain')
70
- BUILDER_CONFIGS = [
71
- datasets.BuilderConfig(
72
- name="bimnli", version=VERSION, description="BiMNLI dataset created from the MNLI dataset."
73
- ),
74
- datasets.BuilderConfig(
75
- name="schemaorg-atomic-SI",
76
- version=VERSION,
77
- description="Atomic SI dataset created from the Schema.org Ontology.",
78
- ),
79
- datasets.BuilderConfig(
80
- name="doid-atomic-SI", version=VERSION, description="Atomic SI dataset created from the Disease Ontology."
81
- ),
82
- datasets.BuilderConfig(
83
- name="foodon-atomic-SI", version=VERSION, description="Atomic SI dataset created from the Food Ontology."
84
- ),
85
- datasets.BuilderConfig(
86
- name="foodon-complex-SI", version=VERSION, description="Complex SI dataset created from the Gene Ontology."
87
- ),
88
- datasets.BuilderConfig(
89
- name="go-atomic-SI", version=VERSION, description="Atomic SI dataset created from the Gene Ontology."
90
- ),
91
- datasets.BuilderConfig(
92
- name="go-complex-SI", version=VERSION, description="Complex SI dataset created from the Gene Ontology."
93
- ),
94
- ]
95
-
96
- def _info(self):
97
- # TODO: This method specifies the datasets.DatasetInfo object which contains informations and typings for the dataset
98
- if "atomic-SI" in self.config.name: # This is the name of the configuration selected in BUILDER_CONFIGS above
99
- features = datasets.Features(
100
- {
101
- "v_sub_concept": datasets.Value("string"),
102
- "v_super_concept": datasets.Value("string"),
103
- "label": datasets.ClassLabel(
104
- num_classes=2, names=["negative_subsumption", "positive_subsumption"], names_file=None, id=None
105
- ),
106
- "axiom": datasets.Value("string"),
107
- # These are the features of your dataset like images, labels ...
108
- }
109
- )
110
- elif (
111
- "complex-SI" in self.config.name
112
- ): # This is an example to show how to have different features for "first_domain" and "second_domain"
113
- features = datasets.Features(
114
- {
115
- "v_sub_concept": datasets.Value("string"),
116
- "v_super_concept": datasets.Value("string"),
117
- "label": datasets.ClassLabel(
118
- num_classes=2, names=["negative_subsumption", "positive_subsumption"], names_file=None, id=None
119
- ),
120
- "axiom": datasets.Value("string"),
121
- "anchor_axiom": datasets.Value("string") # the equivalence axiom used as anchor
122
- # These are the features of your dataset like images, labels ...
123
- }
124
- )
125
- elif self.config.name == "bimnli":
126
- features = datasets.Features(
127
- {
128
- "premise": datasets.Value("string"),
129
- "hypothesis": datasets.Value("string"),
130
- "label": datasets.ClassLabel(
131
- num_classes=2, names=["contradiction", "entailment"], names_file=None, id=None
132
- ),
133
- }
134
- )
135
-
136
- return datasets.DatasetInfo(
137
- # This is the description that will appear on the datasets page.
138
- description=_DESCRIPTION,
139
- # This defines the different columns of the dataset and their types
140
- features=features, # Here we define them above because they are different between the two configurations
141
- # If there's a common (input, target) tuple from the features, uncomment supervised_keys line below and
142
- # specify them. They'll be used if as_supervised=True in builder.as_dataset.
143
- # supervised_keys=("sentence", "label"),
144
- # Homepage of the dataset for documentation
145
- homepage=_HOMEPAGE,
146
- # License for the dataset if available
147
- license=_LICENSE,
148
- # Citation for the dataset
149
- citation=_CITATION,
150
- )
151
-
152
- def _split_generators(self, dl_manager):
153
- # TODO: This method is tasked with downloading/extracting the data and defining the splits depending on the configuration
154
- # If several configurations are possible (listed in BUILDER_CONFIGS), the configuration selected by the user is in self.config.name
155
-
156
- # dl_manager is a datasets.download.DownloadManager that can be used to download and extract URLS
157
- # It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
158
- # By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
159
- urls = _URL(self.config.name)
160
- data_dir = dl_manager.download_and_extract(urls)
161
- return [
162
- datasets.SplitGenerator(
163
- name=datasets.Split.TRAIN,
164
- # These kwargs will be passed to _generate_examples
165
- gen_kwargs={
166
- "filepath": os.path.join(data_dir, self.config.name, "train.jsonl"),
167
- "split": "train",
168
- },
169
- ),
170
- datasets.SplitGenerator(
171
- name=datasets.Split.VALIDATION,
172
- # These kwargs will be passed to _generate_examples
173
- gen_kwargs={
174
- "filepath": os.path.join(data_dir, self.config.name, "dev.jsonl"),
175
- "split": "dev",
176
- },
177
- ),
178
- datasets.SplitGenerator(
179
- name=datasets.Split.TEST,
180
- # These kwargs will be passed to _generate_examples
181
- gen_kwargs={"filepath": os.path.join(data_dir, self.config.name, "test.jsonl"), "split": "test"},
182
- ),
183
- ]
184
-
185
- # method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
186
- def _generate_examples(self, filepath, split):
187
- # TODO: This method handles input defined in _split_generators to yield (key, example) tuples from the dataset.
188
- # The `key` is for legacy reasons (tfds) and is not important in itself, but must be unique for each example.
189
- with open(filepath, encoding="utf-8") as f:
190
- for key, row in enumerate(f):
191
- data = json.loads(row)
192
- if "atomic-SI" in self.config.name:
193
- # Yields examples as (key, example) tuples
194
- yield key, {
195
- "v_sub_concept": data["v_sub_concept"],
196
- "v_super_concept": data["v_super_concept"],
197
- "label": data["label"],
198
- "axiom": data["axiom"],
199
- }
200
- elif "complex-SI" in self.config.name:
201
- yield key, {
202
- "v_sub_concept": data["v_sub_concept"],
203
- "v_super_concept": data["v_super_concept"],
204
- "label": data["label"],
205
- "axiom": data["axiom"],
206
- "anchor_axiom": data["anchor_axiom"],
207
- }
208
- elif self.config.name == "bimnli":
209
- yield key, {
210
- "premise": data["premise"],
211
- "hypothesis": data["hypothesis"],
212
- "label": data["label"],
213
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md CHANGED
@@ -14,32 +14,30 @@ size_categories:
14
  language:
15
  - en
16
  dataset_info:
17
- - config_name: schemaorg-atomic-SI
18
  features:
19
- - name: v_sub_concept
20
  dtype: string
21
- - name: v_super_concept
22
  dtype: string
23
  - name: label
24
  dtype:
25
  class_label:
26
  names:
27
- '0': negative_subsumption
28
- '1': positive_subsumption
29
- - name: axiom
30
- dtype: string
31
  splits:
32
  - name: train
33
- num_bytes: 103485
34
- num_examples: 808
35
  - name: validation
36
- num_bytes: 51523
37
- num_examples: 404
38
  - name: test
39
- num_bytes: 361200
40
- num_examples: 2830
41
- download_size: 82558
42
- dataset_size: 516208
43
  - config_name: doid-atomic-SI
44
  features:
45
  - name: v_sub_concept
@@ -64,7 +62,7 @@ dataset_info:
64
  - name: test
65
  num_bytes: 1977582
66
  num_examples: 11314
67
- download_size: 3184028
68
  dataset_size: 19759219
69
  - config_name: foodon-atomic-SI
70
  features:
@@ -90,9 +88,9 @@ dataset_info:
90
  - name: test
91
  num_bytes: 16098373
92
  num_examples: 96062
93
- download_size: 28499028
94
  dataset_size: 160926634
95
- - config_name: go-atomic-SI
96
  features:
97
  - name: v_sub_concept
98
  dtype: string
@@ -106,43 +104,47 @@ dataset_info:
106
  '1': positive_subsumption
107
  - name: axiom
108
  dtype: string
 
 
109
  splits:
110
  - name: train
111
- num_bytes: 152537233
112
- num_examples: 772870
113
  - name: validation
114
- num_bytes: 19060490
115
- num_examples: 96608
116
  - name: test
117
- num_bytes: 19069265
118
- num_examples: 96610
119
- download_size: 32379717
120
- dataset_size: 190666988
121
- - config_name: bimnli
122
  features:
123
- - name: premise
124
  dtype: string
125
- - name: hypothesis
126
  dtype: string
127
  - name: label
128
  dtype:
129
  class_label:
130
  names:
131
- '0': contradiction
132
- '1': entailment
 
 
133
  splits:
134
  - name: train
135
- num_bytes: 43363266
136
- num_examples: 235622
137
  - name: validation
138
- num_bytes: 4818648
139
- num_examples: 26180
140
  - name: test
141
- num_bytes: 2420273
142
- num_examples: 12906
143
- download_size: 19264134
144
- dataset_size: 50602187
145
- - config_name: foodon-complex-SI
146
  features:
147
  - name: v_sub_concept
148
  dtype: string
@@ -160,17 +162,17 @@ dataset_info:
160
  dtype: string
161
  splits:
162
  - name: train
163
- num_bytes: 2553731
164
- num_examples: 3754
165
  - name: validation
166
- num_bytes: 1271721
167
- num_examples: 1850
168
  - name: test
169
- num_bytes: 8926305
170
- num_examples: 13080
171
- download_size: 1064602
172
- dataset_size: 12751757
173
- - config_name: go-complex-SI
174
  features:
175
  - name: v_sub_concept
176
  dtype: string
@@ -184,20 +186,75 @@ dataset_info:
184
  '1': positive_subsumption
185
  - name: axiom
186
  dtype: string
187
- - name: anchor_axiom
188
- dtype: string
189
  splits:
190
  - name: train
191
- num_bytes: 45328802
192
- num_examples: 72318
193
  - name: validation
194
- num_bytes: 5671713
195
- num_examples: 9040
196
  - name: test
197
- num_bytes: 5667069
198
- num_examples: 9040
199
- download_size: 5059364
200
- dataset_size: 56667584
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  ---
202
 
203
  # OntoLAMA: LAnguage Model Analysis for Ontology Subsumption Inference
 
14
  language:
15
  - en
16
  dataset_info:
17
+ - config_name: bimnli
18
  features:
19
+ - name: premise
20
  dtype: string
21
+ - name: hypothesis
22
  dtype: string
23
  - name: label
24
  dtype:
25
  class_label:
26
  names:
27
+ '0': contradiction
28
+ '1': entailment
 
 
29
  splits:
30
  - name: train
31
+ num_bytes: 43363266
32
+ num_examples: 235622
33
  - name: validation
34
+ num_bytes: 4818648
35
+ num_examples: 26180
36
  - name: test
37
+ num_bytes: 2420273
38
+ num_examples: 12906
39
+ download_size: 34515774
40
+ dataset_size: 50602187
41
  - config_name: doid-atomic-SI
42
  features:
43
  - name: v_sub_concept
 
62
  - name: test
63
  num_bytes: 1977582
64
  num_examples: 11314
65
+ download_size: 5117922
66
  dataset_size: 19759219
67
  - config_name: foodon-atomic-SI
68
  features:
 
88
  - name: test
89
  num_bytes: 16098373
90
  num_examples: 96062
91
+ download_size: 45668013
92
  dataset_size: 160926634
93
+ - config_name: foodon-complex-SI
94
  features:
95
  - name: v_sub_concept
96
  dtype: string
 
104
  '1': positive_subsumption
105
  - name: axiom
106
  dtype: string
107
+ - name: anchor_axiom
108
+ dtype: string
109
  splits:
110
  - name: train
111
+ num_bytes: 2553731
112
+ num_examples: 3754
113
  - name: validation
114
+ num_bytes: 1271721
115
+ num_examples: 1850
116
  - name: test
117
+ num_bytes: 8926305
118
+ num_examples: 13080
119
+ download_size: 2028889
120
+ dataset_size: 12751757
121
+ - config_name: go-atomic-SI
122
  features:
123
+ - name: v_sub_concept
124
  dtype: string
125
+ - name: v_super_concept
126
  dtype: string
127
  - name: label
128
  dtype:
129
  class_label:
130
  names:
131
+ '0': negative_subsumption
132
+ '1': positive_subsumption
133
+ - name: axiom
134
+ dtype: string
135
  splits:
136
  - name: train
137
+ num_bytes: 152537233
138
+ num_examples: 772870
139
  - name: validation
140
+ num_bytes: 19060490
141
+ num_examples: 96608
142
  - name: test
143
+ num_bytes: 19069265
144
+ num_examples: 96610
145
+ download_size: 52657016
146
+ dataset_size: 190666988
147
+ - config_name: go-complex-SI
148
  features:
149
  - name: v_sub_concept
150
  dtype: string
 
162
  dtype: string
163
  splits:
164
  - name: train
165
+ num_bytes: 45328802
166
+ num_examples: 72318
167
  - name: validation
168
+ num_bytes: 5671713
169
+ num_examples: 9040
170
  - name: test
171
+ num_bytes: 5667069
172
+ num_examples: 9040
173
+ download_size: 9668613
174
+ dataset_size: 56667584
175
+ - config_name: schemaorg-atomic-SI
176
  features:
177
  - name: v_sub_concept
178
  dtype: string
 
186
  '1': positive_subsumption
187
  - name: axiom
188
  dtype: string
 
 
189
  splits:
190
  - name: train
191
+ num_bytes: 103485
192
+ num_examples: 808
193
  - name: validation
194
+ num_bytes: 51523
195
+ num_examples: 404
196
  - name: test
197
+ num_bytes: 361200
198
+ num_examples: 2830
199
+ download_size: 144649
200
+ dataset_size: 516208
201
+ configs:
202
+ - config_name: bimnli
203
+ data_files:
204
+ - split: train
205
+ path: bimnli/train-*
206
+ - split: validation
207
+ path: bimnli/validation-*
208
+ - split: test
209
+ path: bimnli/test-*
210
+ - config_name: doid-atomic-SI
211
+ data_files:
212
+ - split: train
213
+ path: doid-atomic-SI/train-*
214
+ - split: validation
215
+ path: doid-atomic-SI/validation-*
216
+ - split: test
217
+ path: doid-atomic-SI/test-*
218
+ - config_name: foodon-atomic-SI
219
+ data_files:
220
+ - split: train
221
+ path: foodon-atomic-SI/train-*
222
+ - split: validation
223
+ path: foodon-atomic-SI/validation-*
224
+ - split: test
225
+ path: foodon-atomic-SI/test-*
226
+ - config_name: foodon-complex-SI
227
+ data_files:
228
+ - split: train
229
+ path: foodon-complex-SI/train-*
230
+ - split: validation
231
+ path: foodon-complex-SI/validation-*
232
+ - split: test
233
+ path: foodon-complex-SI/test-*
234
+ - config_name: go-atomic-SI
235
+ data_files:
236
+ - split: train
237
+ path: go-atomic-SI/train-*
238
+ - split: validation
239
+ path: go-atomic-SI/validation-*
240
+ - split: test
241
+ path: go-atomic-SI/test-*
242
+ - config_name: go-complex-SI
243
+ data_files:
244
+ - split: train
245
+ path: go-complex-SI/train-*
246
+ - split: validation
247
+ path: go-complex-SI/validation-*
248
+ - split: test
249
+ path: go-complex-SI/test-*
250
+ - config_name: schemaorg-atomic-SI
251
+ data_files:
252
+ - split: train
253
+ path: schemaorg-atomic-SI/train-*
254
+ - split: validation
255
+ path: schemaorg-atomic-SI/validation-*
256
+ - split: test
257
+ path: schemaorg-atomic-SI/test-*
258
  ---
259
 
260
  # OntoLAMA: LAnguage Model Analysis for Ontology Subsumption Inference
bimnli/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ec00ad5ec522017c6875a375112f6e899e363ca1719b5375b761ea869c5a15a
3
+ size 1633618
bimnli/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e6ebc1b1fb9608c1c2a41475c55df339e4e9518ca4bcf0f3d56daf86d99ea8
3
+ size 29586283
bimnli/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf16daf1176cd2cce6de2703711907e8947fc16e58f504d91e1c25a11dcb9e1e
3
+ size 3295873
doid-atomic-SI/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4d020736b42b56308e401bd84bdd1abaf21af27021a660e5db31c5e3a5209bc
3
+ size 517226
doid-atomic-SI/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f90ea761d426883011e040ecf1418cac68612a2b08952a4673a1f5d218fd12
3
+ size 4084651
doid-atomic-SI/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6db9c3187c88098812d68bbf457d5e4c347dd46488dcfb0853447b8e2619acb5
3
+ size 516045
foodon-atomic-SI/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5601600c8e59dbb6f903325a39bbb0a6aa8e5473cd19d89c8b5966d6d6b5ad7
3
+ size 4573029
foodon-atomic-SI/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b417200ab36f43daba5891d84c4dc6e62ae58086a4297902c5f5904c72cbd4e2
3
+ size 36526578
foodon-atomic-SI/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d29e805c5fc0b227ea0d5f0d647bbd1c79088e1433b4df63f8b9594f0b9925c
3
+ size 4568406
foodon-complex-SI/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2588ae58f7a77f64b9f0eab6e9e5af74911a56217d87a8fd9e2c870f979e7f7
3
+ size 1415513
foodon-complex-SI/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f20c603cce4334826d808299a6e137d77abc1f0678f9f22d4014f5ca879fa23
3
+ size 410364
foodon-complex-SI/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb0ebb39e35074b40c144ed711f00171eb2932347902a6cade635c07aa58cd80
3
+ size 203012
go-atomic-SI/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a78b3bdd6a8c52b2cc0bf48df3f8b215950d3a09591a28a561d4a28efbbe5d06
3
+ size 5270283
go-atomic-SI/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73eaaf33ab385c7197e6e4d46f6832eb3977a4ec7a027327ed41c30554ffb26a
3
+ size 42123289
go-atomic-SI/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8767754dfa2ca2128f5685ecbd9f81b1ee8c919c0d2af61925a025b57f1aa756
3
+ size 5263444
go-complex-SI/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5482d3038414c26e83a67cfa879a02d50d9b0cfdda0d81c93ca9b44fe7acd7e4
3
+ size 970439
go-complex-SI/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c37ac6f92e796dd299a95be0280d5a454188ab655feab55761c6fa212c8a2756
3
+ size 7725341
go-complex-SI/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c1f51e7a0f48a9cc8bdfcbb3579a5322e0a814bf0d7998cbd621dcbecdf9bf6
3
+ size 972833
schemaorg-atomic-SI/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3ed1beda695dd00c5dbe0a835b78188ab755e8ca4da017ef7a02d965ba31016
3
+ size 96968
schemaorg-atomic-SI/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:675c41d715978903e2a3485c0b9f40c470f7e6133cc9e8ac71288dc223334d88
3
+ size 29696
schemaorg-atomic-SI/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61655d9acd5bb905dbd027fccd26dcbc876df6ad864cb676ce6dcd1d890be2d2
3
+ size 17985