orafandina commited on
Commit
c82ef74
1 Parent(s): 3c9ae25

Upload 9 files

Browse files
wiki_600K_long_inputs/data-00000-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317b5c5071a067ed1663193f653e47c391a8ac052f829be0c9eb25b9011228db
3
+ size 837050800
wiki_600K_long_inputs/data-00001-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c45b07c65f8f96e0cf69c38ecb654aeb32ccc4f26d3f320bba99aad3ade9cc46
3
+ size 496444448
wiki_600K_long_inputs/data-00002-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5620e0201c10ca1f20354aa86b89fe8731cf68d7e8d59347a15be7115a138148
3
+ size 388555192
wiki_600K_long_inputs/data-00003-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3449565bc618fc97b9f23f283a617c22af1263110a698a146d0a850b6702de
3
+ size 386005952
wiki_600K_long_inputs/data-00004-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3e54f6758c07cc8ecb94e4928b58ebe72fa987fdec100e0cac41fe237be3d15
3
+ size 352191984
wiki_600K_long_inputs/data-00005-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dd9a0e69f6c7303504f210892c07190054ef33d288ff6e30c335f28e28003bc
3
+ size 356734016
wiki_600K_long_inputs/data-00006-of-00007.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf54fe9a2e18df9f8a66e538dfa3684a19cd3537ccd22fd0bab86c180e0ddf81
3
+ size 341596088
wiki_600K_long_inputs/dataset_info.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "wikipedia",
3
+ "citation": "@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n",
4
+ "config_name": "20220301.en",
5
+ "dataset_size": 20275516160,
6
+ "description": "Wikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(https://dumps.wikimedia.org/) with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n",
7
+ "download_checksums": {},
8
+ "download_size": 20598313936,
9
+ "features": {
10
+ "text": {
11
+ "dtype": "string",
12
+ "_type": "Value"
13
+ },
14
+ "input_ids": {
15
+ "feature": {
16
+ "dtype": "int32",
17
+ "_type": "Value"
18
+ },
19
+ "_type": "Sequence"
20
+ },
21
+ "token_type_ids": {
22
+ "feature": {
23
+ "dtype": "int8",
24
+ "_type": "Value"
25
+ },
26
+ "_type": "Sequence"
27
+ },
28
+ "attention_mask": {
29
+ "feature": {
30
+ "dtype": "int8",
31
+ "_type": "Value"
32
+ },
33
+ "_type": "Sequence"
34
+ }
35
+ },
36
+ "homepage": "https://dumps.wikimedia.org",
37
+ "license": "",
38
+ "size_in_bytes": 40873830096,
39
+ "splits": {
40
+ "train": {
41
+ "name": "train",
42
+ "num_bytes": 20275516160,
43
+ "num_examples": 6458670,
44
+ "dataset_name": "wikipedia"
45
+ }
46
+ },
47
+ "version": {
48
+ "version_str": "2.0.0",
49
+ "description": "",
50
+ "major": 2,
51
+ "minor": 0,
52
+ "patch": 0
53
+ }
54
+ }
wiki_600K_long_inputs/state.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00007.arrow"
5
+ },
6
+ {
7
+ "filename": "data-00001-of-00007.arrow"
8
+ },
9
+ {
10
+ "filename": "data-00002-of-00007.arrow"
11
+ },
12
+ {
13
+ "filename": "data-00003-of-00007.arrow"
14
+ },
15
+ {
16
+ "filename": "data-00004-of-00007.arrow"
17
+ },
18
+ {
19
+ "filename": "data-00005-of-00007.arrow"
20
+ },
21
+ {
22
+ "filename": "data-00006-of-00007.arrow"
23
+ }
24
+ ],
25
+ "_fingerprint": "bd434c95b5f63fc6",
26
+ "_format_columns": null,
27
+ "_format_kwargs": {},
28
+ "_format_type": null,
29
+ "_output_all_columns": false,
30
+ "_split": "train"
31
+ }