orionweller commited on
Commit
4b0b56a
1 Parent(s): ae7c587

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +30 -0
  2. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  4. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  5. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  6. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds +3 -0
  7. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds +3 -0
  8. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds +3 -0
  9. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds +3 -0
  10. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds +3 -0
  11. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  12. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  13. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  14. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  15. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  16. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  17. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  18. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds +3 -0
  19. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds +3 -0
  20. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds +3 -0
  21. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds +3 -0
  22. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds +3 -0
  23. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds +3 -0
  24. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds +3 -0
  25. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds +3 -0
  26. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds +3 -0
  27. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds +3 -0
  28. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds +3 -0
  29. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds +3 -0
  30. train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds +3 -0
  31. train/cc_en_head/cc_en_head_0185-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds +3 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10716-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10716-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11596-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11596-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12641-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12641-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1320-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1320-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18308-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18308-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1895-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1895-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20476-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20476-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21324-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21324-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22092-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22092-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22863-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
.gitattributes CHANGED
@@ -11167,3 +11167,33 @@ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/
11167
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
11168
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
11169
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11167
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
11168
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
11169
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
11170
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
11171
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds filter=lfs diff=lfs merge=lfs -text
11172
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
11173
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds filter=lfs diff=lfs merge=lfs -text
11174
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text
11175
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds filter=lfs diff=lfs merge=lfs -text
11176
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
11177
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text
11178
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
11179
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
11180
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
11181
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
11182
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text
11183
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
11184
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text
11185
+ train/cc_en_head/cc_en_head_0185-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
11186
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11187
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text
11188
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
11189
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
11190
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds filter=lfs diff=lfs merge=lfs -text
11191
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text
11192
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
11193
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text
11194
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11195
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
11196
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
11197
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
11198
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11199
+ train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5b839beb3a265356311a860d4030fd722692657a4c4634654ea19a88dbad764
3
+ size 67107484
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9932226e5074992f2621959a8a2f37074e00b799eeffaf71465c099534333c03
3
+ size 67108436
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f9febb42d49a4b5f4e68918f35aefcf81cc1f1615e982f92009644fc85c1782
3
+ size 67106890
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3199f3390dde98c433c427923b789354c0154d69bd6c40a981df0945f241dfc
3
+ size 67108303
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:144308d877a848eddbe7cb97d9844ad54fd91252fb6d33828d6285cde267f007
3
+ size 67108475
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71ce36752129a439f5fe25e41e0733e6a979b5e49cd56308412856a751828e0d
3
+ size 67106970
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8071e92ed1a24b917dba08123b477b36e079f32ff938637e74c46ae1b1cc0ad7
3
+ size 67108557
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59a3b8dbb69c97e0eb6c1099d17bb82f1b961dc6b1721b0601a8fb3f9eb2e8d3
3
+ size 67108435
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b880df07349f98c43417a095205d8dc182e69efcf0434146ab6a03da7c3500b4
3
+ size 67107231
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbc45d8b0ee255b7489a925ac8975c50c99cec42bd37d9a85cefc3ca9f282941
3
+ size 67108330
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea652f3d4866144e3b5249aa02a64a70cd1a725fd776b03ea64d46dea8b3eaef
3
+ size 67107555
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b45c8d898f4e103ca7fa9a923bb5e936c5cbc17cf38cb3121a3234aae30af875
3
+ size 67107337
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22df17a506b1b28cafe8758c3937d72e1616f72e8adae825f5a59d32e1a37044
3
+ size 67107425
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06bea9ac22b405beb18515ae9a4ab001b286b24bceb1aa7b594df67034c8d58d
3
+ size 67108626
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d3b388dc0186a7fa8ce3460e7588ebbc58644621ab90f23e9e1d9414bd73609
3
+ size 67108501
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24a2f07a0a36e619b6a6e264adf9f1b5cb925273c9ed3875cf2443243e339bc1
3
+ size 67107266
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91ed933f23d9e12f4724e5cc1ea45d0456bbea9f4b73ce691d117b311f90c91f
3
+ size 67107888
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7635cc42614d120315b8ca5606f1fc6213606bf73238450f5cdbf07c5d0601d4
3
+ size 67108089
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:996e191a5df3d95844f6fd7659712b95dfb155dca7ef9c03d54a4f667923f9c0
3
+ size 67108364
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56a25cd8d1219a78e88b3aa79b3234cf91596a34fe3a536e1c27863dd275c778
3
+ size 67108389
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eea076ca34bbc1a1c38f9388175d5dba82f4e55dd80200b4e9ee6c39eb214d1
3
+ size 67106833
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:033f956d89ff27b45fef4c00c90ce203079b6f6809cec6c487f39c32676d58b0
3
+ size 67108308
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5410a38711625ca22f7e89cb57755659caf6ee6d9289664070461191bb701504
3
+ size 67107565
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:411c615ede8cf2b1fa3fd725ba46dd0a20cdb9aedfcb76628af697918b5b6227
3
+ size 67108507
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bba42566cb4c42876001ea59a9f1022c792f130f8caa35bed935024a8b0c8df8
3
+ size 67108602
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05c097d9de356504a171cba8826f23c380bc40d357110e25f4608ed2d853244e
3
+ size 67107423
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3896cafa94aae429d34fee49df050b7aee51aea377b5f89ceb2fa0e4fd078ae2
3
+ size 67107759
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8f6034456bd1961283d19cc75fe4b62f48f9af035f6cbe5761145e25e76bb21
3
+ size 67107908
train/cc_en_head/cc_en_head_0120-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98633753ed474bcb80e6c40ab611b1669dcc7694798a1d25ae24979eafc93e99
3
+ size 67107363
train/cc_en_head/cc_en_head_0185-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:718ef4feaac4715ba7cc201fe930501eaea7c6b16c1bbfc753d49240d036e442
3
+ size 67107228
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10716-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108390, "hashes": {}}, "samples": 43598, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47736984, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15363473, "hashes": {}}, "samples": 9956, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10886830, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10716-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39578442,
3
+ "num_truncated_tokens": 39545938
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11596-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107018, "hashes": {}}, "samples": 44924, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47889031, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 6875243, "hashes": {}}, "samples": 4524, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4898286, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11596-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 35460843,
3
+ "num_truncated_tokens": 35435375
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12641-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107845, "hashes": {}}, "samples": 44537, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47875640, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9530733, "hashes": {}}, "samples": 6351, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6825909, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12641-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36744543,
3
+ "num_truncated_tokens": 36716160
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1320-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108786, "hashes": {}}, "samples": 45005, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48010312, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 7100395, "hashes": {}}, "samples": 4675, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5119954, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1320-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 35567231,
3
+ "num_truncated_tokens": 35540252
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18308-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108346, "hashes": {}}, "samples": 43646, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47979923, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13570439, "hashes": {}}, "samples": 9025, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9728907, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18308-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38709173,
3
+ "num_truncated_tokens": 38677885
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1895-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107375, "hashes": {}}, "samples": 43883, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47628552, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12364212, "hashes": {}}, "samples": 8044, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8814725, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1895-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38128635,
3
+ "num_truncated_tokens": 38098808
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20476-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107222, "hashes": {}}, "samples": 43325, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47608659, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16154401, "hashes": {}}, "samples": 10624, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11424813, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20476-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39961078,
3
+ "num_truncated_tokens": 39927437
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21324-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108015, "hashes": {}}, "samples": 43136, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47538160, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17596487, "hashes": {}}, "samples": 11282, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12494594, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21324-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40667912,
3
+ "num_truncated_tokens": 40634213
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22092-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108451, "hashes": {}}, "samples": 42976, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47656172, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19187036, "hashes": {}}, "samples": 12152, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13616002, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22092-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41441356,
3
+ "num_truncated_tokens": 41406119
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22863-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108333, "hashes": {}}, "samples": 44231, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48065175, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10794445, "hashes": {}}, "samples": 7190, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7730835, "hashes": {}}}], "version": 2}