Datasets:
license: mit
dataset_info:
features:
- name: URL
dtype: string
- name: Wiki
dtype: string
- name: Language
dtype: string
- name: Title
dtype: string
- name: Abstract
dtype: string
- name: Version Control
dtype: string
splits:
- name: data
num_bytes: 3849026435
num_examples: 6575217
- name: data1
num_bytes: 1339234317
num_examples: 2565263
- name: data2
num_bytes: 299141019
num_examples: 345314
- name: data3
num_bytes: 121110572
num_examples: 194533
- name: data4
num_bytes: 763143777
num_examples: 1474721
- name: data5
num_bytes: 4402642
num_examples: 4238
- name: data6
num_bytes: 921270984
num_examples: 1777137
- name: data7
num_bytes: 634758702
num_examples: 1080168
- name: data8
num_bytes: 1490198322
num_examples: 2501467
- name: data9
num_bytes: 1116300754
num_examples: 1890629
- name: data10
num_bytes: 568620663
num_examples: 1315710
- name: data11
num_bytes: 1555258518
num_examples: 1854603
- name: data12
num_bytes: 787383297
num_examples: 1197801
- name: data13
num_bytes: 160792644
num_examples: 153847
- name: data14
num_bytes: 3861239238
num_examples: 6082427
- name: data15
num_bytes: 1143640444
num_examples: 2522834
- name: data16
num_bytes: 244946
num_examples: 727
download_size: 8522442108
dataset_size: 18615767274
configs:
- config_name: default
data_files:
- split: data
path: data/data-*
- split: data1
path: data/data1-*
- split: data2
path: data/data2-*
- split: data3
path: data/data3-*
- split: data4
path: data/data4-*
- split: data5
path: data/data5-*
- split: data6
path: data/data6-*
- split: data7
path: data/data7-*
- split: data8
path: data/data8-*
- split: data9
path: data/data9-*
- split: data10
path: data/data10-*
- split: data11
path: data/data11-*
- split: data12
path: data/data12-*
- split: data13
path: data/data13-*
- split: data14
path: data/data14-*
- split: data15
path: data/data15-*
- split: data16
path: data/data16-*
Introducing Wikipedia X, a comprehensive dataset encompassing abstracts, complete articles, and a popularity score index for both widely spoken and lesser-known Wikipedia subsets. Our dedication to Wikipedia-X ensures a centralized Wikipedia dataset that undergoes regular updates and adheres to the highest standards.
A central focus of our efforts was to include exotic languages that often lack up-to-date Wikipedia dumps or may not have any dumps at all. Languages such as Hebrew, Urdu, Bengali, Aramaic, Uighur, and Polish were prioritized to ensure high-quality processed Wikipedia datasets are accessible for these languages with substantial speaker bases. This initiative aims to enable Artificial Intelligence to thrive across all languages, breaking down language barriers and fostering inclusivity.
Notice: We're continuously updating this dataset every 8 months as part of a broader effort at LAION AI dedicated to textual embeddings. If you'd like to see a specific language added, please don't hesitate to reach out to us.
(Please note that this dataset is actively under development.)
Language | Code |
---|---|
English | en |
German | de |
Polish | pl |
Spanish | es |
Hebrew | he |
French | fr |
Chinese | zh |
Italian | it |
Russian | ru |
Urdu | ur |
Portuguese | pt |
Aramaic | ar |
Cebuano | ceb |
Swedish | sv |
Uighur | ug |
Bengali | bn |