mC4-hindi / README.md
zicsx's picture
Update README.md
bd9746d
|
raw
history blame
2.18 kB
---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: text
dtype: string
- name: timestamp
dtype: string
- name: url
dtype: string
splits:
- name: train
num_bytes: 137146387873
num_examples: 18507273
- name: validation
num_bytes: 138079468
num_examples: 18392
download_size: 4087107539
dataset_size: 137284467341
license: apache-2.0
task_categories:
- text-generation
language:
- hi
---
# Dataset Card for "mC4-hindi"
This dataset is a subset of the mC4 dataset, which is a multilingual colossal, cleaned version of Common Crawl's web crawl corpus. It contains natural text in 101 languages, including Hindi. This dataset is specifically focused on Hindi text, and contains a variety of different types of text, including news articles, blog posts, and social media posts.
This dataset is intended to be used for training and evaluating natural language processing models for Hindi. It can be used for a variety of tasks, such as pretraining language models, machine translation, text summarization, and question-answering.
**Data format**
The dataset is in JSONL format. Each line in the file contains a JSON object with the following fields:
* `text`: field contains the text of the document.
* `timestamp`: field contains the date and time when the document was crawled.
* `url`: field contains the URL of the document.
**Data splits**
The dataset is split into two parts: train and validation. The train split contains 90% of the data, the validation split contains 5% of the data, and the test split contains 5% of the data.
**Usage**
To use the dataset, you can load it into a Hugging Face Dataset object using the following code:
```python
import datasets
dataset = datasets.load_dataset("zicsx/mC4-hindi")
```
Once you have loaded the dataset, you can access the train and validation splits using the following code:
```python
train_dataset = dataset["train"]
validation_dataset = dataset["validation"]
```
You can then use the dataset to train and evaluate your natural language processing model.