mC4-hindi / README.md
zicsx's picture
Update README.md
bd9746d
|
raw
history blame
2.18 kB
metadata
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
      - split: validation
        path: data/validation-*
dataset_info:
  features:
    - name: text
      dtype: string
    - name: timestamp
      dtype: string
    - name: url
      dtype: string
  splits:
    - name: train
      num_bytes: 137146387873
      num_examples: 18507273
    - name: validation
      num_bytes: 138079468
      num_examples: 18392
  download_size: 4087107539
  dataset_size: 137284467341
license: apache-2.0
task_categories:
  - text-generation
language:
  - hi

Dataset Card for "mC4-hindi"

This dataset is a subset of the mC4 dataset, which is a multilingual colossal, cleaned version of Common Crawl's web crawl corpus. It contains natural text in 101 languages, including Hindi. This dataset is specifically focused on Hindi text, and contains a variety of different types of text, including news articles, blog posts, and social media posts.

This dataset is intended to be used for training and evaluating natural language processing models for Hindi. It can be used for a variety of tasks, such as pretraining language models, machine translation, text summarization, and question-answering.

Data format

The dataset is in JSONL format. Each line in the file contains a JSON object with the following fields:

  • text: field contains the text of the document.
  • timestamp: field contains the date and time when the document was crawled.
  • url: field contains the URL of the document.

Data splits

The dataset is split into two parts: train and validation. The train split contains 90% of the data, the validation split contains 5% of the data, and the test split contains 5% of the data.

Usage

To use the dataset, you can load it into a Hugging Face Dataset object using the following code:

import datasets

dataset = datasets.load_dataset("zicsx/mC4-hindi")

Once you have loaded the dataset, you can access the train and validation splits using the following code:

train_dataset = dataset["train"]
validation_dataset = dataset["validation"]

You can then use the dataset to train and evaluate your natural language processing model.