albertvillanova's picture
Convert dataset to Parquet (#4)
d1e9fb2 verified
|
raw
history blame
7.82 kB
metadata
annotations_creators:
  - expert-generated
language_creators:
  - crowdsourced
language:
  - ur
license:
  - mit
multilinguality:
  - monolingual
size_categories:
  - 1K<n<10K
source_datasets:
  - original
task_categories:
  - text-classification
task_ids:
  - multi-class-classification
pretty_name: roman_urdu_hate_speech
tags:
  - binary classification
dataset_info:
  - config_name: Coarse_Grained
    features:
      - name: tweet
        dtype: string
      - name: label
        dtype:
          class_label:
            names:
              '0': Abusive/Offensive
              '1': Normal
    splits:
      - name: train
        num_bytes: 725715
        num_examples: 7208
      - name: test
        num_bytes: 202318
        num_examples: 2002
      - name: validation
        num_bytes: 79755
        num_examples: 800
    download_size: 730720
    dataset_size: 1007788
  - config_name: Fine_Grained
    features:
      - name: tweet
        dtype: string
      - name: label
        dtype:
          class_label:
            names:
              '0': Abusive/Offensive
              '1': Normal
              '2': Religious Hate
              '3': Sexism
              '4': Profane/Untargeted
    splits:
      - name: train
        num_bytes: 723666
        num_examples: 7208
      - name: test
        num_bytes: 203590
        num_examples: 2002
      - name: validation
        num_bytes: 723666
        num_examples: 7208
    download_size: 1199660
    dataset_size: 1650922
configs:
  - config_name: Coarse_Grained
    data_files:
      - split: train
        path: Coarse_Grained/train-*
      - split: test
        path: Coarse_Grained/test-*
      - split: validation
        path: Coarse_Grained/validation-*
    default: true
  - config_name: Fine_Grained
    data_files:
      - split: train
        path: Fine_Grained/train-*
      - split: test
        path: Fine_Grained/test-*
      - split: validation
        path: Fine_Grained/validation-*

Dataset Card for roman_urdu_hate_speech

Table of Contents

Dataset Description

Dataset Summary

The Roman Urdu Hate-Speech and Offensive Language Detection (RUHSOLD) dataset is a Roman Urdu dataset of tweets annotated by experts in the relevant language. The authors develop the gold-standard for two sub-tasks. First sub-task is based on binary labels of Hate-Offensive content and Normal content (i.e., inoffensive language). These labels are self-explanatory. The authors refer to this sub-task as coarse-grained classification. Second sub-task defines Hate-Offensive content with four labels at a granular level. These labels are the most relevant for the demographic of users who converse in RU and are defined in related literature. The authors refer to this sub-task as fine-grained classification. The objective behind creating two gold-standards is to enable the researchers to evaluate the hate speech detection approaches on both easier (coarse-grained) and challenging (fine-grained) scenarios.

Supported Tasks and Leaderboards

  • 'multi-class-classification', 'text-classification-other-binary classification': The dataset can be used for both multi class classification as well as for binary classification as it contains both coarse grained and fine grained labels.

Languages

The text of this dataset is Roman Urdu. The associated BCP-47 code is 'ur'.

Dataset Structure

Data Instances

The dataset consists of two parts divided as a set of two types, Coarse grained examples and Fine Grained examples. The difference is that in the coarse grained example the tweets are labelled as abusive or normal whereas in the fine grained version there are several classes of hate associated with a tweet.

For the Coarse grained segment of the dataset the label mapping is:- Task 1: Coarse-grained Classification Labels 0: Abusive/Offensive 1: Normal

Whereas for the Fine Grained segment of the dataset the label mapping is:- Task 2: Fine-grained Classification Labels 0: Abusive/Offensive 1: Normal 2: Religious Hate 3: Sexism 4: Profane/Untargeted

An example from Roman Urdu Hate Speech looks as follows:

{
  'tweet': 'there are some yahodi daboo like imran chore zakat khore'
  'label': 0
}

Data Fields

-tweet:a string denoting the tweet which has been selected by using a random sampling from a tweet base of 50000 tweets to select 10000 tweets and annotated for the dataset.

-label:An annotation manually labeled by three independent annotators, during the annotation process, all conflicts are resolved by a majority vote among three annotators.

Data Splits

The data of each of the segments, Coarse Grained and Fine Grained is further split into training, validation and test set. The data is split in train, test, and validation sets with 70,20,10 split ratio using stratification based on fine-grained labels.

The use of stratified sampling is deemed necessary to preserve the same labels ratio across all splits.

The Final split sizes are as follows:

Train Valid Test 7209 2003 801

Dataset Creation

Curation Rationale

[More Information Needed]

Source Data

Initial Data Collection and Normalization

[More Information Needed]

Who are the source language producers?

[More Information Needed]

Annotations

Annotation process

[More Information Needed]

Who are the annotators?

[More Information Needed]

Personal and Sensitive Information

[More Information Needed]

Considerations for Using the Data

Social Impact of Dataset

[More Information Needed]

Discussion of Biases

[More Information Needed]

Other Known Limitations

[More Information Needed]

Additional Information

Dataset Curators

The dataset was created by Hammad Rizwan, Muhammad Haroon Shakeel, Asim Karim during work done at Department of Computer Science, Lahore University of Management Sciences (LUMS), Lahore, Pakistan.

Licensing Information

The licensing status of the dataset hinges on the legal status of the Roman Urdu Hate Speech Dataset Repository which is under MIT License.

Citation Information

@inproceedings{rizwan2020hate,
  title={Hate-speech and offensive language detection in roman Urdu},
  author={Rizwan, Hammad and Shakeel, Muhammad Haroon and Karim, Asim},
  booktitle={Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)},
  pages={2512--2522},
  year={2020}
}

Contributions

Thanks to @bp-high, for adding this dataset.