Datasets:
Tasks:
Text Classification
Formats:
csv
Languages:
Portuguese
Size:
10K - 100K
ArXiv:
Tags:
hate-speech-detection
DOI:
License:
File size: 5,584 Bytes
471f1b9 6c8d705 6864451 9650b6a 6c8d705 9650b6a 6c8d705 5008c87 6864451 6c8d705 6864451 6c8d705 2adb3f1 0eab748 2adb3f1 e33e66a 5649254 0eab748 00bf85b 2adb3f1 00bf85b 2adb3f1 e39a97d 00bf85b 9cf3d78 00bf85b 9cf3d78 df7764a b71d8ba 9cf3d78 5d968a2 1d90000 5d968a2 9982541 9cf3d78 e39a97d 9982541 7a3aee5 9982541 9cf3d78 7e8446f 2adb3f1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 |
---
license: cc-by-4.0
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- pt
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids: []
pretty_name: TuPy
language_bcp47:
- pt-BR
tags:
- hate-speech-detection
configs:
- config_name: multilabel
data_files:
- split: train
path: multilabel/multilabel_train.csv
- split: test
path: multilabel/multilabel_test.csv
default: true
- config_name: binary
data_files:
- split: train
path: binary/binary_train.csv
- split: test
path: binary/binary_test.csv
---
# Portuguese Hate Speech Dataset (TuPy)
The Portuguese hate speech dataset (TuPy) is an annotated corpus designed to facilitate the development of advanced hate speech detection models using machine learning (ML)
and natural language processing (NLP) techniques. TuPy is comprised of 10,000 (ten thousand) unpublished, annotated, and anonymized documents collected
on Twitter (currently known as X) in 2023.
This repository is organized as follows:
```sh
root.
├── binary : binary dataset (including training and testing split)
├── multilabel : multilabel dataset (including training and testing split)
└── README.md : documentation and card metadata
```
## Security measures
To safeguard user identity and uphold the integrity of this dataset, all user mentions have been anonymized as "@user," and any references to external websites have been omitted
## Annotation and voting process
To generate the binary matrices, we employed a straightforward voting process. Three distinct evaluations were assigned to each document. In cases where a document received two or more identical classifications, the adopted value is set to 1; otherwise, it is marked as 0.Raw data can be checked into the repository in the [project repository](https://github.com/Silly-Machine/TuPy-Dataset)
The subsequent table provides a concise summary of the annotators' profiles and qualifications:
#### Table 1 – Annotators
| Annotator | Gender | Education | Political | Color |
|--------------|--------|-----------------------------------------------|------------|--------|
| Annotator 1 | Female | Ph.D. Candidate in civil engineering | Far-left | White |
| Annotator 2 | Male | Master's candidate in human rights | Far-left | Black |
| Annotator 3 | Female | Master's degree in behavioral psychology | Liberal | White |
| Annotator 4 | Male | Master's degree in behavioral psychology | Right-wing | Black |
| Annotator 5 | Female | Ph.D. Candidate in behavioral psychology | Liberal | Black |
| Annotator 6 | Male | Ph.D. Candidate in linguistics | Far-left | White |
| Annotator 7 | Female | Ph.D. Candidate in civil engineering | Liberal | White |
| Annotator 8 | Male | Ph.D. Candidate in civil engineering | Liberal | Black |
| Annotator 9 | Male | Master's degree in behavioral psychology | Far-left | White |
## Data structure
A data point comprises the tweet text (a string) along with thirteen categories, each category is assigned a value of 0 when there is an absence of aggressive or hateful content and a value of 1 when such content is present. These values represent the consensus of annotators regarding the presence of aggressive, hate, ageism, aporophobia, body shame, capacitism, lgbtphobia, political, racism, religious intolerance, misogyny, xenophobia, and others. An illustration from the multilabel TuPy dataset is depicted below:
```python
{
text: "e tem pobre de direita imbecil que ainda defendia a manutenção da política de preços atrelada ao dólar link",
aggressive: 1, hate: 1, ageism: 0, aporophobia: 1, body shame: 0, capacitism: 0, lgbtphobia: 0, political: 1, racism : 0,
religious intolerance : 0, misogyny : 0, xenophobia : 0, other : 0
}
```
# Dataset content
Table 2 provides a detailed breakdown of the dataset, delineating the volume of data based on the occurrence of aggressive speech and the manifestation of hate speech within the documents
#### Table 2 - Count of non-aggressive and aggressive documents
| Label | Count |
|----------------------|--------|
| Non-aggressive | 8013 |
| Aggressive - Not hate| 689 |
| Aggressive - Hate | 1298 |
| Total | 10000 |
Table 3 provides a detailed analysis of the dataset, delineating the data volume in relation to the occurrence of distinct categories of hate speech.
#### Table 3 - Count of documents for hate categories
| Label | Count |
|--------------------------|-------|
| Ageism | 53 |
| Aporophobia | 61 |
| Body shame | 120 |
| Capacitism | 92 |
| LGBTphobia | 96 |
| Political | 532 |
| Racism | 38 |
| Religious intolerance | 28 |
| Misogyny | 207 |
| Xenophobia | 70 |
| Other | 1 |
| Total | 1298 |
# Acknowledge
The TuPy project is the result of the development of Felipe Oliveira's thesis and the work of several collaborators. This project is financed by the Federal University of Rio de Janeiro ([UFRJ](https://ufrj.br/)) and the Alberto Luiz Coimbra Institute for Postgraduate Studies and Research in Engineering ([COPPE](https://coppe.ufrj.br/)). |