mapsoriano's picture
Update README.md
9f6dc22 verified
|
raw
history blame
2.76 kB
metadata
license: cc-by-sa-4.0
base_model: jcblaise/roberta-tagalog-base
tags:
  - generated_from_trainer
  - tagalog
  - filipino
  - twitter
metrics:
  - accuracy
  - precision
  - recall
  - f1
model-index:
  - name: roberta-tagalog-base-philippine-elections-2016-2022-hate-speech
    results: []
datasets:
  - hate_speech_filipino
  - mapsoriano/2016_2022_hate_speech_filipino
language:
  - tl
  - en

roberta-tagalog-base-philippine-elections-2016-2022-hate-speech

This model is a fine-tuned version of jcblaise/roberta-tagalog-base for the task of Text Classification, classifying hate and non-hate tweets.

The model was fine-tuned on a combined dataset mapsoriano/2016_2022_hate_speech_filipino consisting of the hate_speech_filipino dataset and a newly crawled 2022 Philippine Presidential Elections-related Tweets Hate Speech Dataset.

It achieves the following results on the evaluation (validation) set:

  • Loss: 0.3574
  • Accuracy: 0.8743

It achieves the following results on the test set:

  • Accuracy: 0.8783
  • Precision: 0.8563
  • Recall: 0.9077
  • F1: 0.8813

Feel free to connect via LinkedIn for further information on this model or on the study that it was used on.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Accuracy
0.3423 1.0 1361 0.3167 0.8693
0.2194 2.0 2722 0.3574 0.8743

Framework versions

  • Transformers 4.33.2
  • Pytorch 2.0.1+cu118
  • Datasets 2.14.5
  • Tokenizers 0.13.3

Citation Information

Research Title: Application of BERT in Detecting Online Hate

Published: 2023

Authors:

  • Castro, D.
  • Dizon, L. J.
  • Sarip, A. J.
  • Soriano, M. A.

Feel free to connect via LinkedIn for further information on this model or on the study that it was used on.