Edit model card

cardiffnlp/roberta-base-topic-multi

This model is a fine-tuned version of roberta-base on the cardiffnlp/tweet_topic_multi via tweetnlp. Training split is train_all and parameters have been tuned on the validation split validation_2021.

Following metrics are achieved on the test split test_2021 (link).

  • F1 (micro): 0.7546616383825687
  • F1 (macro): 0.5959450154471646
  • Accuracy: 0.5318642048838594

Usage

Install tweetnlp via pip.

pip install tweetnlp

Load the model in python.

import tweetnlp
model = tweetnlp.Classifier("cardiffnlp/roberta-base-topic-multi", max_length=128)
model.predict('Get the all-analog Classic Vinyl Edition of "Takin Off" Album from {@herbiehancock@} via {@bluenoterecords@} link below {{URL}}')

Reference

@inproceedings{camacho-collados-etal-2022-tweetnlp,
    title={{T}weet{NLP}: {C}utting-{E}dge {N}atural {L}anguage {P}rocessing for {S}ocial {M}edia},
    author={Camacho-Collados, Jose and Rezaee, Kiamehr and Riahi, Talayeh and Ushio, Asahi and Loureiro, Daniel and Antypas, Dimosthenis and Boisson, Joanne and Espinosa-Anke, Luis and Liu, Fangyu and Mart{'\i}nez-C{'a}mara, Eugenio and others},
    author = "Ushio, Asahi  and
      Camacho-Collados, Jose",
    booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
    month = nov,
    year = "2022",
    address = "Abu Dhabi, U.A.E.",
    publisher = "Association for Computational Linguistics",
}
Downloads last month
11
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train cardiffnlp/roberta-base-topic-multi

Evaluation results

  • Micro F1 (cardiffnlp/tweet_topic_multi) on cardiffnlp/tweet_topic_multi
    self-reported
    0.755
  • Macro F1 (cardiffnlp/tweet_topic_multi) on cardiffnlp/tweet_topic_multi
    self-reported
    0.596
  • Accuracy (cardiffnlp/tweet_topic_multi) on cardiffnlp/tweet_topic_multi
    self-reported
    0.532