Edit model card

t2t-assert-ade-balanced

t2t-assert-ade-balanced is a text-to-text (t2t) adverse drug event (ade) detection model trained with over- and undersampled (balanced) English tweets reporting adverse drug events. It is trained as part of BOUN-TABI system for the Social Media Mining for Health (SMM4H) 2022 shared task. The system description paper has been accepted for publication in Proceedings of the Seventh Social Media Mining for Health (#SMM4H) Workshop and Shared Task and will be available soon. The source code has been released on GitHub at https://github.com/gokceuludogan/boun-tabi-smm4h22.

The model utilizes the T5 model and its text-to-text formulation. The inputs are fed to the model with the task prefix "assert ade:", followed with a sentence/tweet. In turn, the output "adverse event problem" or "healthy okay" is received.

Requirements

sentencepiece
transformers

Usage

from transformers import pipeline, AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("yirmibesogluz/t2t-assert-ade-balanced")
model = AutoModelForSeq2SeqLM.from_pretrained("yirmibesogluz/t2t-assert-ade-balanced")
predictor = pipeline("text2text-generation", model=model, tokenizer=tokenizer)
predictor("assert ade: joints killing me now i have gone back up on the lamotrigine. sick of side effects. sick of meds. want my own self back. knackered today")

Citation

@inproceedings{uludogan-gokce-yirmibesoglu-zeynep-2022-boun-tabi-smm4h22,
    title = "{BOUN}-{TABI}@{SMM4H}'22: Text-to-{T}ext {A}dverse {D}rug {E}vent {E}xtraction with {D}ata {B}alancing and {P}rompting",
    author = "Uludo{\u{g}}an, G{\"{o}}k{\c{c}}e  and Yirmibe{\c{s}}o{\u{g}}lu, Zeynep",
    booktitle = "Proceedings of the Seventh Social Media Mining for Health ({\#}SMM4H) Workshop and Shared Task",
    year = "2022",
}
Downloads last month
5
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.