|
--- |
|
license: cc-by-nc-3.0 |
|
--- |
|
|
|
PragS2: Pragmatic Masked Language Modeling with Emoji_any dataset followed by Hashtag-Based Surrogate Fine-Tuning |
|
|
|
You can load this model and use for downstream fine-tuning. For example: |
|
```python |
|
from transformers import AutoTokenizer, AutoModelForSequenceClassification |
|
|
|
tokenizer = AutoTokenizer.from_pretrained('UBC-NLP/prags2', use_fast = True) |
|
model = AutoModelForSequenceClassification.from_pretrained('UBC-NLP/prags2',num_labels=lable_size) |
|
``` |
|
|
|
|
|
More details are in our paper: |
|
``` |
|
@inproceedings{zhang-abdul-mageed-2022-improving, |
|
title = "Improving Social Meaning Detection with Pragmatic Masking and Surrogate Fine-Tuning", |
|
author = "Zhang, Chiyu and |
|
Abdul-Mageed, Muhammad", |
|
booktitle = "Proceedings of the 12th Workshop on Computational Approaches to Subjectivity, Sentiment {\&} Social Media Analysis", |
|
month = may, |
|
year = "2022", |
|
address = "Dublin, Ireland", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2022.wassa-1.14", |
|
pages = "141--156", |
|
} |
|
``` |