WRAP / README.md
TomatenMarc's picture
Upload README.md
a0c027d
|
raw
history blame
7.9 kB
metadata
license: cc-by-nc-4.0
language:
  - en
metrics:
  - f1
pipeline_tag: text-classification
tags:
  - transformers
  - argument-mining
  - opinion-mining
  - information-extraction
  - inference-extraction
  - Twitter
widget:
  - text: 'Men shouldn’t be making laws about women’s bodies #abortion #Texas'
    example_title: Statement
  - text: >-
      ’Bitter truth’: EU chief pours cold water on idea of Brits keeping EU
      citizenship after #Brexit HTTPURL via @USER
    example_title: Notification
  - text: >-
      Opinion: As the draconian (and then some) abortion law takes effect in
      #Texas, this is not an idle question for millions of Americans. A slippery
      slope towards more like-minded Republican state legislatures to try to
      follow suit. #abortion #F24 HTTPURL
    example_title: Reason
  - text: '@USER Blah blah blah blah blah blah'
    example_title: None
  - text: republican men and karens make me sick
    example_title: Unlabeled 1
  - text: No empire lives forever! Historical fact! GodWins! 🙏💪🇺🇲
    example_title: Unlabeled 2
  - text: >-
      Further author information regarding registration and visa support letters
      will be sent to the authors soon. #CIKM2023
    example_title: Unlabeled 3
  - text: Ummmmmm
    example_title: Unlabeled 4
  - text: >-
      whoever says that The Last Jedi is a good movie is lying or trolling
      everyone
    example_title: Unlabeled 5
  - text: >-
      I don’t think people realize how big this story is GBI Strategies, the
      group paid $11M+ by Biden PACs to harvest fraudulent voter registrations
      in *20 states*, may be the root source of Democrat election rigging @USER
      may have just exposed their entire fraud machine HTTPURL
    example_tite: Unlabeled 6

WRAP -- A TACO-based Classifier For Inference and Information-Driven Argument Mining on Twitter

Introducing WRAP, an advanced classification model built upon AutoModelForSequenceClassification, designed to identify tweets belonging to four distinct classes: Reason, Statement, Notification, and None of the TACO dataset. Designed specifically for extracting information and inferences from Twitter data, this specialized classification model utilizes WRAPresentations, from which WRAP acquires its name. WRAPresentations is an advancement of the BERTweet-base architecture, whose embeddings were extended on augmented tweets using contrastive learning for better encoding inference and information in tweets.

Class Semantics

The TACO framework revolves around the two key elements of an argument, as defined by the Cambridge Dictionary. It encodes inference as a guess that you make or an opinion that you form based on the information that you have, and it also leverages the definition of information as facts or details about a person, company, product, etc..

Taken together, WRAP can identify specific classes of tweets, where inferences and information can be aggregated in relation to these distinct classes containing these components:

  • Statement, which refers to unique cases where only the inference is presented as something that someone says or writes officially, or an action done to express an opinion.
  • Reason, which represents a full argument where the inference is based on direct information mentioned in the tweet, such as a source-reference or quotation, and thus reveals the author’s motivation to try to understand and to make judgments based on practical facts.
  • Notification, which refers to a tweet that limits itself to providing information, such as media channels promoting their latest articles.
  • None, a tweet that provides neither inference nor information.

In its entirety, WRAP can classify the following hierarchy for tweets:

Component Space

Usage

Using this model becomes easy when you have transformers installed:

pip install - U transformers

Then you can use the model to generate tweet classifications like this:

from transformers import pipeline

pipe = pipeline("text-classification", model="TomatenMarc/WRAP")
prediction = pipe("Huggingface is awesome")

print(prediction)
Notice: The tweets need to undergo preprocessing before classification.

Training

The final model underwent training using the entire shuffled ground truth dataset known as TACO, encompassing a total of 1734 tweets. This dataset showcases the distribution of topics as: #abortion (25.9%), #brexit (29.0%), #got (11.0%), #lotrrop (12.1%), #squidgame (12.7%), and #twittertakeover (9.3%). For training, we utilized SimpleTransformers.

Additionally, the category and class distribution of the dataset TACO is as follows:

Inference No-Inference
865 (49.88%) 869 (50.12%)
Information No-Information
1081 (62.34%) 653 (37.66%)
Reason Statement Notification None
581 (33.50%) 284 (16.38%) 500 (28.84%) 369 (21.28%)

Notice: Our training involved WRAP to forecast class predictions, where the categories (information/inference) represent class aggregations based on the inference or information component.

Dataloader

"data_loader": {
    "type": "torch.utils.data.dataloader.DataLoader",
    "args": {
        "batch_size": 8,
        "sampler": "torch.utils.data.sampler.RandomSampler"
    }
}

Parameters of the fit()-Method:

{
    "epochs": 5,
    "max_grad_norm": 1,
    "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
    "optimizer_params": {
        "lr": 4e-05
    },
    "scheduler": "WarmupLinear",
    "warmup_steps": 66,
    "weight_decay": 0.06
}

Evaluation

We applied a 6-fold (In-Topic) cross-validation method to demonstrate WRAP's optimal performance. This involved using the same dataset and parameters described in the Training section, where we trained on k-1 splits and made predictions using the kth split.

Additionally, we assessed its ability to generalize across the 6 topics (Cross-Topic) of TACO. Each of the k topics was utilized for testing, while the remaining k-1 topics were used for training purposes.

In total, the WRAP classifier performs as follows:

Content Management

Macro-F1 Inference Information Multiclass
In-Topic 87.71% 85.34% 75.80%
Cross-Topic 86.71% 84.58% 73.92%

Classification

Micro-F1 Reason Statement Notification None
In-Topic 77.82% 61.10% 80.56% 83.71%
Cross-Topic 76.52% 58.99% 78.43% 81.73%

Environmental Impact

Licensing

WRAP © 2023 is licensed under CC BY-NC-SA 4.0