|
--- |
|
base_model: INSAIT-Institute/BgGPT-7B-Instruct-v0.2 |
|
library_name: peft |
|
license: apache-2.0 |
|
language: |
|
- en |
|
tags: |
|
- propaganda |
|
--- |
|
|
|
# Model Card for identrics/BG_propaganda_detector |
|
|
|
|
|
|
|
## Model Description |
|
|
|
- **Developed by:** [`Identrics`](https://identrics.ai/) |
|
- **Language:** English |
|
- **License:** apache-2.0 |
|
- **Finetuned from model:** [`google-bert/bert-base-cased`](https://huggingface.co/google-bert/bert-base-cased) |
|
- **Context window :** 512 tokens |
|
|
|
## Model Description |
|
|
|
This model consists of a fine-tuned version of google-bert/bert-base-cased for a propaganda detection task. It is effectively a binary classifier, determining whether propaganda is present in the output string. |
|
This model was created by [`Identrics`](https://identrics.ai/), in the scope of the WASPer project. The detailed taxonomy of the full pipeline could be found [here](https://github.com/Identrics/wasper/). |
|
|
|
|
|
## Uses |
|
|
|
Designed as a binary classifier to determine whether a traditional or social media comment contains propaganda. |
|
|
|
### Example |
|
|
|
First install direct dependencies: |
|
``` |
|
pip install transformers torch accelerate |
|
``` |
|
|
|
Then the model can be downloaded and used for inference: |
|
```py |
|
from transformers import AutoModelForSequenceClassification, AutoTokenizer |
|
|
|
model = AutoModelForSequenceClassification.from_pretrained("identrics/EN_propaganda_detector", num_labels=2) |
|
tokenizer = AutoTokenizer.from_pretrained("identrics/EN_propaganda_detector") |
|
|
|
tokens = tokenizer("Our country is the most powerful country in the world!", return_tensors="pt") |
|
output = model(**tokens) |
|
print(output.logits) |
|
``` |
|
|
|
## Training Details |
|
|
|
The training dataset for the model consists of a balanced collection of English examples, including both propaganda and non-propaganda content. These examples were sourced from a variety of traditional media and social media platforms and manually annotated by domain experts. Additionally, the dataset is enriched with AI-generated samples. |
|
|
|
The model achieved an F1 score of **0.807** during evaluation. |
|
|
|
## Compute Infrastructure |
|
The model was fine-tuned using a **GPU / 2xNVIDIA Tesla V100 32GB**. |
|
|
|
## Citation [this section is to be updated soon] |
|
|
|
If you find our work useful, please consider citing WASPer: |
|
|
|
``` |
|
@article{...2024wasper, |
|
title={WASPer: Propaganda Detection in Bulgarian and English}, |
|
author={....}, |
|
journal={arXiv preprint arXiv:...}, |
|
year={2024} |
|
} |
|
``` |
|
|
|
|
|
|