BIAS-CONLL / README.md
AnanyaRaval's picture
Fix formatting for showing code snippet. (#2)
ac4428f verified
# Hugging Face with Bias Data in CoNLL Format
## Introduction
This README provides guidance on how to use the Hugging Face platform with bias-tagged datasets in the CoNLL format.
Such datasets are essential for studying and mitigating bias in AI models.
This dataset is curated by **Shaina Raza**.
The methods and formatting discussed here are based on the seminal work "Nbias: A natural language processing framework for BIAS identification in text" by Raza et al. (2024) (see citation below).
## Prerequisites
- Install the Hugging Face `transformers` and `datasets` libraries:
```bash
pip install transformers datasets
```
## Data Format
Bias data in CoNLL format can be structured similarly to standard CoNLL, but with labels indicating bias instead of named entities:
```
The O
book O
written B-BIAS
by I-BIAS
egoist I-BIAS
women I-BIAS
is O
good O
. O
```
Here, `B-` prefixes indicate the beginning of a biased term,`I-` indicates inside biased terms, and `O` stands for outside any biased entity.
## Steps to Use with Hugging Face
1. **Loading Bias-tagged CoNLL Data with Hugging Face**
- If your bias-tagged dataset in CoNLL format is publicly available on the Hugging Face `datasets` hub, use:
```python
from datasets import load_dataset
dataset = load_dataset("newsmediabias/BIAS-CONLL")
```
- For custom datasets, ensure they are formatted correctly and use a local path to load them.
If the dataset is gated/private, make sure you have run `huggingface-cli login`
2. **Preprocessing the Data**
- Tokenization:
```python
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("YOUR_PREFERRED_MODEL_CHECKPOINT")
tokenized_input = tokenizer(dataset['train']['tokens'])
```
3. **Training a Model on Bias-tagged CoNLL Data**
- Depending on your task, you may fine-tune a model on the bias data using Hugging Face's `Trainer` class or native PyTorch/TensorFlow code.
4. **Evaluation**
- After training, evaluate the model's ability to recognize and possibly mitigate bias.
- This might involve measuring the model's precision, recall, and F1 score on recognizing bias in text.
5. **Deployment**
- Once satisfied with the model's performance, deploy it for real-world applications, always being mindful of its limitations and potential implications.
Please cite us if you use it.
**Reference to cite us**
```
@article{raza2024nbias,
title={Nbias: A natural language processing framework for BIAS identification in text},
author={Raza, Shaina and Garg, Muskan and Reji, Deepak John and Bashir, Syed Raza and Ding, Chen},
journal={Expert Systems with Applications},
volume={237},
pages={121542},
year={2024},
publisher={Elsevier}
}
```