|
--- |
|
language: "en" |
|
license: "mit" |
|
tags: |
|
- distilbert |
|
- sentiment-analysis |
|
- multilingual |
|
widgets: |
|
- text: "I love this movie!" |
|
--- |
|
|
|
# Model Name: DistilBERT for Sentiment Analysis |
|
|
|
## Model Description |
|
|
|
### Overview |
|
|
|
This model is a fine-tuned version of `distilbert-base-uncased` on a social media dataset for the purpose of sentiment analysis. It can classify text into non-negative and negative sentiments. |
|
|
|
### Intended Use |
|
|
|
This model is intended for sentiment analysis tasks, particularly for analyzing social media texts. |
|
|
|
### Model Architecture |
|
|
|
This model is based on the `DistilBertForSequenceClassification` architecture, a distilled version of BERT that maintains comparable performance on downstream tasks while being more computationally efficient. |
|
|
|
## Training |
|
|
|
### Training Data |
|
|
|
The model was trained on a dataset consisting of social media posts, surveys and interviews, labeled for sentiment (non-negative and negative). The dataset includes texts from a variety of sources and demographics. |
|
|
|
### Training Procedure |
|
|
|
The model was trained using the following parameters: |
|
- Optimizer: AdamW |
|
- Learning Rate: 5e-5 |
|
- Batch Size: 32 |
|
- Epochs: 30 |
|
|
|
Training was conducted on Kaggle, utilizing two GPUs for accelerated training. |