DT12the's picture
Create README.md
8b4a863 verified
|
raw
history blame
1.33 kB
---
language: "en"
license: "mit"
tags:
- distilbert
- sentiment-analysis
- multilingual
widgets:
- text: "I love this movie!"
---
# Model Name: DistilBERT for Sentiment Analysis
## Model Description
### Overview
This model is a fine-tuned version of `distilbert-base-uncased` on a social media dataset for the purpose of sentiment analysis. It can classify text into positive, negative, and neutral sentiments.
### Intended Use
This model is intended for sentiment analysis tasks, particularly for analyzing social media texts. It supports multiple languages, making it versatile for international applications.
### Model Architecture
This model is based on the `DistilBertForSequenceClassification` architecture, a distilled version of BERT that maintains comparable performance on downstream tasks while being more computationally efficient.
## Training
### Training Data
The model was trained on a dataset consisting of social media posts, labeled for sentiment (positive, negative, neutral). The dataset includes multiple languages, enhancing the model's multilingual capabilities.
### Training Procedure
The model was trained using the following parameters:
- Optimizer: AdamW
- Learning Rate: 5e-5
- Batch Size: 32
- Epochs: 30
Training was conducted on Kaggle, utilizing two GPUs for accelerated training.