language: en
license: mit
tags:
- distilbert
- sentiment-analysis
- multilingual
widgets:
- text: I love this movie!
Model Name: DistilBERT for Sentiment Analysis
Model Description
Overview
This model is a fine-tuned version of distilbert-base-uncased
on a social media dataset for the purpose of sentiment analysis. It can classify text into positive, negative, and neutral sentiments.
Intended Use
This model is intended for sentiment analysis tasks, particularly for analyzing social media texts. It supports multiple languages, making it versatile for international applications.
Model Architecture
This model is based on the DistilBertForSequenceClassification
architecture, a distilled version of BERT that maintains comparable performance on downstream tasks while being more computationally efficient.
Training
Training Data
The model was trained on a dataset consisting of social media posts, labeled for sentiment (positive, negative, neutral). The dataset includes multiple languages, enhancing the model's multilingual capabilities.
Training Procedure
The model was trained using the following parameters:
- Optimizer: AdamW
- Learning Rate: 5e-5
- Batch Size: 32
- Epochs: 30
Training was conducted on Kaggle, utilizing two GPUs for accelerated training.