DT12the's picture
Create README.md
8b4a863 verified
|
raw
history blame
1.33 kB
metadata
language: en
license: mit
tags:
  - distilbert
  - sentiment-analysis
  - multilingual
widgets:
  - text: I love this movie!

Model Name: DistilBERT for Sentiment Analysis

Model Description

Overview

This model is a fine-tuned version of distilbert-base-uncased on a social media dataset for the purpose of sentiment analysis. It can classify text into positive, negative, and neutral sentiments.

Intended Use

This model is intended for sentiment analysis tasks, particularly for analyzing social media texts. It supports multiple languages, making it versatile for international applications.

Model Architecture

This model is based on the DistilBertForSequenceClassification architecture, a distilled version of BERT that maintains comparable performance on downstream tasks while being more computationally efficient.

Training

Training Data

The model was trained on a dataset consisting of social media posts, labeled for sentiment (positive, negative, neutral). The dataset includes multiple languages, enhancing the model's multilingual capabilities.

Training Procedure

The model was trained using the following parameters:

  • Optimizer: AdamW
  • Learning Rate: 5e-5
  • Batch Size: 32
  • Epochs: 30

Training was conducted on Kaggle, utilizing two GPUs for accelerated training.