YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

emotion-classification-model

This model is a fine-tuned version of distilbert-base-uncased on the dair-ai/emotion dataset. It is designed to classify text into the following emotional categories: sadness (0), joy (1), love (2), anger (3), fear (4), surprise (5).

It achieves the following results:

  • Validation Accuracy: 94.25%
  • Test Accuracy: 93.2%

Model Description

This model uses the DistilBERT architecture, which is a lighter and faster variant of BERT. It has been fine-tuned specifically for emotion classification, making it suitable for tasks such as sentiment analysis, customer feedback analysis, and user emotion detection.

Key Features

  • Efficient and lightweight for deployment.
  • High accuracy for emotion detection tasks.
  • Pretrained on a diverse dataset and fine-tuned for high specificity to emotions.

Intended Uses & Limitations

Intended Uses

  • Emotion analysis in text data.
  • Sentiment detection in customer reviews, tweets, or user feedback.
  • Psychological or behavioral studies to analyze emotional tone in communications.

Limitations

  • May not generalize well to datasets with highly domain-specific language.
  • Might struggle with sarcasm, irony, or other nuanced forms of language.
  • The model is English-specific and may not perform well on non-English text.

Training and Evaluation Data

Training Dataset

  • Dataset: dair-ai/emotion
  • Training Set Size: 16,000 examples
  • Dataset Description: The dataset contains English sentences labeled with six emotional categories: sadness (0), joy (1), love (2), anger (3), fear (4), surprise (5)

Results

  • Training Time: ~226 seconds
  • Training Loss: 0.0520
  • Validation Accuracy: 94.25%
  • Test Accuracy: 93.2%

Training Procedure

Hyperparameters

  • Learning Rate: 5e-05
  • Batch Size: 16 (train and evaluation)
  • Epochs: 3
  • Seed: 42
  • Optimizer: AdamW (betas=(0.9,0.999), epsilon=1e-08)
  • Learning Rate Scheduler: Linear
  • Mixed Precision Training: Native AMP

Training and Validation Results

Epoch Training Loss Validation Loss Validation Accuracy
1 0.5383 0.1845 92.90%
2 0.2254 0.1589 93.55%
3 0.0520 0.1485 94.25%

Final Evaluation

  • Validation Loss: 0.1485
  • Validation Accuracy: 94.25%
  • Test Loss: 0.1758
  • Test Accuracy: 93.2%

Performance Metrics

  • Training Speed: ~212 samples/second
  • Evaluation Speed: ~1144 samples/second

Usage Example

from transformers import pipeline

# Load the fine-tuned model
classifier = pipeline("text-classification", model="Panda0116/emotion-classification-model")

# Example usage
text = "I am so happy to see you!"
emotion = classifier(text)
print(emotion)
Downloads last month
6
Safetensors
Model size
67M params
Tensor type
F32
·
Inference API
Unable to determine this model's library. Check the docs .