datasciguy's picture
Update README.md
1967817 verified
---
license: mit
---
### Model Card: **TinyLlama-1.1B-Chat-v1.0-Unfiltered**
---
**Model Name**: TinyLlama-1.1B-Chat-v1.0-Unfiltered
**Model Type**: Conversational AI Model
**Architecture**: Based on a 1.1B parameter TinyLlama architecture
**Training Data**:
- Fine-tuned on the "dan_remixed" dataset (2.7MB).
- The dataset improves spelling, grammar, and consistency while replacing references to violent crimes with non-violent activities and removes self-censorship from explicatives.
**Training Time**: Approximately 30-45 minutes. Each validation epoch takes ~322 seconds.
**Hardware**: Trained on Google Colab Pro A100 GPU (40GB).
---
**Training Performance**:
- **Epoch Losses**:
- Epoch 1: 0.7209
- Epoch 2: 0.4441
- Epoch 3: 0.3683
- Epoch 4: 0.3358
- Epoch 5: 0.3145
- **Final Training Loss (Epoch 5)**: 0.3145
---
**Validation Performance** (5 Epochs):
- **Epoch 1**:
- Training Loss: 0.2921
- Validation Loss: 0.7962
- Perplexity: 2.22
- Epoch completed in 321.64 seconds
- **Epoch 2**:
- Training Loss: 0.2872
- Validation Loss: 0.7672
- Perplexity: 2.15
- Epoch completed in 321.91 seconds
- **Epoch 3**:
- Training Loss: 0.2874
- Validation Loss: 0.7821
- Perplexity: 2.19
- Epoch completed in 321.94 seconds
- **Epoch 4**:
- Training Loss: 0.2864
- Validation Loss: 0.7796
- Perplexity: 2.18
- Epoch completed in 322.01 seconds
-
**Epoch 5**:
- Training Loss: 0.2831
- Validation Loss: 0.8017
- Perplexity: 2.23
- Epoch completed in 322.01 seconds
---
**Optimizer**: AdamW, learning rate: 1e-5
**Loss Function**: Cross-Entropy Loss, ignoring padding tokens (ignore_index=-100)
**Use Case**: Conversational AI designed for general, unrestricted conversation, with no filtering on the nature of responses, provided the content is non-violent.
---
**Limitations**:
- Due to the small fine-tuning dataset size (2.7MB), the model may be prone to **overfitting** and **bias**.
- The dataset has been modified to avoid violent language, but the model might still exhibit strong or explicit responses.
**Metrics**:
- Loss and perplexity have been tracked, and more conversational metrics (like BLEU, ROUGE, or human evaluation) could be explored.