|
--- |
|
language: ti |
|
tags: |
|
- tigrinya |
|
- gpt2 |
|
- text-generation |
|
--- |
|
|
|
# GPT-2 for Tigrinya Language |
|
|
|
This repository contains a GPT-2 model trained from scratch on Tigrinya text data. The model was trained using the Hugging Face Transformers library. |
|
|
|
## Model Details |
|
|
|
- Model Type: GPT-2 |
|
- Language: Tigrinya |
|
- Vocabulary Size: 16000 |
|
- Maximum Length: 128 |
|
- Model Size: Small |
|
- Number of Parameters: 33,523,200 |
|
|
|
## Training Details |
|
|
|
- Number of Epochs: 12 |
|
- Batch Size: 1 (with gradient accumulation steps of 4) |
|
- Learning Rate: 5e-4 |
|
|
|
## Dataset Statistics |
|
- Total number of words: 16061839 |
|
- Total number of unique words: 458901 |
|
|
|
## Usage |
|
|
|
```python |
|
from transformers import pipeline |
|
|
|
# Load the model |
|
generator = pipeline('text-generation', model='luel/gpt2-tigrinya-small') |
|
|
|
# Generate text |
|
text = generator("ትግራይ", max_length=60) |
|
print(text) |