res_nw_yem_1.5
This model is a fine-tuned version of riotu-lab/ArabianGPT-1.5B on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 0.4809
- Bleu: 0.3835
- Rouge1: 0.5750
- Rouge2: 0.2922
- Rougel: 0.5730
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 20.0
Training results
Training Loss | Epoch | Step | Validation Loss | Bleu | Rouge1 | Rouge2 | Rougel |
---|---|---|---|---|---|---|---|
9.5242 | 1.0 | 153 | 3.5205 | 0.0 | 0.1746 | 0.0016 | 0.1744 |
1.708 | 2.0 | 306 | 0.6233 | 0.3359 | 0.4685 | 0.1802 | 0.4680 |
0.544 | 3.0 | 459 | 0.5348 | 0.3493 | 0.5121 | 0.2174 | 0.5119 |
0.4496 | 4.0 | 612 | 0.5109 | 0.3575 | 0.5235 | 0.2210 | 0.5210 |
0.3818 | 5.0 | 765 | 0.4967 | 0.3741 | 0.5472 | 0.2526 | 0.5450 |
0.3292 | 6.0 | 918 | 0.4865 | 0.3766 | 0.5491 | 0.2610 | 0.5472 |
0.2841 | 7.0 | 1071 | 0.4831 | 0.3775 | 0.5573 | 0.2705 | 0.5556 |
0.2454 | 8.0 | 1224 | 0.4809 | 0.3835 | 0.5750 | 0.2922 | 0.5730 |
0.214 | 9.0 | 1377 | 0.4823 | 0.3843 | 0.5801 | 0.3014 | 0.5780 |
0.1875 | 10.0 | 1530 | 0.4839 | 0.3862 | 0.5860 | 0.3070 | 0.5841 |
0.1662 | 11.0 | 1683 | 0.4854 | 0.3874 | 0.5880 | 0.3116 | 0.5860 |
0.1485 | 12.0 | 1836 | 0.4899 | 0.3920 | 0.5932 | 0.3196 | 0.5916 |
0.1349 | 13.0 | 1989 | 0.4938 | 0.3963 | 0.5931 | 0.3214 | 0.5912 |
Framework versions
- Transformers 4.45.0.dev0
- Pytorch 2.3.1+cu121
- Datasets 2.19.2
- Tokenizers 0.19.1
- Downloads last month
- 1
Model tree for nlparabic/res_nw_yem_1.5
Base model
riotu-lab/ArabianGPT-1.5B