my_fine_tuning_nllb_200_3B3_peft_en_vi_model_kc

This model is a fine-tuned version of facebook/nllb-200-3.3B on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 5.4198
  • Bleu: 37.4195
  • Gen Len: 30.9488

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 4e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 3
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Bleu Gen Len
5.9131 1.0 33330 5.4224 37.4001 30.8936
5.8974 2.0 66660 5.4207 37.4719 30.8582
5.911 3.0 99990 5.4198 37.4195 30.9488

Framework versions

  • PEFT 0.8.2
  • Transformers 4.37.2
  • Pytorch 2.0.1+cu117
  • Datasets 2.17.0
  • Tokenizers 0.15.2
Downloads last month
0
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for Kudod/my_fine_tuning_nllb_200_3B3_peft_en_vi_model_kc

Adapter
(4)
this model