Edit model card

whisper2

This model is a fine-tuned version of openai/whisper-tiny.en on the tiny dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5233
  • Wer: 31.1083

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 64
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 500
  • training_steps: 500

Training results

Training Loss Epoch Step Validation Loss Wer
3.9553 0.1408 10 3.9646 74.8741
3.9548 0.2817 20 3.8794 77.6763
3.8127 0.4225 30 3.7405 76.4169
3.6178 0.5634 40 3.5547 75.3149
3.3992 0.7042 50 3.3235 70.2771
3.1416 0.8451 60 3.0402 67.8526
2.8052 0.9859 70 2.6852 65.9635
2.3513 1.1268 80 2.2235 68.3249
1.893 1.2676 90 1.6708 63.8224
1.2871 1.4085 100 1.1645 63.2557
0.9146 1.5493 110 0.8785 56.8325
0.8044 1.6901 120 0.7907 46.9773
0.6634 1.8310 130 0.7425 47.4811
0.6722 1.9718 140 0.7100 45.9068
0.6823 2.1127 150 0.6854 42.4118
0.5802 2.2535 160 0.6659 40.4282
0.6084 2.3944 170 0.6503 40.8375
0.6038 2.5352 180 0.6346 41.4987
0.5095 2.6761 190 0.6247 42.0340
0.5251 2.8169 200 0.6155 39.3577
0.5699 2.9577 210 0.6046 38.3501
0.4839 3.0986 220 0.5945 37.2796
0.4843 3.2394 230 0.5861 48.3942
0.4538 3.3803 240 0.5794 34.6662
0.4741 3.5211 250 0.5737 33.8161
0.4542 3.6620 260 0.5663 41.9710
0.4163 3.8028 270 0.5623 46.0957
0.3496 3.9437 280 0.5605 42.2544
0.3835 4.0845 290 0.5557 41.6562
0.3462 4.2254 300 0.5507 36.3980
0.3133 4.3662 310 0.5452 42.5693
0.3638 4.5070 320 0.5435 35.9572
0.3826 4.6479 330 0.5396 31.9584
0.3581 4.7887 340 0.5361 33.7846
0.3127 4.9296 350 0.5339 37.3426
0.2988 5.0704 360 0.5348 38.7280
0.2807 5.2113 370 0.5344 35.5164
0.2612 5.3521 380 0.5305 34.6662
0.2762 5.4930 390 0.5306 32.2733
0.299 5.6338 400 0.5267 36.8703
0.2718 5.7746 410 0.5232 41.6877
0.2618 5.9155 420 0.5208 34.0995
0.2121 6.0563 430 0.5220 28.0542
0.1929 6.1972 440 0.5256 35.7997
0.2504 6.3380 450 0.5296 32.8715
0.2064 6.4789 460 0.5265 35.3904
0.2044 6.6197 470 0.5267 38.3186
0.1844 6.7606 480 0.5231 35.1071
0.1867 6.9014 490 0.5235 31.5806
0.1562 7.0423 500 0.5233 31.1083

Framework versions

  • Transformers 4.40.1
  • Pytorch 2.2.1+cu121
  • Datasets 2.19.1.dev0
  • Tokenizers 0.19.1
Downloads last month
7
Safetensors
Model size
37.8M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for khaingsmon/whisper2

Finetuned
(60)
this model