|
--- |
|
library_name: transformers |
|
language: |
|
- nl |
|
license: apache-2.0 |
|
base_model: openai/whisper-large-v2 |
|
tags: |
|
- generated_from_trainer |
|
metrics: |
|
- wer |
|
model-index: |
|
- name: Whisper Large V2 |
|
results: [] |
|
--- |
|
|
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You |
|
should probably proofread and complete it, then remove this comment. --> |
|
|
|
# Whisper Large V2 |
|
|
|
This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the None dataset. |
|
It achieves the following results on the evaluation set: |
|
- Loss: 0.4599 |
|
- Wer: 24.2092 |
|
|
|
## Model description |
|
|
|
More information needed |
|
|
|
## Intended uses & limitations |
|
|
|
More information needed |
|
|
|
## Training and evaluation data |
|
|
|
More information needed |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 3e-05 |
|
- train_batch_size: 12 |
|
- eval_batch_size: 8 |
|
- seed: 42 |
|
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 |
|
- lr_scheduler_type: linear |
|
- lr_scheduler_warmup_steps: 20 |
|
- num_epochs: 5 |
|
|
|
### Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | Wer | |
|
|:-------------:|:------:|:----:|:---------------:|:-------:| |
|
| 0.8662 | 0.0833 | 15 | 0.5819 | 33.0573 | |
|
| 0.5255 | 0.1667 | 30 | 0.4760 | 37.3151 | |
|
| 0.4103 | 0.25 | 45 | 0.4493 | 28.9104 | |
|
| 0.4344 | 0.3333 | 60 | 0.4362 | 25.2990 | |
|
| 0.3882 | 0.4167 | 75 | 0.4279 | 37.7798 | |
|
| 0.402 | 0.5 | 90 | 0.4190 | 40.5043 | |
|
| 0.4105 | 0.5833 | 105 | 0.4136 | 42.3254 | |
|
| 0.4126 | 0.6667 | 120 | 0.3959 | 24.2446 | |
|
| 0.3672 | 0.75 | 135 | 0.3956 | 36.5367 | |
|
| 0.4401 | 0.8333 | 150 | 0.3805 | 24.8744 | |
|
| 0.3698 | 0.9167 | 165 | 0.3834 | 44.2526 | |
|
| 0.3728 | 1.0 | 180 | 0.3742 | 42.0447 | |
|
| 0.1642 | 1.0833 | 195 | 0.3942 | 27.6555 | |
|
| 0.1904 | 1.1667 | 210 | 0.3793 | 25.7967 | |
|
| 0.1801 | 1.25 | 225 | 0.3859 | 23.6879 | |
|
| 0.1693 | 1.3333 | 240 | 0.3934 | 25.1928 | |
|
| 0.1839 | 1.4167 | 255 | 0.3853 | 29.6629 | |
|
| 0.19 | 1.5 | 270 | 0.3763 | 27.3489 | |
|
| 0.1977 | 1.5833 | 285 | 0.3764 | 21.0436 | |
|
| 0.1922 | 1.6667 | 300 | 0.3719 | 30.1040 | |
|
| 0.185 | 1.75 | 315 | 0.3716 | 25.9736 | |
|
| 0.1873 | 1.8333 | 330 | 0.3671 | 22.8127 | |
|
| 0.1802 | 1.9167 | 345 | 0.3621 | 21.2582 | |
|
| 0.1931 | 2.0 | 360 | 0.3662 | 24.4262 | |
|
| 0.0848 | 2.0833 | 375 | 0.3989 | 34.8949 | |
|
| 0.0823 | 2.1667 | 390 | 0.3888 | 23.3718 | |
|
| 0.0817 | 2.25 | 405 | 0.3914 | 22.8057 | |
|
| 0.0952 | 2.3333 | 420 | 0.3784 | 23.8530 | |
|
| 0.0961 | 2.4167 | 435 | 0.3917 | 33.5315 | |
|
| 0.0954 | 2.5 | 450 | 0.3822 | 20.7959 | |
|
| 0.0909 | 2.5833 | 465 | 0.3877 | 22.4282 | |
|
| 0.084 | 2.6667 | 480 | 0.3878 | 26.7025 | |
|
| 0.0769 | 2.75 | 495 | 0.3890 | 21.8597 | |
|
| 0.0879 | 2.8333 | 510 | 0.3899 | 24.0724 | |
|
| 0.0835 | 2.9167 | 525 | 0.3788 | 20.9327 | |
|
| 0.0845 | 3.0 | 540 | 0.3807 | 26.7379 | |
|
| 0.0383 | 3.0833 | 555 | 0.4227 | 24.4333 | |
|
| 0.0408 | 3.1667 | 570 | 0.4173 | 31.6868 | |
|
| 0.0393 | 3.25 | 585 | 0.4202 | 21.5413 | |
|
| 0.035 | 3.3333 | 600 | 0.4141 | 23.4355 | |
|
| 0.034 | 3.4167 | 615 | 0.4193 | 24.1927 | |
|
| 0.0383 | 3.5 | 630 | 0.4160 | 27.6319 | |
|
| 0.0295 | 3.5833 | 645 | 0.4243 | 26.9644 | |
|
| 0.0323 | 3.6667 | 660 | 0.4201 | 25.1221 | |
|
| 0.0337 | 3.75 | 675 | 0.4195 | 26.5445 | |
|
| 0.0328 | 3.8333 | 690 | 0.4229 | 23.7775 | |
|
| 0.0344 | 3.9167 | 705 | 0.4213 | 23.9025 | |
|
| 0.0257 | 4.0 | 720 | 0.4209 | 23.6643 | |
|
| 0.0134 | 4.0833 | 735 | 0.4392 | 22.7372 | |
|
| 0.0113 | 4.1667 | 750 | 0.4556 | 21.2511 | |
|
| 0.0122 | 4.25 | 765 | 0.4596 | 21.6899 | |
|
| 0.0117 | 4.3333 | 780 | 0.4652 | 21.7890 | |
|
| 0.0111 | 4.4167 | 795 | 0.4637 | 21.6946 | |
|
| 0.0115 | 4.5 | 810 | 0.4627 | 23.1571 | |
|
| 0.0127 | 4.5833 | 825 | 0.4567 | 24.0040 | |
|
| 0.0108 | 4.6667 | 840 | 0.4592 | 23.0415 | |
|
| 0.0107 | 4.75 | 855 | 0.4610 | 23.4661 | |
|
| 0.0094 | 4.8333 | 870 | 0.4602 | 25.0112 | |
|
| 0.0104 | 4.9167 | 885 | 0.4599 | 24.6621 | |
|
| 0.0125 | 5.0 | 900 | 0.4599 | 24.2092 | |
|
|
|
|
|
### Framework versions |
|
|
|
- Transformers 4.45.0.dev0 |
|
- Pytorch 2.1.0+cu121 |
|
- Datasets 2.20.0 |
|
- Tokenizers 0.19.1 |
|
|