scream_sextusdecimus_virtual_tsfix_medium_1e5
This model is a fine-tuned version of openai/whisper-medium on the NbAiLab/ncc_speech dataset. It achieves the following results on the evaluation set:
- step: 19999
- eval_loss: 1.6336
- train_loss: 0.6795
- eval_wer: 7.9120
- eval_cer: 3.4474
- eval_exact_wer: 7.9120
- eval_exact_cer: 3.4474
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- lr_scheduler_type: linear
- per_device_train_batch_size: 16
- total_train_batch_size_per_node: 64
- total_train_batch_size: 512
- total_optimization_steps: 20,000
- starting_optimization_step: None
- finishing_optimization_step: 20,000
- num_train_dataset_workers: 32
- num_hosts: 8
- total_num_training_examples: 10,240,000
- steps_per_epoch: To be computed after first epoch
- num_beams: None
- dropout: True
- bpe_dropout_probability: 0.1
- activation_dropout_probability: 0.1
Training results
step | eval_loss | train_loss | eval_wer | eval_cer | eval_exact_wer | eval_exact_cer |
---|---|---|---|---|---|---|
0 | 5.5890 | 2.8362 | 17.4598 | 5.3906 | 17.4598 | 5.3906 |
1000 | 5.2798 | 1.0896 | 12.4926 | 3.8321 | 12.4926 | 3.8321 |
2000 | 5.2432 | 0.9018 | 11.0351 | 3.9899 | 11.0351 | 3.9899 |
3000 | 4.1719 | 0.8159 | 9.8453 | 3.8173 | 9.8453 | 3.8173 |
4000 | 3.0758 | 0.7799 | 9.6371 | 3.8716 | 9.6371 | 3.8716 |
5000 | 2.2223 | 0.7803 | 9.7264 | 3.9110 | 9.7264 | 3.9110 |
6000 | 2.0574 | 0.7206 | 9.5181 | 3.8864 | 9.5181 | 3.8864 |
7000 | 1.7271 | 0.7088 | 8.7745 | 3.7039 | 8.7745 | 3.7039 |
8000 | 1.5868 | 0.7528 | 8.2391 | 3.5362 | 8.2391 | 3.5362 |
9000 | 1.5781 | 0.6747 | 8.2094 | 3.5313 | 8.2094 | 3.5313 |
10000 | 1.6658 | 0.6830 | 8.1499 | 3.4277 | 8.1499 | 3.4277 |
11000 | 1.5514 | 0.7141 | 8.6853 | 3.8814 | 8.6853 | 3.8814 |
12000 | 1.8042 | 0.6941 | 8.5366 | 3.6792 | 8.5366 | 3.6792 |
13000 | 1.7561 | 0.6732 | 8.6258 | 3.8666 | 8.6258 | 3.8666 |
14000 | 1.7517 | 0.7050 | 8.2094 | 3.5066 | 8.2094 | 3.5066 |
15000 | 1.7413 | 0.7191 | 7.8822 | 3.3389 | 7.8822 | 3.3389 |
16000 | 1.7014 | 0.6850 | 8.0309 | 3.4178 | 8.0309 | 3.4178 |
17000 | 1.7205 | 0.6937 | 7.8822 | 3.4524 | 7.8822 | 3.4524 |
18000 | 1.5928 | 0.7014 | 7.8227 | 3.4425 | 7.8227 | 3.4425 |
19000 | 1.5883 | 0.7102 | 7.9417 | 3.4573 | 7.9417 | 3.4573 |
19999 | 1.6336 | 0.6795 | 7.9120 | 3.4474 | 7.9120 | 3.4474 |
Framework versions
- Transformers 4.30.0.dev0
- Datasets 2.12.1.dev0
- Tokenizers 0.13.3
- Downloads last month
- 12
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.