zionia commited on
Commit
2ca0402
·
verified ·
1 Parent(s): e0f597c

Add model card

Browse files
Files changed (1) hide show
  1. README.md +27 -63
README.md CHANGED
@@ -1,77 +1,41 @@
1
  ---
2
- library_name: transformers
3
  license: apache-2.0
4
- base_model: openai/whisper-small
5
  tags:
6
- - generated_from_trainer
7
- metrics:
8
- - wer
9
- model-index:
10
- - name: whisper-small-isizulu-noisy
11
- results: []
 
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
-
17
- # whisper-small-isizulu-noisy
18
-
19
- This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 0.9578
22
- - Wer: 77.9164
23
-
24
- ## Model description
25
-
26
- More information needed
27
-
28
- ## Intended uses & limitations
29
-
30
- More information needed
31
-
32
- ## Training and evaluation data
33
 
34
- More information needed
35
 
36
- ## Training procedure
37
 
38
- ### Training hyperparameters
 
 
39
 
40
- The following hyperparameters were used during training:
41
- - learning_rate: 1e-05
42
- - train_batch_size: 32
43
- - eval_batch_size: 16
44
- - seed: 42
45
- - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
- - lr_scheduler_type: linear
47
- - lr_scheduler_warmup_steps: 100
48
- - num_epochs: 15
49
- - mixed_precision_training: Native AMP
50
 
51
- ### Training results
52
 
53
- | Training Loss | Epoch | Step | Validation Loss | Wer |
54
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
55
- | 2.7094 | 1.0 | 33 | 2.2550 | 108.7960 |
56
- | 1.5701 | 2.0 | 66 | 1.4053 | 90.1435 |
57
- | 1.0347 | 3.0 | 99 | 1.0760 | 103.7430 |
58
- | 0.5594 | 4.0 | 132 | 0.9268 | 100.6238 |
59
- | 0.2742 | 5.0 | 165 | 0.9051 | 80.2246 |
60
- | 0.1504 | 6.0 | 198 | 0.8892 | 89.5820 |
61
- | 0.065 | 7.0 | 231 | 0.9149 | 81.1603 |
62
- | 0.0369 | 8.0 | 264 | 0.9126 | 66.1260 |
63
- | 0.0215 | 9.0 | 297 | 0.9266 | 66.9370 |
64
- | 0.0129 | 10.0 | 330 | 0.9321 | 63.5683 |
65
- | 0.0076 | 11.0 | 363 | 0.9407 | 66.1884 |
66
- | 0.006 | 12.0 | 396 | 0.9498 | 64.3793 |
67
- | 0.0051 | 13.0 | 429 | 0.9536 | 77.7293 |
68
- | 0.0046 | 14.0 | 462 | 0.9570 | 77.9164 |
69
- | 0.0044 | 15.0 | 495 | 0.9578 | 77.9164 |
70
 
 
71
 
72
- ### Framework versions
 
73
 
74
- - Transformers 4.57.0
75
- - Pytorch 2.8.0+cu128
76
- - Datasets 4.2.0
77
- - Tokenizers 0.22.1
 
1
  ---
2
+ language: zu
3
  license: apache-2.0
 
4
  tags:
5
+ - whisper
6
+ - automatic-speech-recognition
7
+ - south-african-languages
8
+ datasets:
9
+ - zionia/isizulu-asr-train
10
+ - zionia/isizulu-asr-gaussian-noise
11
+ base_model: openai/whisper-small
12
  ---
13
 
14
+ # Whisper Small - Combined South African Languages
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on multiple South African language datasets.
17
 
18
+ ## Training Data
19
 
20
+ This model was trained on 2 combined datasets:
21
+ - zionia/isizulu-asr-train
22
+ - zionia/isizulu-asr-gaussian-noise
23
 
24
+ Total training samples: 1050
25
+ Total test samples: 263
 
 
 
 
 
 
 
 
26
 
27
+ ## Training Details
28
 
29
+ - **Training epochs:** 15
30
+ - **Learning rate:** 1e-05
31
+ - **Batch size:** 16
32
+ - **Best WER:** 63.57%
 
 
 
 
 
 
 
 
 
 
 
 
 
33
 
34
+ ## Usage
35
 
36
+ ```python
37
+ from transformers import pipeline
38
 
39
+ pipe = pipeline("automatic-speech-recognition", model="zionia/whisper-small-isizulu-noisy")
40
+ result = pipe("path/to/audio.wav")
41
+ ```