Drazcat commited on
Commit
0a328c7
1 Parent(s): b3bd754

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -3
README.md CHANGED
@@ -7,9 +7,22 @@ tags:
7
  - generated_from_trainer
8
  datasets:
9
  - Drazcat/Cencosud
 
 
10
  model-index:
11
  - name: Whisper Small Es - GoCloud
12
- results: []
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -18,6 +31,9 @@ should probably proofread and complete it, then remove this comment. -->
18
  # Whisper Small Es - GoCloud
19
 
20
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the 30seg dataset.
 
 
 
21
 
22
  ## Model description
23
 
@@ -37,17 +53,23 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
- - train_batch_size: 16
41
  - eval_batch_size: 8
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
- - lr_scheduler_warmup_steps: 500
46
  - training_steps: 200
47
  - mixed_precision_training: Native AMP
48
 
49
  ### Training results
50
 
 
 
 
 
 
 
51
 
52
 
53
  ### Framework versions
 
7
  - generated_from_trainer
8
  datasets:
9
  - Drazcat/Cencosud
10
+ metrics:
11
+ - wer
12
  model-index:
13
  - name: Whisper Small Es - GoCloud
14
+ results:
15
+ - task:
16
+ name: Automatic Speech Recognition
17
+ type: automatic-speech-recognition
18
+ dataset:
19
+ name: 30seg
20
+ type: Drazcat/Cencosud
21
+ args: 'config: es, split: test'
22
+ metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 0.0
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
31
  # Whisper Small Es - GoCloud
32
 
33
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the 30seg dataset.
34
+ It achieves the following results on the evaluation set:
35
+ - Loss: 0.0028
36
+ - Wer: 0.0
37
 
38
  ## Model description
39
 
 
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 1e-05
56
+ - train_batch_size: 8
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
+ - lr_scheduler_warmup_steps: 25
62
  - training_steps: 200
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
66
 
67
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
68
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|
69
+ | 0.2944 | 5.56 | 50 | 0.1392 | 79.6117 |
70
+ | 0.08 | 11.11 | 100 | 0.0569 | 46.0472 |
71
+ | 0.0204 | 16.67 | 150 | 0.0086 | 0.0 |
72
+ | 0.0028 | 22.22 | 200 | 0.0028 | 0.0 |
73
 
74
 
75
  ### Framework versions