shahukareem
commited on
Commit
·
d4582e9
1
Parent(s):
9236d2a
update model card README.md
Browse files
README.md
CHANGED
@@ -1,29 +1,12 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
-
- automatic-speech-recognition
|
5 |
-
- mozilla-foundation/common_voice_8_0
|
6 |
- generated_from_trainer
|
7 |
-
- robust-speech-event
|
8 |
datasets:
|
9 |
- common_voice
|
10 |
model-index:
|
11 |
- name: xls-r-300m-dv
|
12 |
-
results:
|
13 |
-
- task:
|
14 |
-
name: Automatic Speech Recognition
|
15 |
-
type: automatic-speech-recognition
|
16 |
-
dataset:
|
17 |
-
name: Common Voice 8
|
18 |
-
type: mozilla-foundation/common_voice_8_0
|
19 |
-
args: dv
|
20 |
-
metrics:
|
21 |
-
- name: Test WER
|
22 |
-
type: wer
|
23 |
-
value:
|
24 |
-
- name: Test CER
|
25 |
-
type: cer
|
26 |
-
value:
|
27 |
---
|
28 |
|
29 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -31,10 +14,10 @@ should probably proofread and complete it, then remove this comment. -->
|
|
31 |
|
32 |
# xls-r-300m-dv
|
33 |
|
34 |
-
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the
|
35 |
It achieves the following results on the evaluation set:
|
36 |
-
- Loss: 0.
|
37 |
-
- Wer: 0.
|
38 |
|
39 |
## Model description
|
40 |
|
@@ -62,24 +45,88 @@ The following hyperparameters were used during training:
|
|
62 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
63 |
- lr_scheduler_type: linear
|
64 |
- lr_scheduler_warmup_steps: 500
|
65 |
-
- num_epochs:
|
66 |
- mixed_precision_training: Native AMP
|
67 |
|
68 |
### Training results
|
69 |
|
70 |
-
| Training Loss | Epoch | Step
|
71 |
-
|
72 |
-
|
|
73 |
-
| 0.
|
74 |
-
| 0.
|
75 |
-
| 0.
|
76 |
-
| 0.
|
77 |
-
| 0.
|
78 |
-
| 0.
|
79 |
-
| 0.
|
80 |
-
| 0.
|
81 |
-
| 0.
|
82 |
-
| 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
83 |
|
84 |
|
85 |
### Framework versions
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
tags:
|
|
|
|
|
4 |
- generated_from_trainer
|
|
|
5 |
datasets:
|
6 |
- common_voice
|
7 |
model-index:
|
8 |
- name: xls-r-300m-dv
|
9 |
+
results: []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
---
|
11 |
|
12 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
14 |
|
15 |
# xls-r-300m-dv
|
16 |
|
17 |
+
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset.
|
18 |
It achieves the following results on the evaluation set:
|
19 |
+
- Loss: 0.3149
|
20 |
+
- Wer: 0.2947
|
21 |
|
22 |
## Model description
|
23 |
|
|
|
45 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
46 |
- lr_scheduler_type: linear
|
47 |
- lr_scheduler_warmup_steps: 500
|
48 |
+
- num_epochs: 50
|
49 |
- mixed_precision_training: Native AMP
|
50 |
|
51 |
### Training results
|
52 |
|
53 |
+
| Training Loss | Epoch | Step | Validation Loss | Wer |
|
54 |
+
|:-------------:|:-----:|:-----:|:---------------:|:------:|
|
55 |
+
| 3.9617 | 0.66 | 400 | 1.4251 | 0.9768 |
|
56 |
+
| 0.9081 | 1.33 | 800 | 0.6068 | 0.7290 |
|
57 |
+
| 0.6575 | 1.99 | 1200 | 0.4700 | 0.6234 |
|
58 |
+
| 0.548 | 2.65 | 1600 | 0.4158 | 0.5868 |
|
59 |
+
| 0.5031 | 3.32 | 2000 | 0.4067 | 0.5728 |
|
60 |
+
| 0.4792 | 3.98 | 2400 | 0.3965 | 0.5673 |
|
61 |
+
| 0.4344 | 4.64 | 2800 | 0.3862 | 0.5383 |
|
62 |
+
| 0.4237 | 5.31 | 3200 | 0.3794 | 0.5316 |
|
63 |
+
| 0.3984 | 5.97 | 3600 | 0.3395 | 0.5177 |
|
64 |
+
| 0.3788 | 6.63 | 4000 | 0.3528 | 0.5329 |
|
65 |
+
| 0.3685 | 7.3 | 4400 | 0.3404 | 0.5060 |
|
66 |
+
| 0.3535 | 7.96 | 4800 | 0.3425 | 0.5069 |
|
67 |
+
| 0.3391 | 8.62 | 5200 | 0.3576 | 0.5118 |
|
68 |
+
| 0.331 | 9.29 | 5600 | 0.3259 | 0.4783 |
|
69 |
+
| 0.3192 | 9.95 | 6000 | 0.3145 | 0.4794 |
|
70 |
+
| 0.2956 | 10.61 | 6400 | 0.3111 | 0.4650 |
|
71 |
+
| 0.2936 | 11.28 | 6800 | 0.3303 | 0.4741 |
|
72 |
+
| 0.2868 | 11.94 | 7200 | 0.3109 | 0.4597 |
|
73 |
+
| 0.2743 | 12.6 | 7600 | 0.3191 | 0.4557 |
|
74 |
+
| 0.2654 | 13.27 | 8000 | 0.3286 | 0.4570 |
|
75 |
+
| 0.2556 | 13.93 | 8400 | 0.3186 | 0.4468 |
|
76 |
+
| 0.2452 | 14.59 | 8800 | 0.3405 | 0.4582 |
|
77 |
+
| 0.241 | 15.26 | 9200 | 0.3418 | 0.4533 |
|
78 |
+
| 0.2313 | 15.92 | 9600 | 0.3388 | 0.4405 |
|
79 |
+
| 0.2234 | 16.58 | 10000 | 0.3659 | 0.4421 |
|
80 |
+
| 0.2194 | 17.25 | 10400 | 0.3559 | 0.4490 |
|
81 |
+
| 0.2168 | 17.91 | 10800 | 0.3452 | 0.4355 |
|
82 |
+
| 0.2036 | 18.57 | 11200 | 0.3496 | 0.4259 |
|
83 |
+
| 0.2046 | 19.24 | 11600 | 0.3282 | 0.4245 |
|
84 |
+
| 0.1917 | 19.9 | 12000 | 0.3201 | 0.4052 |
|
85 |
+
| 0.1908 | 20.56 | 12400 | 0.3439 | 0.4165 |
|
86 |
+
| 0.1838 | 21.23 | 12800 | 0.3165 | 0.3950 |
|
87 |
+
| 0.1828 | 21.89 | 13200 | 0.3332 | 0.4079 |
|
88 |
+
| 0.1774 | 22.55 | 13600 | 0.3485 | 0.4072 |
|
89 |
+
| 0.1776 | 23.22 | 14000 | 0.3308 | 0.3868 |
|
90 |
+
| 0.1693 | 23.88 | 14400 | 0.3153 | 0.3906 |
|
91 |
+
| 0.1656 | 24.54 | 14800 | 0.3408 | 0.3899 |
|
92 |
+
| 0.1629 | 25.21 | 15200 | 0.3333 | 0.3854 |
|
93 |
+
| 0.164 | 25.87 | 15600 | 0.3172 | 0.3775 |
|
94 |
+
| 0.1505 | 26.53 | 16000 | 0.3105 | 0.3777 |
|
95 |
+
| 0.1524 | 27.2 | 16400 | 0.3136 | 0.3726 |
|
96 |
+
| 0.1482 | 27.86 | 16800 | 0.3110 | 0.3710 |
|
97 |
+
| 0.1423 | 28.52 | 17200 | 0.3299 | 0.3687 |
|
98 |
+
| 0.1419 | 29.19 | 17600 | 0.3271 | 0.3645 |
|
99 |
+
| 0.135 | 29.85 | 18000 | 0.3333 | 0.3638 |
|
100 |
+
| 0.1319 | 30.51 | 18400 | 0.3272 | 0.3640 |
|
101 |
+
| 0.131 | 31.18 | 18800 | 0.3438 | 0.3636 |
|
102 |
+
| 0.1252 | 31.84 | 19200 | 0.3266 | 0.3557 |
|
103 |
+
| 0.1238 | 32.5 | 19600 | 0.3195 | 0.3516 |
|
104 |
+
| 0.1203 | 33.17 | 20000 | 0.3405 | 0.3534 |
|
105 |
+
| 0.1159 | 33.83 | 20400 | 0.3287 | 0.3509 |
|
106 |
+
| 0.115 | 34.49 | 20800 | 0.3474 | 0.3433 |
|
107 |
+
| 0.108 | 35.16 | 21200 | 0.3245 | 0.3381 |
|
108 |
+
| 0.1091 | 35.82 | 21600 | 0.3185 | 0.3448 |
|
109 |
+
| 0.1043 | 36.48 | 22000 | 0.3309 | 0.3363 |
|
110 |
+
| 0.1034 | 37.15 | 22400 | 0.3288 | 0.3349 |
|
111 |
+
| 0.1015 | 37.81 | 22800 | 0.3222 | 0.3284 |
|
112 |
+
| 0.0953 | 38.47 | 23200 | 0.3272 | 0.3315 |
|
113 |
+
| 0.0966 | 39.14 | 23600 | 0.3196 | 0.3239 |
|
114 |
+
| 0.0938 | 39.8 | 24000 | 0.3199 | 0.3280 |
|
115 |
+
| 0.0905 | 40.46 | 24400 | 0.3193 | 0.3166 |
|
116 |
+
| 0.0893 | 41.13 | 24800 | 0.3224 | 0.3222 |
|
117 |
+
| 0.0858 | 41.79 | 25200 | 0.3216 | 0.3142 |
|
118 |
+
| 0.0839 | 42.45 | 25600 | 0.3241 | 0.3135 |
|
119 |
+
| 0.0819 | 43.12 | 26000 | 0.3260 | 0.3071 |
|
120 |
+
| 0.0782 | 43.78 | 26400 | 0.3202 | 0.3075 |
|
121 |
+
| 0.0775 | 44.44 | 26800 | 0.3140 | 0.3067 |
|
122 |
+
| 0.0751 | 45.11 | 27200 | 0.3118 | 0.3020 |
|
123 |
+
| 0.0736 | 45.77 | 27600 | 0.3155 | 0.2976 |
|
124 |
+
| 0.071 | 46.43 | 28000 | 0.3105 | 0.2998 |
|
125 |
+
| 0.0715 | 47.1 | 28400 | 0.3065 | 0.2993 |
|
126 |
+
| 0.0668 | 47.76 | 28800 | 0.3161 | 0.2972 |
|
127 |
+
| 0.0698 | 48.42 | 29200 | 0.3137 | 0.2967 |
|
128 |
+
| 0.0681 | 49.09 | 29600 | 0.3130 | 0.2971 |
|
129 |
+
| 0.0651 | 49.75 | 30000 | 0.3149 | 0.2947 |
|
130 |
|
131 |
|
132 |
### Framework versions
|