csikasote commited on
Commit
becbb57
1 Parent(s): 07d3820

End of training

Browse files
README.md CHANGED
@@ -3,6 +3,9 @@ library_name: transformers
3
  license: cc-by-nc-4.0
4
  base_model: facebook/mms-1b-all
5
  tags:
 
 
 
6
  - generated_from_trainer
7
  metrics:
8
  - wer
@@ -16,10 +19,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # mms-1b-bemgen-combined-model
18
 
19
- This model is a fine-tuned version of [facebook/mms-1b-all](https://huggingface.co/facebook/mms-1b-all) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 0.2591
22
- - Wer: 0.4135
23
 
24
  ## Model description
25
 
 
3
  license: cc-by-nc-4.0
4
  base_model: facebook/mms-1b-all
5
  tags:
6
+ - automatic-speech-recognition
7
+ - genbed
8
+ - mms
9
  - generated_from_trainer
10
  metrics:
11
  - wer
 
19
 
20
  # mms-1b-bemgen-combined-model
21
 
22
+ This model is a fine-tuned version of [facebook/mms-1b-all](https://huggingface.co/facebook/mms-1b-all) on the GENBED - BEM dataset.
23
  It achieves the following results on the evaluation set:
24
  - Loss: 0.2591
25
+ - Wer: 0.4134
26
 
27
  ## Model description
28
 
adapter.bem.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9f76e1bf613746e8d243ec0f994efd46275a53aa7358fcfabc6a9855bb3093
3
+ size 8798532
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.2893243940175347,
3
+ "eval_loss": 0.2591020166873932,
4
+ "eval_runtime": 57.7458,
5
+ "eval_samples": 976,
6
+ "eval_samples_per_second": 16.902,
7
+ "eval_steps_per_second": 2.113,
8
+ "eval_wer": 0.413365259033522,
9
+ "total_flos": 5.958212987780215e+18,
10
+ "train_loss": 0.6475233551025391,
11
+ "train_runtime": 3633.5111,
12
+ "train_samples": 7756,
13
+ "train_samples_per_second": 64.037,
14
+ "train_steps_per_second": 16.009
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.2893243940175347,
3
+ "eval_loss": 0.2591020166873932,
4
+ "eval_runtime": 57.7458,
5
+ "eval_samples": 976,
6
+ "eval_samples_per_second": 16.902,
7
+ "eval_steps_per_second": 2.113,
8
+ "eval_wer": 0.413365259033522
9
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.2893243940175347,
3
+ "total_flos": 5.958212987780215e+18,
4
+ "train_loss": 0.6475233551025391,
5
+ "train_runtime": 3633.5111,
6
+ "train_samples": 7756,
7
+ "train_samples_per_second": 64.037,
8
+ "train_steps_per_second": 16.009
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,451 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.258684903383255,
3
+ "best_model_checkpoint": "/scratch/skscla001/speech/results/mms-1b-bemgen-combined-model/checkpoint-2200",
4
+ "epoch": 1.2893243940175347,
5
+ "eval_steps": 100,
6
+ "global_step": 2500,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.05157297576070139,
13
+ "grad_norm": 3.681674003601074,
14
+ "learning_rate": 0.00028799999999999995,
15
+ "loss": 6.7553,
16
+ "step": 100
17
+ },
18
+ {
19
+ "epoch": 0.05157297576070139,
20
+ "eval_loss": 0.8774181604385376,
21
+ "eval_runtime": 57.8286,
22
+ "eval_samples_per_second": 16.877,
23
+ "eval_steps_per_second": 2.11,
24
+ "eval_wer": 0.847627340008707,
25
+ "step": 100
26
+ },
27
+ {
28
+ "epoch": 0.10314595152140278,
29
+ "grad_norm": 2.3965742588043213,
30
+ "learning_rate": 0.00029950404684002066,
31
+ "loss": 0.5648,
32
+ "step": 200
33
+ },
34
+ {
35
+ "epoch": 0.10314595152140278,
36
+ "eval_loss": 0.34078654646873474,
37
+ "eval_runtime": 57.7359,
38
+ "eval_samples_per_second": 16.905,
39
+ "eval_steps_per_second": 2.113,
40
+ "eval_wer": 0.5031562908141054,
41
+ "step": 200
42
+ },
43
+ {
44
+ "epoch": 0.15471892728210418,
45
+ "grad_norm": 12.905903816223145,
46
+ "learning_rate": 0.00029898742896504216,
47
+ "loss": 0.4827,
48
+ "step": 300
49
+ },
50
+ {
51
+ "epoch": 0.15471892728210418,
52
+ "eval_loss": 0.3261478543281555,
53
+ "eval_runtime": 57.6285,
54
+ "eval_samples_per_second": 16.936,
55
+ "eval_steps_per_second": 2.117,
56
+ "eval_wer": 0.4930343926861123,
57
+ "step": 300
58
+ },
59
+ {
60
+ "epoch": 0.20629190304280556,
61
+ "grad_norm": 1.9552323818206787,
62
+ "learning_rate": 0.00029847081109006366,
63
+ "loss": 0.4321,
64
+ "step": 400
65
+ },
66
+ {
67
+ "epoch": 0.20629190304280556,
68
+ "eval_loss": 0.303559273481369,
69
+ "eval_runtime": 57.7126,
70
+ "eval_samples_per_second": 16.911,
71
+ "eval_steps_per_second": 2.114,
72
+ "eval_wer": 0.48541575968654765,
73
+ "step": 400
74
+ },
75
+ {
76
+ "epoch": 0.25786487880350695,
77
+ "grad_norm": 2.4878430366516113,
78
+ "learning_rate": 0.0002979541932150852,
79
+ "loss": 0.4168,
80
+ "step": 500
81
+ },
82
+ {
83
+ "epoch": 0.25786487880350695,
84
+ "eval_loss": 0.2988507151603699,
85
+ "eval_runtime": 57.9159,
86
+ "eval_samples_per_second": 16.852,
87
+ "eval_steps_per_second": 2.107,
88
+ "eval_wer": 0.4783413147583805,
89
+ "step": 500
90
+ },
91
+ {
92
+ "epoch": 0.30943785456420836,
93
+ "grad_norm": 3.2817885875701904,
94
+ "learning_rate": 0.0002974375753401067,
95
+ "loss": 0.3965,
96
+ "step": 600
97
+ },
98
+ {
99
+ "epoch": 0.30943785456420836,
100
+ "eval_loss": 0.2906985282897949,
101
+ "eval_runtime": 57.5951,
102
+ "eval_samples_per_second": 16.946,
103
+ "eval_steps_per_second": 2.118,
104
+ "eval_wer": 0.45134958641706574,
105
+ "step": 600
106
+ },
107
+ {
108
+ "epoch": 0.36101083032490977,
109
+ "grad_norm": 1.416237235069275,
110
+ "learning_rate": 0.0002969209574651283,
111
+ "loss": 0.4199,
112
+ "step": 700
113
+ },
114
+ {
115
+ "epoch": 0.36101083032490977,
116
+ "eval_loss": 0.29259440302848816,
117
+ "eval_runtime": 57.5412,
118
+ "eval_samples_per_second": 16.962,
119
+ "eval_steps_per_second": 2.12,
120
+ "eval_wer": 0.4718110579016108,
121
+ "step": 700
122
+ },
123
+ {
124
+ "epoch": 0.4125838060856111,
125
+ "grad_norm": 2.3475992679595947,
126
+ "learning_rate": 0.0002964043395901498,
127
+ "loss": 0.3975,
128
+ "step": 800
129
+ },
130
+ {
131
+ "epoch": 0.4125838060856111,
132
+ "eval_loss": 0.288595974445343,
133
+ "eval_runtime": 57.3013,
134
+ "eval_samples_per_second": 17.033,
135
+ "eval_steps_per_second": 2.129,
136
+ "eval_wer": 0.44590770570309096,
137
+ "step": 800
138
+ },
139
+ {
140
+ "epoch": 0.46415678184631254,
141
+ "grad_norm": 1.5925949811935425,
142
+ "learning_rate": 0.00029588772171517134,
143
+ "loss": 0.3839,
144
+ "step": 900
145
+ },
146
+ {
147
+ "epoch": 0.46415678184631254,
148
+ "eval_loss": 0.29075172543525696,
149
+ "eval_runtime": 57.8589,
150
+ "eval_samples_per_second": 16.869,
151
+ "eval_steps_per_second": 2.109,
152
+ "eval_wer": 0.4722464083587288,
153
+ "step": 900
154
+ },
155
+ {
156
+ "epoch": 0.5157297576070139,
157
+ "grad_norm": 2.456458568572998,
158
+ "learning_rate": 0.00029537110384019285,
159
+ "loss": 0.3673,
160
+ "step": 1000
161
+ },
162
+ {
163
+ "epoch": 0.5157297576070139,
164
+ "eval_loss": 0.2836114764213562,
165
+ "eval_runtime": 57.7048,
166
+ "eval_samples_per_second": 16.914,
167
+ "eval_steps_per_second": 2.114,
168
+ "eval_wer": 0.44449281671745755,
169
+ "step": 1000
170
+ },
171
+ {
172
+ "epoch": 0.5673027333677153,
173
+ "grad_norm": 2.249444007873535,
174
+ "learning_rate": 0.00029485448596521435,
175
+ "loss": 0.3777,
176
+ "step": 1100
177
+ },
178
+ {
179
+ "epoch": 0.5673027333677153,
180
+ "eval_loss": 0.27841705083847046,
181
+ "eval_runtime": 57.4013,
182
+ "eval_samples_per_second": 17.003,
183
+ "eval_steps_per_second": 2.125,
184
+ "eval_wer": 0.4365476708750544,
185
+ "step": 1100
186
+ },
187
+ {
188
+ "epoch": 0.6188757091284167,
189
+ "grad_norm": 1.6211791038513184,
190
+ "learning_rate": 0.0002943378680902359,
191
+ "loss": 0.3764,
192
+ "step": 1200
193
+ },
194
+ {
195
+ "epoch": 0.6188757091284167,
196
+ "eval_loss": 0.2790738344192505,
197
+ "eval_runtime": 57.3594,
198
+ "eval_samples_per_second": 17.016,
199
+ "eval_steps_per_second": 2.127,
200
+ "eval_wer": 0.4278406617326948,
201
+ "step": 1200
202
+ },
203
+ {
204
+ "epoch": 0.6704486848891181,
205
+ "grad_norm": 2.047067165374756,
206
+ "learning_rate": 0.0002938212502152574,
207
+ "loss": 0.3918,
208
+ "step": 1300
209
+ },
210
+ {
211
+ "epoch": 0.6704486848891181,
212
+ "eval_loss": 0.27573078870773315,
213
+ "eval_runtime": 57.7956,
214
+ "eval_samples_per_second": 16.887,
215
+ "eval_steps_per_second": 2.111,
216
+ "eval_wer": 0.42511972137570747,
217
+ "step": 1300
218
+ },
219
+ {
220
+ "epoch": 0.7220216606498195,
221
+ "grad_norm": 1.4619171619415283,
222
+ "learning_rate": 0.00029330979851902874,
223
+ "loss": 0.3669,
224
+ "step": 1400
225
+ },
226
+ {
227
+ "epoch": 0.7220216606498195,
228
+ "eval_loss": 0.2721499502658844,
229
+ "eval_runtime": 57.7153,
230
+ "eval_samples_per_second": 16.911,
231
+ "eval_steps_per_second": 2.114,
232
+ "eval_wer": 0.41815411406181974,
233
+ "step": 1400
234
+ },
235
+ {
236
+ "epoch": 0.7735946364105208,
237
+ "grad_norm": 1.6276806592941284,
238
+ "learning_rate": 0.0002927931806440503,
239
+ "loss": 0.377,
240
+ "step": 1500
241
+ },
242
+ {
243
+ "epoch": 0.7735946364105208,
244
+ "eval_loss": 0.27284711599349976,
245
+ "eval_runtime": 57.3961,
246
+ "eval_samples_per_second": 17.005,
247
+ "eval_steps_per_second": 2.126,
248
+ "eval_wer": 0.47572921201567264,
249
+ "step": 1500
250
+ },
251
+ {
252
+ "epoch": 0.8251676121712223,
253
+ "grad_norm": 3.681490659713745,
254
+ "learning_rate": 0.0002922765627690718,
255
+ "loss": 0.4174,
256
+ "step": 1600
257
+ },
258
+ {
259
+ "epoch": 0.8251676121712223,
260
+ "eval_loss": 0.2684435546398163,
261
+ "eval_runtime": 57.3852,
262
+ "eval_samples_per_second": 17.008,
263
+ "eval_steps_per_second": 2.126,
264
+ "eval_wer": 0.4242490204614715,
265
+ "step": 1600
266
+ },
267
+ {
268
+ "epoch": 0.8767405879319237,
269
+ "grad_norm": 1.2958589792251587,
270
+ "learning_rate": 0.0002917599448940933,
271
+ "loss": 0.3641,
272
+ "step": 1700
273
+ },
274
+ {
275
+ "epoch": 0.8767405879319237,
276
+ "eval_loss": 0.2649287283420563,
277
+ "eval_runtime": 57.8551,
278
+ "eval_samples_per_second": 16.87,
279
+ "eval_steps_per_second": 2.109,
280
+ "eval_wer": 0.4194601654331737,
281
+ "step": 1700
282
+ },
283
+ {
284
+ "epoch": 0.9283135636926251,
285
+ "grad_norm": 1.8431603908538818,
286
+ "learning_rate": 0.00029124332701911486,
287
+ "loss": 0.3882,
288
+ "step": 1800
289
+ },
290
+ {
291
+ "epoch": 0.9283135636926251,
292
+ "eval_loss": 0.2646693289279938,
293
+ "eval_runtime": 57.7373,
294
+ "eval_samples_per_second": 16.904,
295
+ "eval_steps_per_second": 2.113,
296
+ "eval_wer": 0.41249455811928604,
297
+ "step": 1800
298
+ },
299
+ {
300
+ "epoch": 0.9798865394533265,
301
+ "grad_norm": 2.4119603633880615,
302
+ "learning_rate": 0.00029072670914413636,
303
+ "loss": 0.3861,
304
+ "step": 1900
305
+ },
306
+ {
307
+ "epoch": 0.9798865394533265,
308
+ "eval_loss": 0.2668125033378601,
309
+ "eval_runtime": 57.5111,
310
+ "eval_samples_per_second": 16.971,
311
+ "eval_steps_per_second": 2.121,
312
+ "eval_wer": 0.44253373966042664,
313
+ "step": 1900
314
+ },
315
+ {
316
+ "epoch": 1.0314595152140278,
317
+ "grad_norm": 1.174663782119751,
318
+ "learning_rate": 0.00029021009126915787,
319
+ "loss": 0.3647,
320
+ "step": 2000
321
+ },
322
+ {
323
+ "epoch": 1.0314595152140278,
324
+ "eval_loss": 0.26750749349594116,
325
+ "eval_runtime": 57.5411,
326
+ "eval_samples_per_second": 16.962,
327
+ "eval_steps_per_second": 2.12,
328
+ "eval_wer": 0.42457553330430997,
329
+ "step": 2000
330
+ },
331
+ {
332
+ "epoch": 1.0830324909747293,
333
+ "grad_norm": 0.9301519393920898,
334
+ "learning_rate": 0.0002896934733941794,
335
+ "loss": 0.3467,
336
+ "step": 2100
337
+ },
338
+ {
339
+ "epoch": 1.0830324909747293,
340
+ "eval_loss": 0.2629115581512451,
341
+ "eval_runtime": 57.7541,
342
+ "eval_samples_per_second": 16.899,
343
+ "eval_steps_per_second": 2.112,
344
+ "eval_wer": 0.40977361776229865,
345
+ "step": 2100
346
+ },
347
+ {
348
+ "epoch": 1.1346054667354306,
349
+ "grad_norm": 0.7093687057495117,
350
+ "learning_rate": 0.00028917685551920093,
351
+ "loss": 0.3579,
352
+ "step": 2200
353
+ },
354
+ {
355
+ "epoch": 1.1346054667354306,
356
+ "eval_loss": 0.258684903383255,
357
+ "eval_runtime": 57.7529,
358
+ "eval_samples_per_second": 16.9,
359
+ "eval_steps_per_second": 2.112,
360
+ "eval_wer": 0.41858946451893775,
361
+ "step": 2200
362
+ },
363
+ {
364
+ "epoch": 1.1861784424961321,
365
+ "grad_norm": 1.027114748954773,
366
+ "learning_rate": 0.00028866023764422243,
367
+ "loss": 0.3544,
368
+ "step": 2300
369
+ },
370
+ {
371
+ "epoch": 1.1861784424961321,
372
+ "eval_loss": 0.2609032392501831,
373
+ "eval_runtime": 57.5179,
374
+ "eval_samples_per_second": 16.969,
375
+ "eval_steps_per_second": 2.121,
376
+ "eval_wer": 0.412712233347845,
377
+ "step": 2300
378
+ },
379
+ {
380
+ "epoch": 1.2377514182568334,
381
+ "grad_norm": 0.9472237825393677,
382
+ "learning_rate": 0.000288143619769244,
383
+ "loss": 0.35,
384
+ "step": 2400
385
+ },
386
+ {
387
+ "epoch": 1.2377514182568334,
388
+ "eval_loss": 0.2592073678970337,
389
+ "eval_runtime": 57.1938,
390
+ "eval_samples_per_second": 17.065,
391
+ "eval_steps_per_second": 2.133,
392
+ "eval_wer": 0.4061819764910753,
393
+ "step": 2400
394
+ },
395
+ {
396
+ "epoch": 1.2893243940175347,
397
+ "grad_norm": 1.6566214561462402,
398
+ "learning_rate": 0.00028762700189426555,
399
+ "loss": 0.3519,
400
+ "step": 2500
401
+ },
402
+ {
403
+ "epoch": 1.2893243940175347,
404
+ "eval_loss": 0.2590978443622589,
405
+ "eval_runtime": 57.823,
406
+ "eval_samples_per_second": 16.879,
407
+ "eval_steps_per_second": 2.11,
408
+ "eval_wer": 0.4134740966478015,
409
+ "step": 2500
410
+ },
411
+ {
412
+ "epoch": 1.2893243940175347,
413
+ "step": 2500,
414
+ "total_flos": 5.958212987780215e+18,
415
+ "train_loss": 0.6475233551025391,
416
+ "train_runtime": 3633.5111,
417
+ "train_samples_per_second": 64.037,
418
+ "train_steps_per_second": 16.009
419
+ }
420
+ ],
421
+ "logging_steps": 100,
422
+ "max_steps": 58170,
423
+ "num_input_tokens_seen": 0,
424
+ "num_train_epochs": 30,
425
+ "save_steps": 400,
426
+ "stateful_callbacks": {
427
+ "EarlyStoppingCallback": {
428
+ "args": {
429
+ "early_stopping_patience": 3,
430
+ "early_stopping_threshold": 0.0
431
+ },
432
+ "attributes": {
433
+ "early_stopping_patience_counter": 2
434
+ }
435
+ },
436
+ "TrainerControl": {
437
+ "args": {
438
+ "should_epoch_stop": false,
439
+ "should_evaluate": false,
440
+ "should_log": false,
441
+ "should_save": true,
442
+ "should_training_stop": false
443
+ },
444
+ "attributes": {}
445
+ }
446
+ },
447
+ "total_flos": 5.958212987780215e+18,
448
+ "train_batch_size": 4,
449
+ "trial_name": null,
450
+ "trial_params": null
451
+ }