metadata
language:
- pt
license: apache-2.0
tags:
- automatic-speech-recognition
- mozilla-foundation/common_voice_8_0
- generated_from_trainer
datasets:
- common_voice
model-index:
- name: output
results: []
output
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - PT dataset. It achieves the following results on the evaluation set:
- Loss: 0.1505
- Wer: 0.1352
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 7.5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 2000
- num_epochs: 50.0
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer |
---|---|---|---|---|
4.1367 | 0.64 | 500 | 3.8825 | 1.0 |
2.9677 | 1.29 | 1000 | 2.9498 | 1.0 |
1.5884 | 1.93 | 1500 | 0.6722 | 0.6493 |
1.2292 | 2.57 | 2000 | 0.3635 | 0.3202 |
1.1314 | 3.22 | 2500 | 0.2970 | 0.2680 |
1.0879 | 3.86 | 3000 | 0.2671 | 0.2486 |
1.0344 | 4.5 | 3500 | 0.2625 | 0.2239 |
1.0109 | 5.15 | 4000 | 0.2520 | 0.2230 |
0.9966 | 5.79 | 4500 | 0.2280 | 0.2105 |
0.9815 | 6.43 | 5000 | 0.2254 | 0.2179 |
0.9744 | 7.08 | 5500 | 0.2301 | 0.2137 |
0.9487 | 7.72 | 6000 | 0.2224 | 0.2051 |
0.9431 | 8.37 | 6500 | 0.2105 | 0.1992 |
0.9365 | 9.01 | 7000 | 0.2114 | 0.2019 |
0.9268 | 9.65 | 7500 | 0.2097 | 0.1988 |
0.9292 | 10.3 | 8000 | 0.2120 | 0.1986 |
0.929 | 10.94 | 8500 | 0.2048 | 0.1998 |
0.9017 | 11.58 | 9000 | 0.2035 | 0.1999 |
0.8898 | 12.23 | 9500 | 0.1961 | 0.1908 |
0.8799 | 12.87 | 10000 | 0.1945 | 0.1817 |
0.869 | 13.51 | 10500 | 0.1929 | 0.1844 |
0.8572 | 14.16 | 11000 | 0.1941 | 0.1888 |
0.8691 | 14.8 | 11500 | 0.1912 | 0.1804 |
0.8645 | 15.44 | 12000 | 0.1950 | 0.1851 |
0.8468 | 16.09 | 12500 | 0.1879 | 0.1770 |
0.8405 | 16.73 | 13000 | 0.1881 | 0.1759 |
0.8647 | 17.37 | 13500 | 0.1861 | 0.1740 |
0.8477 | 18.02 | 14000 | 0.1782 | 0.1702 |
0.811 | 18.66 | 14500 | 0.1915 | 0.1757 |
0.8165 | 19.3 | 15000 | 0.1820 | 0.1724 |
0.8166 | 19.95 | 15500 | 0.1798 | 0.1697 |
0.8167 | 20.59 | 16000 | 0.1805 | 0.1752 |
0.7908 | 21.24 | 16500 | 0.1761 | 0.1699 |
0.7925 | 21.88 | 17000 | 0.1740 | 0.1709 |
0.7803 | 22.52 | 17500 | 0.1815 | 0.1727 |
0.7839 | 23.17 | 18000 | 0.1737 | 0.1694 |
0.7815 | 23.81 | 18500 | 0.1732 | 0.1630 |
0.767 | 24.45 | 19000 | 0.1724 | 0.1648 |
0.7672 | 25.1 | 19500 | 0.1706 | 0.1596 |
0.7691 | 25.74 | 20000 | 0.1718 | 0.1618 |
0.7547 | 26.38 | 20500 | 0.1694 | 0.1565 |
0.7498 | 27.03 | 21000 | 0.1706 | 0.1582 |
0.7459 | 27.67 | 21500 | 0.1663 | 0.1586 |
0.7374 | 28.31 | 22000 | 0.1651 | 0.1567 |
0.7499 | 28.96 | 22500 | 0.1668 | 0.1549 |
0.7471 | 29.6 | 23000 | 0.1667 | 0.1553 |
0.7369 | 30.24 | 23500 | 0.1659 | 0.1556 |
0.7389 | 30.89 | 24000 | 0.1668 | 0.1538 |
0.7197 | 31.53 | 24500 | 0.1687 | 0.1561 |
0.71 | 32.17 | 25000 | 0.1666 | 0.1516 |
0.7199 | 32.82 | 25500 | 0.1640 | 0.1523 |
0.7194 | 33.46 | 26000 | 0.1659 | 0.1528 |
0.6923 | 34.11 | 26500 | 0.1662 | 0.1507 |
0.7054 | 34.75 | 27000 | 0.1641 | 0.1486 |
0.6955 | 35.39 | 27500 | 0.1634 | 0.1497 |
0.7084 | 36.04 | 28000 | 0.1618 | 0.1478 |
0.6917 | 36.68 | 28500 | 0.1589 | 0.1471 |
0.687 | 37.32 | 29000 | 0.1589 | 0.1450 |
0.6914 | 37.97 | 29500 | 0.1588 | 0.1465 |
0.6646 | 38.61 | 30000 | 0.1602 | 0.1468 |
0.6667 | 39.25 | 30500 | 0.1588 | 0.1444 |
0.6754 | 39.9 | 31000 | 0.1587 | 0.1455 |
0.6632 | 40.54 | 31500 | 0.1586 | 0.1461 |
0.6619 | 41.18 | 32000 | 0.1571 | 0.1441 |
0.6561 | 41.83 | 32500 | 0.1564 | 0.1420 |
0.6492 | 42.47 | 33000 | 0.1539 | 0.1437 |
0.6649 | 43.11 | 33500 | 0.1512 | 0.1406 |
0.6511 | 43.76 | 34000 | 0.1539 | 0.1384 |
0.6551 | 44.4 | 34500 | 0.1520 | 0.1384 |
0.6452 | 45.05 | 35000 | 0.1510 | 0.1368 |
0.6155 | 45.69 | 35500 | 0.1522 | 0.1375 |
0.628 | 46.33 | 36000 | 0.1522 | 0.1366 |
0.6389 | 46.97 | 36500 | 0.1513 | 0.1377 |
0.6265 | 47.62 | 37000 | 0.1512 | 0.1369 |
0.6197 | 48.26 | 37500 | 0.1511 | 0.1362 |
0.621 | 48.91 | 38000 | 0.1510 | 0.1357 |
0.6259 | 49.55 | 38500 | 0.1506 | 0.1353 |
Framework versions
- Transformers 4.17.0
- Pytorch 1.9.1+cu102
- Datasets 2.0.0
- Tokenizers 0.11.6