elsayedissa commited on
Commit
7b1e60e
·
1 Parent(s): 8f31d06

Training in progress, step 8000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8b93e7897bc3f43ea6b1a664fc0daee99a1df27feebf5e9743f2a0bbfb65f30
3
  size 2524414853
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01ff7248d086cb5c4cedadeb579adbb675a1c64e4fb09e5ae750359a44febb76
3
  size 2524414853
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4c04132f7025e27d9cf890f119fc19b9fbfc35fa926c5345bb77dcfcf19e918
3
  size 1262229869
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4607f812b37e97d000dcc622b3e637fb3031eca1d8fd3c7555cdaadfec40180b
3
  size 1262229869
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aebbece3849c4ea65c57e14ac7218389ce4023006361ce66491299183a563542
3
  size 14639
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:977a8fef6282492c80e91318169668264512de99e37019c60391b3f14bdd2233
3
  size 14639
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e97f75ec4a68ce2d88e985bf5af3c9d9064b2fdb2fc3093a5ded26ae274c6de
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3c549db73c256944038c972af144f55d1e0fbea4d86305beed62e57f9cae640
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.06791821116824107,
3
  "best_model_checkpoint": "./wav2vec2-xls-r-300m-ar/checkpoint-6000",
4
- "epoch": 61.94690265486726,
5
- "global_step": 7000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -1749,11 +1749,260 @@
1749
  "eval_steps_per_second": 1.206,
1750
  "eval_wer": 0.07174458926222647,
1751
  "step": 7000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1752
  }
1753
  ],
1754
  "max_steps": 10000,
1755
  "num_train_epochs": 89,
1756
- "total_flos": 2.756451071078706e+19,
1757
  "trial_name": null,
1758
  "trial_params": null
1759
  }
 
1
  {
2
  "best_metric": 0.06791821116824107,
3
  "best_model_checkpoint": "./wav2vec2-xls-r-300m-ar/checkpoint-6000",
4
+ "epoch": 70.79646017699115,
5
+ "global_step": 8000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
1749
  "eval_steps_per_second": 1.206,
1750
  "eval_wer": 0.07174458926222647,
1751
  "step": 7000
1752
+ },
1753
+ {
1754
+ "epoch": 62.17,
1755
+ "learning_rate": 9.394736842105262e-05,
1756
+ "loss": 0.0116,
1757
+ "step": 7025
1758
+ },
1759
+ {
1760
+ "epoch": 62.39,
1761
+ "learning_rate": 9.31578947368421e-05,
1762
+ "loss": 0.0102,
1763
+ "step": 7050
1764
+ },
1765
+ {
1766
+ "epoch": 62.61,
1767
+ "learning_rate": 9.236842105263157e-05,
1768
+ "loss": 0.0118,
1769
+ "step": 7075
1770
+ },
1771
+ {
1772
+ "epoch": 62.83,
1773
+ "learning_rate": 9.157894736842104e-05,
1774
+ "loss": 0.0093,
1775
+ "step": 7100
1776
+ },
1777
+ {
1778
+ "epoch": 63.05,
1779
+ "learning_rate": 9.078947368421052e-05,
1780
+ "loss": 0.0074,
1781
+ "step": 7125
1782
+ },
1783
+ {
1784
+ "epoch": 63.27,
1785
+ "learning_rate": 8.999999999999999e-05,
1786
+ "loss": 0.0079,
1787
+ "step": 7150
1788
+ },
1789
+ {
1790
+ "epoch": 63.5,
1791
+ "learning_rate": 8.921052631578946e-05,
1792
+ "loss": 0.0075,
1793
+ "step": 7175
1794
+ },
1795
+ {
1796
+ "epoch": 63.72,
1797
+ "learning_rate": 8.842105263157893e-05,
1798
+ "loss": 0.0099,
1799
+ "step": 7200
1800
+ },
1801
+ {
1802
+ "epoch": 63.94,
1803
+ "learning_rate": 8.763157894736841e-05,
1804
+ "loss": 0.0098,
1805
+ "step": 7225
1806
+ },
1807
+ {
1808
+ "epoch": 64.16,
1809
+ "learning_rate": 8.68421052631579e-05,
1810
+ "loss": 0.0109,
1811
+ "step": 7250
1812
+ },
1813
+ {
1814
+ "epoch": 64.38,
1815
+ "learning_rate": 8.605263157894737e-05,
1816
+ "loss": 0.0103,
1817
+ "step": 7275
1818
+ },
1819
+ {
1820
+ "epoch": 64.6,
1821
+ "learning_rate": 8.526315789473684e-05,
1822
+ "loss": 0.0074,
1823
+ "step": 7300
1824
+ },
1825
+ {
1826
+ "epoch": 64.82,
1827
+ "learning_rate": 8.44736842105263e-05,
1828
+ "loss": 0.0101,
1829
+ "step": 7325
1830
+ },
1831
+ {
1832
+ "epoch": 65.04,
1833
+ "learning_rate": 8.368421052631578e-05,
1834
+ "loss": 0.009,
1835
+ "step": 7350
1836
+ },
1837
+ {
1838
+ "epoch": 65.27,
1839
+ "learning_rate": 8.289473684210526e-05,
1840
+ "loss": 0.0077,
1841
+ "step": 7375
1842
+ },
1843
+ {
1844
+ "epoch": 65.49,
1845
+ "learning_rate": 8.210526315789474e-05,
1846
+ "loss": 0.009,
1847
+ "step": 7400
1848
+ },
1849
+ {
1850
+ "epoch": 65.71,
1851
+ "learning_rate": 8.131578947368421e-05,
1852
+ "loss": 0.0098,
1853
+ "step": 7425
1854
+ },
1855
+ {
1856
+ "epoch": 65.93,
1857
+ "learning_rate": 8.052631578947368e-05,
1858
+ "loss": 0.0096,
1859
+ "step": 7450
1860
+ },
1861
+ {
1862
+ "epoch": 66.15,
1863
+ "learning_rate": 7.973684210526315e-05,
1864
+ "loss": 0.0136,
1865
+ "step": 7475
1866
+ },
1867
+ {
1868
+ "epoch": 66.37,
1869
+ "learning_rate": 7.894736842105262e-05,
1870
+ "loss": 0.0082,
1871
+ "step": 7500
1872
+ },
1873
+ {
1874
+ "epoch": 66.59,
1875
+ "learning_rate": 7.815789473684209e-05,
1876
+ "loss": 0.0089,
1877
+ "step": 7525
1878
+ },
1879
+ {
1880
+ "epoch": 66.81,
1881
+ "learning_rate": 7.736842105263159e-05,
1882
+ "loss": 0.0089,
1883
+ "step": 7550
1884
+ },
1885
+ {
1886
+ "epoch": 67.04,
1887
+ "learning_rate": 7.657894736842105e-05,
1888
+ "loss": 0.0084,
1889
+ "step": 7575
1890
+ },
1891
+ {
1892
+ "epoch": 67.26,
1893
+ "learning_rate": 7.578947368421052e-05,
1894
+ "loss": 0.0073,
1895
+ "step": 7600
1896
+ },
1897
+ {
1898
+ "epoch": 67.48,
1899
+ "learning_rate": 7.5e-05,
1900
+ "loss": 0.008,
1901
+ "step": 7625
1902
+ },
1903
+ {
1904
+ "epoch": 67.7,
1905
+ "learning_rate": 7.421052631578946e-05,
1906
+ "loss": 0.0083,
1907
+ "step": 7650
1908
+ },
1909
+ {
1910
+ "epoch": 67.92,
1911
+ "learning_rate": 7.342105263157895e-05,
1912
+ "loss": 0.0068,
1913
+ "step": 7675
1914
+ },
1915
+ {
1916
+ "epoch": 68.14,
1917
+ "learning_rate": 7.263157894736842e-05,
1918
+ "loss": 0.007,
1919
+ "step": 7700
1920
+ },
1921
+ {
1922
+ "epoch": 68.36,
1923
+ "learning_rate": 7.184210526315788e-05,
1924
+ "loss": 0.0086,
1925
+ "step": 7725
1926
+ },
1927
+ {
1928
+ "epoch": 68.58,
1929
+ "learning_rate": 7.105263157894735e-05,
1930
+ "loss": 0.0077,
1931
+ "step": 7750
1932
+ },
1933
+ {
1934
+ "epoch": 68.81,
1935
+ "learning_rate": 7.026315789473684e-05,
1936
+ "loss": 0.0076,
1937
+ "step": 7775
1938
+ },
1939
+ {
1940
+ "epoch": 69.03,
1941
+ "learning_rate": 6.947368421052631e-05,
1942
+ "loss": 0.006,
1943
+ "step": 7800
1944
+ },
1945
+ {
1946
+ "epoch": 69.25,
1947
+ "learning_rate": 6.868421052631578e-05,
1948
+ "loss": 0.006,
1949
+ "step": 7825
1950
+ },
1951
+ {
1952
+ "epoch": 69.47,
1953
+ "learning_rate": 6.789473684210526e-05,
1954
+ "loss": 0.0077,
1955
+ "step": 7850
1956
+ },
1957
+ {
1958
+ "epoch": 69.69,
1959
+ "learning_rate": 6.710526315789473e-05,
1960
+ "loss": 0.009,
1961
+ "step": 7875
1962
+ },
1963
+ {
1964
+ "epoch": 69.91,
1965
+ "learning_rate": 6.63157894736842e-05,
1966
+ "loss": 0.0081,
1967
+ "step": 7900
1968
+ },
1969
+ {
1970
+ "epoch": 70.13,
1971
+ "learning_rate": 6.552631578947368e-05,
1972
+ "loss": 0.0085,
1973
+ "step": 7925
1974
+ },
1975
+ {
1976
+ "epoch": 70.35,
1977
+ "learning_rate": 6.473684210526315e-05,
1978
+ "loss": 0.007,
1979
+ "step": 7950
1980
+ },
1981
+ {
1982
+ "epoch": 70.58,
1983
+ "learning_rate": 6.394736842105262e-05,
1984
+ "loss": 0.0088,
1985
+ "step": 7975
1986
+ },
1987
+ {
1988
+ "epoch": 70.8,
1989
+ "learning_rate": 6.315789473684209e-05,
1990
+ "loss": 0.0074,
1991
+ "step": 8000
1992
+ },
1993
+ {
1994
+ "epoch": 70.8,
1995
+ "eval_loss": 0.30043184757232666,
1996
+ "eval_runtime": 10.795,
1997
+ "eval_samples_per_second": 9.264,
1998
+ "eval_steps_per_second": 1.204,
1999
+ "eval_wer": 0.0680377854836781,
2000
+ "step": 8000
2001
  }
2002
  ],
2003
  "max_steps": 10000,
2004
  "num_train_epochs": 89,
2005
+ "total_flos": 3.1517992803917713e+19,
2006
  "trial_name": null,
2007
  "trial_params": null
2008
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4c04132f7025e27d9cf890f119fc19b9fbfc35fa926c5345bb77dcfcf19e918
3
  size 1262229869
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4607f812b37e97d000dcc622b3e637fb3031eca1d8fd3c7555cdaadfec40180b
3
  size 1262229869
runs/Apr09_14-45-21_gpu07.cyverse.org/events.out.tfevents.1681076808.gpu07.cyverse.org.48142.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17afcb646f4bffba0651ee2b2d287faa8b2e8e9a98dac1f49fdb1210b9df2b8a
3
- size 51568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14933acd55b5be1858479eeffcc145290dda1fac71c2c315cec1bd00af72f3df
3
+ size 58166