|
{ |
|
"best_metric": 19.094888228857275, |
|
"best_model_checkpoint": "./checkpoint-19000", |
|
"epoch": 400.0, |
|
"eval_steps": 1000, |
|
"global_step": 40000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.6499999999999997e-06, |
|
"loss": 2.4538, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 3.5249999999999997e-06, |
|
"loss": 1.87, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.399999999999999e-06, |
|
"loss": 1.3171, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.275e-06, |
|
"loss": 1.016, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.149999999999999e-06, |
|
"loss": 0.834, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.1024999999999999e-05, |
|
"loss": 0.7356, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.2899999999999998e-05, |
|
"loss": 0.6513, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.4775e-05, |
|
"loss": 0.5926, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.6649999999999998e-05, |
|
"loss": 0.5086, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.8525e-05, |
|
"loss": 0.4757, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.04e-05, |
|
"loss": 0.4501, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.2274999999999996e-05, |
|
"loss": 0.4266, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.4149999999999997e-05, |
|
"loss": 0.357, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 2.6024999999999996e-05, |
|
"loss": 0.3415, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 2.7899999999999997e-05, |
|
"loss": 0.3319, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 2.9775e-05, |
|
"loss": 0.3189, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 3.165e-05, |
|
"loss": 0.253, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 3.3524999999999995e-05, |
|
"loss": 0.2526, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 3.539999999999999e-05, |
|
"loss": 0.2463, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 3.7275e-05, |
|
"loss": 0.2471, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 3.747911392405063e-05, |
|
"loss": 0.184, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 3.7455379746835444e-05, |
|
"loss": 0.1809, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 3.743164556962025e-05, |
|
"loss": 0.1773, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 3.7407911392405063e-05, |
|
"loss": 0.1852, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 3.738417721518987e-05, |
|
"loss": 0.1259, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 3.736044303797468e-05, |
|
"loss": 0.1287, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 3.733670886075949e-05, |
|
"loss": 0.1289, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 3.73129746835443e-05, |
|
"loss": 0.1312, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 3.7289240506329114e-05, |
|
"loss": 0.0865, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 3.726550632911392e-05, |
|
"loss": 0.0904, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 3.724177215189873e-05, |
|
"loss": 0.0907, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 3.721803797468354e-05, |
|
"loss": 0.0926, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 3.719430379746835e-05, |
|
"loss": 0.0587, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 3.717056962025316e-05, |
|
"loss": 0.061, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 3.714683544303797e-05, |
|
"loss": 0.0625, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 3.7123101265822784e-05, |
|
"loss": 0.0624, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 3.709936708860759e-05, |
|
"loss": 0.0386, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 3.70756329113924e-05, |
|
"loss": 0.0405, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 3.705189873417721e-05, |
|
"loss": 0.0415, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 3.702816455696202e-05, |
|
"loss": 0.0426, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.34511512517929077, |
|
"eval_runtime": 91.8649, |
|
"eval_samples_per_second": 138.704, |
|
"eval_steps_per_second": 1.089, |
|
"eval_wer": 23.200328278880725, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 10.25, |
|
"learning_rate": 3.700443037974683e-05, |
|
"loss": 0.0259, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 3.698069620253164e-05, |
|
"loss": 0.0259, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"learning_rate": 3.6956962025316454e-05, |
|
"loss": 0.0277, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 3.693322784810126e-05, |
|
"loss": 0.0279, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 3.690949367088607e-05, |
|
"loss": 0.0177, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 11.5, |
|
"learning_rate": 3.688575949367088e-05, |
|
"loss": 0.0184, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 11.75, |
|
"learning_rate": 3.686202531645569e-05, |
|
"loss": 0.018, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 3.68382911392405e-05, |
|
"loss": 0.019, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 12.25, |
|
"learning_rate": 3.681455696202531e-05, |
|
"loss": 0.0122, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 3.6790822784810124e-05, |
|
"loss": 0.0136, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 12.75, |
|
"learning_rate": 3.676708860759493e-05, |
|
"loss": 0.0136, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 3.674335443037974e-05, |
|
"loss": 0.014, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 13.25, |
|
"learning_rate": 3.6719620253164556e-05, |
|
"loss": 0.0097, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 13.5, |
|
"learning_rate": 3.669588607594937e-05, |
|
"loss": 0.0096, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 13.75, |
|
"learning_rate": 3.6672151898734175e-05, |
|
"loss": 0.0087, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.664841772151899e-05, |
|
"loss": 0.0088, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 14.25, |
|
"learning_rate": 3.6624683544303794e-05, |
|
"loss": 0.0064, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 14.5, |
|
"learning_rate": 3.660094936708861e-05, |
|
"loss": 0.0064, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 14.75, |
|
"learning_rate": 3.657721518987341e-05, |
|
"loss": 0.0063, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.6553481012658226e-05, |
|
"loss": 0.0063, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 15.25, |
|
"learning_rate": 3.652974683544304e-05, |
|
"loss": 0.0052, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 15.5, |
|
"learning_rate": 3.6506012658227845e-05, |
|
"loss": 0.0057, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 15.75, |
|
"learning_rate": 3.648227848101266e-05, |
|
"loss": 0.0059, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.6458544303797464e-05, |
|
"loss": 0.0057, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 16.25, |
|
"learning_rate": 3.643481012658228e-05, |
|
"loss": 0.0044, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 16.5, |
|
"learning_rate": 3.641107594936708e-05, |
|
"loss": 0.0044, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 16.75, |
|
"learning_rate": 3.6387341772151896e-05, |
|
"loss": 0.0047, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 3.636360759493671e-05, |
|
"loss": 0.0051, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 17.25, |
|
"learning_rate": 3.6339873417721515e-05, |
|
"loss": 0.0047, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 3.631613924050633e-05, |
|
"loss": 0.0052, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 17.75, |
|
"learning_rate": 3.6292405063291134e-05, |
|
"loss": 0.006, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 3.626867088607595e-05, |
|
"loss": 0.0079, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 18.25, |
|
"learning_rate": 3.624493670886075e-05, |
|
"loss": 0.0061, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 18.5, |
|
"learning_rate": 3.6221202531645566e-05, |
|
"loss": 0.0067, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 18.75, |
|
"learning_rate": 3.619746835443038e-05, |
|
"loss": 0.007, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"learning_rate": 3.6173734177215185e-05, |
|
"loss": 0.0083, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 19.25, |
|
"learning_rate": 3.615e-05, |
|
"loss": 0.0072, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 19.5, |
|
"learning_rate": 3.6126265822784804e-05, |
|
"loss": 0.0079, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 19.75, |
|
"learning_rate": 3.610253164556962e-05, |
|
"loss": 0.0078, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 3.607879746835442e-05, |
|
"loss": 0.0077, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.41232848167419434, |
|
"eval_runtime": 81.6367, |
|
"eval_samples_per_second": 156.082, |
|
"eval_steps_per_second": 1.225, |
|
"eval_wer": 22.605322807566047, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 20.25, |
|
"learning_rate": 3.6055063291139236e-05, |
|
"loss": 0.0059, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 20.5, |
|
"learning_rate": 3.603132911392405e-05, |
|
"loss": 0.0063, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 20.75, |
|
"learning_rate": 3.6007594936708855e-05, |
|
"loss": 0.0078, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"learning_rate": 3.598386075949367e-05, |
|
"loss": 0.0081, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 21.25, |
|
"learning_rate": 3.596012658227848e-05, |
|
"loss": 0.0065, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 21.5, |
|
"learning_rate": 3.593639240506329e-05, |
|
"loss": 0.0078, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 21.75, |
|
"learning_rate": 3.59126582278481e-05, |
|
"loss": 0.008, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"learning_rate": 3.588892405063291e-05, |
|
"loss": 0.0081, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 22.25, |
|
"learning_rate": 3.586518987341772e-05, |
|
"loss": 0.0071, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"learning_rate": 3.584145569620253e-05, |
|
"loss": 0.007, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 22.75, |
|
"learning_rate": 3.581772151898734e-05, |
|
"loss": 0.0068, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"learning_rate": 3.579398734177215e-05, |
|
"loss": 0.0073, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 23.25, |
|
"learning_rate": 3.577025316455696e-05, |
|
"loss": 0.0052, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 23.5, |
|
"learning_rate": 3.574651898734177e-05, |
|
"loss": 0.0051, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 3.572278481012658e-05, |
|
"loss": 0.0054, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"learning_rate": 3.569905063291139e-05, |
|
"loss": 0.0053, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 24.25, |
|
"learning_rate": 3.56753164556962e-05, |
|
"loss": 0.0045, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 24.5, |
|
"learning_rate": 3.565158227848101e-05, |
|
"loss": 0.0044, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 24.75, |
|
"learning_rate": 3.562784810126582e-05, |
|
"loss": 0.004, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 3.5604113924050627e-05, |
|
"loss": 0.0044, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 25.25, |
|
"learning_rate": 3.558037974683544e-05, |
|
"loss": 0.0033, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 25.5, |
|
"learning_rate": 3.555664556962025e-05, |
|
"loss": 0.0029, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 25.75, |
|
"learning_rate": 3.553291139240506e-05, |
|
"loss": 0.0028, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"learning_rate": 3.550917721518987e-05, |
|
"loss": 0.0033, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 26.25, |
|
"learning_rate": 3.548544303797468e-05, |
|
"loss": 0.0031, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 26.5, |
|
"learning_rate": 3.546170886075949e-05, |
|
"loss": 0.0028, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 26.75, |
|
"learning_rate": 3.5437974683544297e-05, |
|
"loss": 0.0028, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"learning_rate": 3.541424050632911e-05, |
|
"loss": 0.0027, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 27.25, |
|
"learning_rate": 3.539050632911392e-05, |
|
"loss": 0.0025, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"learning_rate": 3.536677215189873e-05, |
|
"loss": 0.0028, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 27.75, |
|
"learning_rate": 3.534303797468354e-05, |
|
"loss": 0.0025, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"learning_rate": 3.531930379746835e-05, |
|
"loss": 0.0024, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 28.25, |
|
"learning_rate": 3.529556962025316e-05, |
|
"loss": 0.002, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 28.5, |
|
"learning_rate": 3.527183544303797e-05, |
|
"loss": 0.0021, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 28.75, |
|
"learning_rate": 3.524810126582278e-05, |
|
"loss": 0.0019, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"learning_rate": 3.522436708860759e-05, |
|
"loss": 0.002, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 29.25, |
|
"learning_rate": 3.5200632911392405e-05, |
|
"loss": 0.0015, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 29.5, |
|
"learning_rate": 3.517689873417721e-05, |
|
"loss": 0.0014, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 29.75, |
|
"learning_rate": 3.5153164556962024e-05, |
|
"loss": 0.0015, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 3.512943037974684e-05, |
|
"loss": 0.0013, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 0.4287857711315155, |
|
"eval_runtime": 80.2692, |
|
"eval_samples_per_second": 158.741, |
|
"eval_steps_per_second": 1.246, |
|
"eval_wer": 21.19645927778646, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 30.25, |
|
"learning_rate": 3.510569620253164e-05, |
|
"loss": 0.001, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 30.5, |
|
"learning_rate": 3.5081962025316456e-05, |
|
"loss": 0.0009, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 30.75, |
|
"learning_rate": 3.505822784810126e-05, |
|
"loss": 0.0009, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"learning_rate": 3.5034493670886075e-05, |
|
"loss": 0.0009, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 3.501075949367088e-05, |
|
"loss": 0.0008, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 31.5, |
|
"learning_rate": 3.4987025316455694e-05, |
|
"loss": 0.0008, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 31.75, |
|
"learning_rate": 3.496329113924051e-05, |
|
"loss": 0.0008, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"learning_rate": 3.493955696202531e-05, |
|
"loss": 0.0008, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 32.25, |
|
"learning_rate": 3.4915822784810126e-05, |
|
"loss": 0.0007, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"learning_rate": 3.489208860759493e-05, |
|
"loss": 0.0007, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 32.75, |
|
"learning_rate": 3.4868354430379745e-05, |
|
"loss": 0.0007, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"learning_rate": 3.484462025316455e-05, |
|
"loss": 0.0007, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 33.25, |
|
"learning_rate": 3.4820886075949364e-05, |
|
"loss": 0.0006, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 33.5, |
|
"learning_rate": 3.479715189873418e-05, |
|
"loss": 0.0006, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 33.75, |
|
"learning_rate": 3.477341772151898e-05, |
|
"loss": 0.0006, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"learning_rate": 3.4749683544303796e-05, |
|
"loss": 0.0006, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 34.25, |
|
"learning_rate": 3.47259493670886e-05, |
|
"loss": 0.0006, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 34.5, |
|
"learning_rate": 3.4702215189873415e-05, |
|
"loss": 0.0006, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 34.75, |
|
"learning_rate": 3.467848101265822e-05, |
|
"loss": 0.0006, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 3.4654746835443034e-05, |
|
"loss": 0.0006, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 35.25, |
|
"learning_rate": 3.463101265822785e-05, |
|
"loss": 0.0005, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 35.5, |
|
"learning_rate": 3.460727848101265e-05, |
|
"loss": 0.0005, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 35.75, |
|
"learning_rate": 3.4583544303797466e-05, |
|
"loss": 0.0005, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"learning_rate": 3.455981012658227e-05, |
|
"loss": 0.0005, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 36.25, |
|
"learning_rate": 3.4536075949367085e-05, |
|
"loss": 0.0005, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 36.5, |
|
"learning_rate": 3.451234177215189e-05, |
|
"loss": 0.0005, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 36.75, |
|
"learning_rate": 3.4488607594936704e-05, |
|
"loss": 0.0005, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"learning_rate": 3.446487341772152e-05, |
|
"loss": 0.0005, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 37.25, |
|
"learning_rate": 3.444113924050632e-05, |
|
"loss": 0.0005, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"learning_rate": 3.4417405063291136e-05, |
|
"loss": 0.0005, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 37.75, |
|
"learning_rate": 3.439367088607595e-05, |
|
"loss": 0.0005, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"learning_rate": 3.436993670886076e-05, |
|
"loss": 0.0005, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 38.25, |
|
"learning_rate": 3.434620253164557e-05, |
|
"loss": 0.0004, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 38.5, |
|
"learning_rate": 3.432246835443038e-05, |
|
"loss": 0.0004, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 38.75, |
|
"learning_rate": 3.429873417721519e-05, |
|
"loss": 0.0004, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"learning_rate": 3.4275e-05, |
|
"loss": 0.0004, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 39.25, |
|
"learning_rate": 3.4251265822784806e-05, |
|
"loss": 0.0004, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 39.5, |
|
"learning_rate": 3.422753164556962e-05, |
|
"loss": 0.0004, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 39.75, |
|
"learning_rate": 3.420379746835443e-05, |
|
"loss": 0.0004, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 3.418006329113924e-05, |
|
"loss": 0.0004, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 0.45376476645469666, |
|
"eval_runtime": 81.5739, |
|
"eval_samples_per_second": 156.202, |
|
"eval_steps_per_second": 1.226, |
|
"eval_wer": 21.192551195873065, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 40.25, |
|
"learning_rate": 3.415632911392405e-05, |
|
"loss": 0.0004, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 40.5, |
|
"learning_rate": 3.413259493670886e-05, |
|
"loss": 0.0004, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 40.75, |
|
"learning_rate": 3.410886075949367e-05, |
|
"loss": 0.0004, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"learning_rate": 3.4085126582278476e-05, |
|
"loss": 0.0004, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 41.25, |
|
"learning_rate": 3.406139240506329e-05, |
|
"loss": 0.0004, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 41.5, |
|
"learning_rate": 3.40376582278481e-05, |
|
"loss": 0.0004, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 41.75, |
|
"learning_rate": 3.401392405063291e-05, |
|
"loss": 0.0004, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"learning_rate": 3.399018987341772e-05, |
|
"loss": 0.0004, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 42.25, |
|
"learning_rate": 3.396645569620253e-05, |
|
"loss": 0.0003, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 3.394272151898734e-05, |
|
"loss": 0.0003, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 42.75, |
|
"learning_rate": 3.3918987341772146e-05, |
|
"loss": 0.0004, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"learning_rate": 3.389525316455696e-05, |
|
"loss": 0.0004, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 43.25, |
|
"learning_rate": 3.387151898734177e-05, |
|
"loss": 0.0003, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 43.5, |
|
"learning_rate": 3.384778481012658e-05, |
|
"loss": 0.0003, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 43.75, |
|
"learning_rate": 3.382405063291139e-05, |
|
"loss": 0.0003, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"learning_rate": 3.3800316455696197e-05, |
|
"loss": 0.0003, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 44.25, |
|
"learning_rate": 3.377658227848101e-05, |
|
"loss": 0.0003, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 44.5, |
|
"learning_rate": 3.3752848101265816e-05, |
|
"loss": 0.0003, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 44.75, |
|
"learning_rate": 3.372911392405063e-05, |
|
"loss": 0.0003, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 3.370537974683544e-05, |
|
"loss": 0.0003, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 45.25, |
|
"learning_rate": 3.368164556962025e-05, |
|
"loss": 0.0003, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 45.5, |
|
"learning_rate": 3.365791139240506e-05, |
|
"loss": 0.0003, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 45.75, |
|
"learning_rate": 3.363417721518987e-05, |
|
"loss": 0.0003, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"learning_rate": 3.361044303797468e-05, |
|
"loss": 0.0003, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 46.25, |
|
"learning_rate": 3.358670886075949e-05, |
|
"loss": 0.0003, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 46.5, |
|
"learning_rate": 3.3562974683544305e-05, |
|
"loss": 0.0003, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 46.75, |
|
"learning_rate": 3.353924050632911e-05, |
|
"loss": 0.0003, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"learning_rate": 3.3515506329113924e-05, |
|
"loss": 0.0003, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 47.25, |
|
"learning_rate": 3.349177215189873e-05, |
|
"loss": 0.0003, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"learning_rate": 3.346803797468354e-05, |
|
"loss": 0.0003, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 47.75, |
|
"learning_rate": 3.344430379746835e-05, |
|
"loss": 0.0003, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"learning_rate": 3.342056962025316e-05, |
|
"loss": 0.0003, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 48.25, |
|
"learning_rate": 3.3396835443037975e-05, |
|
"loss": 0.0002, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 48.5, |
|
"learning_rate": 3.337310126582278e-05, |
|
"loss": 0.0002, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 48.75, |
|
"learning_rate": 3.3349367088607594e-05, |
|
"loss": 0.0002, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"learning_rate": 3.33256329113924e-05, |
|
"loss": 0.0003, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 49.25, |
|
"learning_rate": 3.330189873417721e-05, |
|
"loss": 0.0002, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 49.5, |
|
"learning_rate": 3.327816455696202e-05, |
|
"loss": 0.0002, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 49.75, |
|
"learning_rate": 3.325443037974683e-05, |
|
"loss": 0.0002, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 3.3230696202531645e-05, |
|
"loss": 0.0003, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_loss": 0.475699245929718, |
|
"eval_runtime": 80.4553, |
|
"eval_samples_per_second": 158.374, |
|
"eval_steps_per_second": 1.243, |
|
"eval_wer": 21.180826950132875, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 50.25, |
|
"learning_rate": 3.320696202531645e-05, |
|
"loss": 0.0002, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 50.5, |
|
"learning_rate": 3.3183227848101264e-05, |
|
"loss": 0.0002, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 50.75, |
|
"learning_rate": 3.315949367088607e-05, |
|
"loss": 0.0002, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"learning_rate": 3.313575949367088e-05, |
|
"loss": 0.0002, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 51.25, |
|
"learning_rate": 3.311202531645569e-05, |
|
"loss": 0.0002, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 51.5, |
|
"learning_rate": 3.30882911392405e-05, |
|
"loss": 0.0002, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 51.75, |
|
"learning_rate": 3.3064556962025315e-05, |
|
"loss": 0.0002, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"learning_rate": 3.304082278481012e-05, |
|
"loss": 0.0002, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 52.25, |
|
"learning_rate": 3.3017088607594934e-05, |
|
"loss": 0.0002, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 52.5, |
|
"learning_rate": 3.299335443037974e-05, |
|
"loss": 0.0002, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 52.75, |
|
"learning_rate": 3.296962025316455e-05, |
|
"loss": 0.0002, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 53.0, |
|
"learning_rate": 3.2945886075949366e-05, |
|
"loss": 0.0002, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 53.25, |
|
"learning_rate": 3.292215189873417e-05, |
|
"loss": 0.0002, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 53.5, |
|
"learning_rate": 3.2898417721518985e-05, |
|
"loss": 0.0002, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 53.75, |
|
"learning_rate": 3.287468354430379e-05, |
|
"loss": 0.0002, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 54.0, |
|
"learning_rate": 3.2850949367088604e-05, |
|
"loss": 0.0002, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 54.25, |
|
"learning_rate": 3.282721518987342e-05, |
|
"loss": 0.0002, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 54.5, |
|
"learning_rate": 3.280348101265823e-05, |
|
"loss": 0.0002, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 54.75, |
|
"learning_rate": 3.2779746835443036e-05, |
|
"loss": 0.0002, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"learning_rate": 3.275601265822785e-05, |
|
"loss": 0.0002, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 55.25, |
|
"learning_rate": 3.2732278481012655e-05, |
|
"loss": 0.0002, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 55.5, |
|
"learning_rate": 3.270854430379747e-05, |
|
"loss": 0.0002, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 55.75, |
|
"learning_rate": 3.2684810126582274e-05, |
|
"loss": 0.0002, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"learning_rate": 3.266107594936709e-05, |
|
"loss": 0.0002, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 56.25, |
|
"learning_rate": 3.26373417721519e-05, |
|
"loss": 0.0002, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 56.5, |
|
"learning_rate": 3.2613607594936706e-05, |
|
"loss": 0.0002, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 56.75, |
|
"learning_rate": 3.258987341772152e-05, |
|
"loss": 0.0002, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 57.0, |
|
"learning_rate": 3.2566139240506325e-05, |
|
"loss": 0.0002, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 57.25, |
|
"learning_rate": 3.254240506329114e-05, |
|
"loss": 0.0002, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 57.5, |
|
"learning_rate": 3.2518670886075944e-05, |
|
"loss": 0.0002, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 57.75, |
|
"learning_rate": 3.249493670886076e-05, |
|
"loss": 0.0002, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 58.0, |
|
"learning_rate": 3.247120253164557e-05, |
|
"loss": 0.0054, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 58.25, |
|
"learning_rate": 3.2447468354430376e-05, |
|
"loss": 0.0387, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 58.5, |
|
"learning_rate": 3.242373417721519e-05, |
|
"loss": 0.0443, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 58.75, |
|
"learning_rate": 3.2399999999999995e-05, |
|
"loss": 0.0406, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 59.0, |
|
"learning_rate": 3.237626582278481e-05, |
|
"loss": 0.0352, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 59.25, |
|
"learning_rate": 3.2352531645569614e-05, |
|
"loss": 0.0224, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 59.5, |
|
"learning_rate": 3.232879746835443e-05, |
|
"loss": 0.0213, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 59.75, |
|
"learning_rate": 3.230506329113924e-05, |
|
"loss": 0.0218, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"learning_rate": 3.2281329113924046e-05, |
|
"loss": 0.0206, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"eval_loss": 0.41719070076942444, |
|
"eval_runtime": 86.1367, |
|
"eval_samples_per_second": 147.928, |
|
"eval_steps_per_second": 1.161, |
|
"eval_wer": 22.275089885884007, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 60.25, |
|
"learning_rate": 3.225759493670886e-05, |
|
"loss": 0.0123, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 60.5, |
|
"learning_rate": 3.2233860759493665e-05, |
|
"loss": 0.0125, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 60.75, |
|
"learning_rate": 3.221012658227848e-05, |
|
"loss": 0.0126, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 61.0, |
|
"learning_rate": 3.2186392405063284e-05, |
|
"loss": 0.0126, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 61.25, |
|
"learning_rate": 3.2162658227848097e-05, |
|
"loss": 0.0078, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 61.5, |
|
"learning_rate": 3.213892405063291e-05, |
|
"loss": 0.0072, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 61.75, |
|
"learning_rate": 3.2115189873417716e-05, |
|
"loss": 0.0068, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 62.0, |
|
"learning_rate": 3.209145569620253e-05, |
|
"loss": 0.0073, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 62.25, |
|
"learning_rate": 3.206772151898734e-05, |
|
"loss": 0.0041, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"learning_rate": 3.2043987341772154e-05, |
|
"loss": 0.004, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 62.75, |
|
"learning_rate": 3.202025316455696e-05, |
|
"loss": 0.0039, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 63.0, |
|
"learning_rate": 3.199651898734177e-05, |
|
"loss": 0.0036, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 63.25, |
|
"learning_rate": 3.197278481012658e-05, |
|
"loss": 0.0022, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 63.5, |
|
"learning_rate": 3.194905063291139e-05, |
|
"loss": 0.0016, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 63.75, |
|
"learning_rate": 3.19253164556962e-05, |
|
"loss": 0.0019, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"learning_rate": 3.190158227848101e-05, |
|
"loss": 0.0017, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 64.25, |
|
"learning_rate": 3.1877848101265824e-05, |
|
"loss": 0.001, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 64.5, |
|
"learning_rate": 3.185411392405063e-05, |
|
"loss": 0.0015, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 64.75, |
|
"learning_rate": 3.183037974683544e-05, |
|
"loss": 0.001, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"learning_rate": 3.180664556962025e-05, |
|
"loss": 0.0013, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 65.25, |
|
"learning_rate": 3.178291139240506e-05, |
|
"loss": 0.0008, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 65.5, |
|
"learning_rate": 3.175917721518987e-05, |
|
"loss": 0.0008, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 65.75, |
|
"learning_rate": 3.173544303797468e-05, |
|
"loss": 0.001, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 66.0, |
|
"learning_rate": 3.1711708860759494e-05, |
|
"loss": 0.0008, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 66.25, |
|
"learning_rate": 3.16879746835443e-05, |
|
"loss": 0.0005, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 66.5, |
|
"learning_rate": 3.166424050632911e-05, |
|
"loss": 0.0007, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 66.75, |
|
"learning_rate": 3.164050632911392e-05, |
|
"loss": 0.0007, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 67.0, |
|
"learning_rate": 3.161677215189873e-05, |
|
"loss": 0.0006, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 67.25, |
|
"learning_rate": 3.159303797468354e-05, |
|
"loss": 0.0004, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 67.5, |
|
"learning_rate": 3.156930379746835e-05, |
|
"loss": 0.0004, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 67.75, |
|
"learning_rate": 3.1545569620253164e-05, |
|
"loss": 0.0004, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"learning_rate": 3.152183544303797e-05, |
|
"loss": 0.0004, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 68.25, |
|
"learning_rate": 3.149810126582278e-05, |
|
"loss": 0.0003, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 68.5, |
|
"learning_rate": 3.147436708860759e-05, |
|
"loss": 0.0003, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 68.75, |
|
"learning_rate": 3.14506329113924e-05, |
|
"loss": 0.0003, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 69.0, |
|
"learning_rate": 3.142689873417721e-05, |
|
"loss": 0.0003, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 69.25, |
|
"learning_rate": 3.140316455696202e-05, |
|
"loss": 0.0003, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 69.5, |
|
"learning_rate": 3.1379430379746834e-05, |
|
"loss": 0.0003, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 69.75, |
|
"learning_rate": 3.135569620253164e-05, |
|
"loss": 0.0003, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"learning_rate": 3.133196202531645e-05, |
|
"loss": 0.0003, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"eval_loss": 0.43740585446357727, |
|
"eval_runtime": 80.2604, |
|
"eval_samples_per_second": 158.758, |
|
"eval_steps_per_second": 1.246, |
|
"eval_wer": 19.513052993590748, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 70.25, |
|
"learning_rate": 3.130822784810126e-05, |
|
"loss": 0.0003, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 70.5, |
|
"learning_rate": 3.128449367088607e-05, |
|
"loss": 0.0003, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 70.75, |
|
"learning_rate": 3.1260759493670885e-05, |
|
"loss": 0.0003, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 71.0, |
|
"learning_rate": 3.12370253164557e-05, |
|
"loss": 0.0003, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 71.25, |
|
"learning_rate": 3.1213291139240504e-05, |
|
"loss": 0.0002, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 71.5, |
|
"learning_rate": 3.118955696202532e-05, |
|
"loss": 0.0002, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 71.75, |
|
"learning_rate": 3.116582278481012e-05, |
|
"loss": 0.0003, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"learning_rate": 3.1142088607594936e-05, |
|
"loss": 0.0002, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 72.25, |
|
"learning_rate": 3.111835443037974e-05, |
|
"loss": 0.0002, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 72.5, |
|
"learning_rate": 3.1094620253164555e-05, |
|
"loss": 0.0002, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 72.75, |
|
"learning_rate": 3.107088607594937e-05, |
|
"loss": 0.0002, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 73.0, |
|
"learning_rate": 3.1047151898734174e-05, |
|
"loss": 0.0002, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 73.25, |
|
"learning_rate": 3.102341772151899e-05, |
|
"loss": 0.0002, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 73.5, |
|
"learning_rate": 3.099968354430379e-05, |
|
"loss": 0.0002, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 73.75, |
|
"learning_rate": 3.0975949367088606e-05, |
|
"loss": 0.0002, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 74.0, |
|
"learning_rate": 3.095221518987341e-05, |
|
"loss": 0.0002, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 74.25, |
|
"learning_rate": 3.0928481012658225e-05, |
|
"loss": 0.0002, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 74.5, |
|
"learning_rate": 3.090474683544304e-05, |
|
"loss": 0.0002, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 74.75, |
|
"learning_rate": 3.0881012658227844e-05, |
|
"loss": 0.0002, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"learning_rate": 3.085727848101266e-05, |
|
"loss": 0.0002, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 75.25, |
|
"learning_rate": 3.083354430379746e-05, |
|
"loss": 0.0002, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 75.5, |
|
"learning_rate": 3.0809810126582276e-05, |
|
"loss": 0.0002, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 75.75, |
|
"learning_rate": 3.078607594936709e-05, |
|
"loss": 0.0002, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"learning_rate": 3.0762341772151895e-05, |
|
"loss": 0.0002, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 76.25, |
|
"learning_rate": 3.073860759493671e-05, |
|
"loss": 0.0002, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 76.5, |
|
"learning_rate": 3.0714873417721514e-05, |
|
"loss": 0.0002, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 76.75, |
|
"learning_rate": 3.069113924050633e-05, |
|
"loss": 0.0002, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 77.0, |
|
"learning_rate": 3.066740506329113e-05, |
|
"loss": 0.0002, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 77.25, |
|
"learning_rate": 3.0643670886075946e-05, |
|
"loss": 0.0002, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 77.5, |
|
"learning_rate": 3.061993670886076e-05, |
|
"loss": 0.0002, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 77.75, |
|
"learning_rate": 3.0596202531645565e-05, |
|
"loss": 0.0002, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 78.0, |
|
"learning_rate": 3.057246835443038e-05, |
|
"loss": 0.0002, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 78.25, |
|
"learning_rate": 3.0548734177215184e-05, |
|
"loss": 0.0002, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 78.5, |
|
"learning_rate": 3.0524999999999997e-05, |
|
"loss": 0.0002, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 78.75, |
|
"learning_rate": 3.050126582278481e-05, |
|
"loss": 0.0002, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 79.0, |
|
"learning_rate": 3.047753164556962e-05, |
|
"loss": 0.0002, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 79.25, |
|
"learning_rate": 3.045379746835443e-05, |
|
"loss": 0.0001, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 79.5, |
|
"learning_rate": 3.0430063291139238e-05, |
|
"loss": 0.0001, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 79.75, |
|
"learning_rate": 3.040632911392405e-05, |
|
"loss": 0.0001, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"learning_rate": 3.038259493670886e-05, |
|
"loss": 0.0002, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"eval_loss": 0.4546888768672943, |
|
"eval_runtime": 80.4032, |
|
"eval_samples_per_second": 158.476, |
|
"eval_steps_per_second": 1.244, |
|
"eval_wer": 19.50914491167735, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 80.25, |
|
"learning_rate": 3.035886075949367e-05, |
|
"loss": 0.0001, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 80.5, |
|
"learning_rate": 3.033512658227848e-05, |
|
"loss": 0.0001, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 80.75, |
|
"learning_rate": 3.031139240506329e-05, |
|
"loss": 0.0001, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 81.0, |
|
"learning_rate": 3.02876582278481e-05, |
|
"loss": 0.0001, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 81.25, |
|
"learning_rate": 3.0263924050632908e-05, |
|
"loss": 0.0001, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 81.5, |
|
"learning_rate": 3.024018987341772e-05, |
|
"loss": 0.0001, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 81.75, |
|
"learning_rate": 3.021645569620253e-05, |
|
"loss": 0.0001, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 82.0, |
|
"learning_rate": 3.019272151898734e-05, |
|
"loss": 0.0001, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 82.25, |
|
"learning_rate": 3.016898734177215e-05, |
|
"loss": 0.0001, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 82.5, |
|
"learning_rate": 3.014525316455696e-05, |
|
"loss": 0.0001, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 82.75, |
|
"learning_rate": 3.012151898734177e-05, |
|
"loss": 0.0001, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 83.0, |
|
"learning_rate": 3.0097784810126578e-05, |
|
"loss": 0.0001, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 83.25, |
|
"learning_rate": 3.007405063291139e-05, |
|
"loss": 0.0001, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 83.5, |
|
"learning_rate": 3.00503164556962e-05, |
|
"loss": 0.0001, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 83.75, |
|
"learning_rate": 3.002658227848101e-05, |
|
"loss": 0.0001, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"learning_rate": 3.000284810126582e-05, |
|
"loss": 0.0001, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 84.25, |
|
"learning_rate": 2.997911392405063e-05, |
|
"loss": 0.0001, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 84.5, |
|
"learning_rate": 2.9955379746835438e-05, |
|
"loss": 0.0001, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 84.75, |
|
"learning_rate": 2.993164556962025e-05, |
|
"loss": 0.0001, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"learning_rate": 2.990791139240506e-05, |
|
"loss": 0.0001, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 85.25, |
|
"learning_rate": 2.988417721518987e-05, |
|
"loss": 0.0001, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 85.5, |
|
"learning_rate": 2.986044303797468e-05, |
|
"loss": 0.0001, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 85.75, |
|
"learning_rate": 2.983670886075949e-05, |
|
"loss": 0.0001, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 86.0, |
|
"learning_rate": 2.98129746835443e-05, |
|
"loss": 0.0001, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 86.25, |
|
"learning_rate": 2.9789240506329108e-05, |
|
"loss": 0.0001, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 86.5, |
|
"learning_rate": 2.976550632911392e-05, |
|
"loss": 0.0001, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 86.75, |
|
"learning_rate": 2.974177215189873e-05, |
|
"loss": 0.0001, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 87.0, |
|
"learning_rate": 2.9718037974683544e-05, |
|
"loss": 0.0001, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 87.25, |
|
"learning_rate": 2.9694303797468353e-05, |
|
"loss": 0.0001, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 87.5, |
|
"learning_rate": 2.9670569620253163e-05, |
|
"loss": 0.0001, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 87.75, |
|
"learning_rate": 2.9646835443037972e-05, |
|
"loss": 0.0001, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"learning_rate": 2.9623101265822785e-05, |
|
"loss": 0.0001, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 88.25, |
|
"learning_rate": 2.9599367088607594e-05, |
|
"loss": 0.0001, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 88.5, |
|
"learning_rate": 2.9575632911392404e-05, |
|
"loss": 0.0001, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 88.75, |
|
"learning_rate": 2.9551898734177213e-05, |
|
"loss": 0.0001, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 89.0, |
|
"learning_rate": 2.9528164556962023e-05, |
|
"loss": 0.0001, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 89.25, |
|
"learning_rate": 2.9504430379746832e-05, |
|
"loss": 0.0001, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 89.5, |
|
"learning_rate": 2.9480696202531645e-05, |
|
"loss": 0.0001, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 89.75, |
|
"learning_rate": 2.9456962025316455e-05, |
|
"loss": 0.0001, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"learning_rate": 2.9433227848101264e-05, |
|
"loss": 0.0001, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"eval_loss": 0.4697396755218506, |
|
"eval_runtime": 80.422, |
|
"eval_samples_per_second": 158.439, |
|
"eval_steps_per_second": 1.243, |
|
"eval_wer": 19.5062138502423, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 90.25, |
|
"learning_rate": 2.9409493670886074e-05, |
|
"loss": 0.0001, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 90.5, |
|
"learning_rate": 2.9385759493670883e-05, |
|
"loss": 0.0001, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 90.75, |
|
"learning_rate": 2.9362025316455693e-05, |
|
"loss": 0.0001, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 91.0, |
|
"learning_rate": 2.9338291139240502e-05, |
|
"loss": 0.0001, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 91.25, |
|
"learning_rate": 2.9314556962025315e-05, |
|
"loss": 0.0001, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 91.5, |
|
"learning_rate": 2.9290822784810125e-05, |
|
"loss": 0.0001, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 91.75, |
|
"learning_rate": 2.9267088607594934e-05, |
|
"loss": 0.0001, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"learning_rate": 2.9243354430379744e-05, |
|
"loss": 0.0001, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 92.25, |
|
"learning_rate": 2.9219620253164553e-05, |
|
"loss": 0.0001, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 92.5, |
|
"learning_rate": 2.9195886075949363e-05, |
|
"loss": 0.0001, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 92.75, |
|
"learning_rate": 2.9172151898734172e-05, |
|
"loss": 0.0001, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 93.0, |
|
"learning_rate": 2.9148417721518985e-05, |
|
"loss": 0.0001, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 93.25, |
|
"learning_rate": 2.9124683544303795e-05, |
|
"loss": 0.0001, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 93.5, |
|
"learning_rate": 2.9100949367088604e-05, |
|
"loss": 0.0001, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 93.75, |
|
"learning_rate": 2.9077215189873414e-05, |
|
"loss": 0.0001, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 94.0, |
|
"learning_rate": 2.9053481012658223e-05, |
|
"loss": 0.0001, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 94.25, |
|
"learning_rate": 2.9029746835443033e-05, |
|
"loss": 0.0001, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 94.5, |
|
"learning_rate": 2.9006012658227842e-05, |
|
"loss": 0.0001, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 94.75, |
|
"learning_rate": 2.8982278481012655e-05, |
|
"loss": 0.0001, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"learning_rate": 2.8958544303797465e-05, |
|
"loss": 0.0001, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 95.25, |
|
"learning_rate": 2.8934810126582278e-05, |
|
"loss": 0.0001, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 95.5, |
|
"learning_rate": 2.8911075949367087e-05, |
|
"loss": 0.0001, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 95.75, |
|
"learning_rate": 2.8887341772151897e-05, |
|
"loss": 0.0001, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"learning_rate": 2.886360759493671e-05, |
|
"loss": 0.0001, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 96.25, |
|
"learning_rate": 2.883987341772152e-05, |
|
"loss": 0.0001, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 96.5, |
|
"learning_rate": 2.881613924050633e-05, |
|
"loss": 0.0001, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 96.75, |
|
"learning_rate": 2.8792405063291138e-05, |
|
"loss": 0.0001, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 97.0, |
|
"learning_rate": 2.8768670886075948e-05, |
|
"loss": 0.0001, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 97.25, |
|
"learning_rate": 2.8744936708860757e-05, |
|
"loss": 0.0001, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 97.5, |
|
"learning_rate": 2.8721202531645567e-05, |
|
"loss": 0.0001, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 97.75, |
|
"learning_rate": 2.869746835443038e-05, |
|
"loss": 0.0001, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 98.0, |
|
"learning_rate": 2.867373417721519e-05, |
|
"loss": 0.0001, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 98.25, |
|
"learning_rate": 2.865e-05, |
|
"loss": 0.0001, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 98.5, |
|
"learning_rate": 2.8626265822784808e-05, |
|
"loss": 0.0001, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 98.75, |
|
"learning_rate": 2.8602531645569617e-05, |
|
"loss": 0.0001, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 99.0, |
|
"learning_rate": 2.8578797468354427e-05, |
|
"loss": 0.0001, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 99.25, |
|
"learning_rate": 2.8555063291139236e-05, |
|
"loss": 0.0001, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 99.5, |
|
"learning_rate": 2.853132911392405e-05, |
|
"loss": 0.0001, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 99.75, |
|
"learning_rate": 2.850759493670886e-05, |
|
"loss": 0.0001, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"learning_rate": 2.848386075949367e-05, |
|
"loss": 0.0001, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"eval_loss": 0.48529908061027527, |
|
"eval_runtime": 80.5352, |
|
"eval_samples_per_second": 158.217, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.51989213693919, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 100.25, |
|
"learning_rate": 2.8460126582278478e-05, |
|
"loss": 0.0001, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 100.5, |
|
"learning_rate": 2.8436392405063287e-05, |
|
"loss": 0.0001, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 100.75, |
|
"learning_rate": 2.8412658227848097e-05, |
|
"loss": 0.0001, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 101.0, |
|
"learning_rate": 2.8388924050632906e-05, |
|
"loss": 0.0001, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 101.25, |
|
"learning_rate": 2.836518987341772e-05, |
|
"loss": 0.0001, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 101.5, |
|
"learning_rate": 2.834145569620253e-05, |
|
"loss": 0.0001, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 101.75, |
|
"learning_rate": 2.831772151898734e-05, |
|
"loss": 0.0001, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 102.0, |
|
"learning_rate": 2.8293987341772148e-05, |
|
"loss": 0.0001, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 102.25, |
|
"learning_rate": 2.8270253164556957e-05, |
|
"loss": 0.0001, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 102.5, |
|
"learning_rate": 2.8246518987341767e-05, |
|
"loss": 0.0001, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 102.75, |
|
"learning_rate": 2.8222784810126576e-05, |
|
"loss": 0.0001, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 103.0, |
|
"learning_rate": 2.819905063291139e-05, |
|
"loss": 0.0001, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 103.25, |
|
"learning_rate": 2.81753164556962e-05, |
|
"loss": 0.0001, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 103.5, |
|
"learning_rate": 2.815158227848101e-05, |
|
"loss": 0.0001, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 103.75, |
|
"learning_rate": 2.812784810126582e-05, |
|
"loss": 0.0001, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 104.0, |
|
"learning_rate": 2.810411392405063e-05, |
|
"loss": 0.0, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 104.25, |
|
"learning_rate": 2.8080379746835444e-05, |
|
"loss": 0.0, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 104.5, |
|
"learning_rate": 2.8056645569620253e-05, |
|
"loss": 0.0, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 104.75, |
|
"learning_rate": 2.8032911392405063e-05, |
|
"loss": 0.0, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 105.0, |
|
"learning_rate": 2.8009177215189872e-05, |
|
"loss": 0.0001, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 105.25, |
|
"learning_rate": 2.798544303797468e-05, |
|
"loss": 0.0, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 105.5, |
|
"learning_rate": 2.796170886075949e-05, |
|
"loss": 0.0, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 105.75, |
|
"learning_rate": 2.79379746835443e-05, |
|
"loss": 0.0001, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 106.0, |
|
"learning_rate": 2.7914240506329113e-05, |
|
"loss": 0.0001, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 106.25, |
|
"learning_rate": 2.7890506329113923e-05, |
|
"loss": 0.0, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 106.5, |
|
"learning_rate": 2.7866772151898732e-05, |
|
"loss": 0.0, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 106.75, |
|
"learning_rate": 2.7843037974683542e-05, |
|
"loss": 0.0001, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 107.0, |
|
"learning_rate": 2.781930379746835e-05, |
|
"loss": 0.0, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 107.25, |
|
"learning_rate": 2.779556962025316e-05, |
|
"loss": 0.0, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 107.5, |
|
"learning_rate": 2.777183544303797e-05, |
|
"loss": 0.0, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 107.75, |
|
"learning_rate": 2.7748101265822783e-05, |
|
"loss": 0.0001, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 108.0, |
|
"learning_rate": 2.7724367088607593e-05, |
|
"loss": 0.0001, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 108.25, |
|
"learning_rate": 2.7700632911392402e-05, |
|
"loss": 0.0001, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 108.5, |
|
"learning_rate": 2.7676898734177212e-05, |
|
"loss": 0.0, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 108.75, |
|
"learning_rate": 2.765316455696202e-05, |
|
"loss": 0.0001, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 109.0, |
|
"learning_rate": 2.762943037974683e-05, |
|
"loss": 0.0, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 109.25, |
|
"learning_rate": 2.7605696202531644e-05, |
|
"loss": 0.0, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 109.5, |
|
"learning_rate": 2.7581962025316453e-05, |
|
"loss": 0.0, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 109.75, |
|
"learning_rate": 2.7558227848101263e-05, |
|
"loss": 0.0, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"learning_rate": 2.7534493670886072e-05, |
|
"loss": 0.0001, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"eval_loss": 0.5008699297904968, |
|
"eval_runtime": 80.5626, |
|
"eval_samples_per_second": 158.163, |
|
"eval_steps_per_second": 1.241, |
|
"eval_wer": 19.568743160856652, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 110.25, |
|
"learning_rate": 2.7510759493670882e-05, |
|
"loss": 0.0, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 110.5, |
|
"learning_rate": 2.748702531645569e-05, |
|
"loss": 0.0, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 110.75, |
|
"learning_rate": 2.74632911392405e-05, |
|
"loss": 0.0, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 111.0, |
|
"learning_rate": 2.7439556962025314e-05, |
|
"loss": 0.0, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 111.25, |
|
"learning_rate": 2.7415822784810123e-05, |
|
"loss": 0.0, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 111.5, |
|
"learning_rate": 2.7392088607594933e-05, |
|
"loss": 0.0, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 111.75, |
|
"learning_rate": 2.7368354430379746e-05, |
|
"loss": 0.0, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 112.0, |
|
"learning_rate": 2.7344620253164555e-05, |
|
"loss": 0.0, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 112.25, |
|
"learning_rate": 2.7320886075949368e-05, |
|
"loss": 0.0, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 112.5, |
|
"learning_rate": 2.7297151898734178e-05, |
|
"loss": 0.0, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 112.75, |
|
"learning_rate": 2.7273417721518987e-05, |
|
"loss": 0.0, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 113.0, |
|
"learning_rate": 2.7249683544303797e-05, |
|
"loss": 0.0, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 113.25, |
|
"learning_rate": 2.7225949367088606e-05, |
|
"loss": 0.0, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 113.5, |
|
"learning_rate": 2.7202215189873416e-05, |
|
"loss": 0.0, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 113.75, |
|
"learning_rate": 2.7178481012658225e-05, |
|
"loss": 0.0, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 114.0, |
|
"learning_rate": 2.7154746835443038e-05, |
|
"loss": 0.0, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 114.25, |
|
"learning_rate": 2.7131012658227848e-05, |
|
"loss": 0.0, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 114.5, |
|
"learning_rate": 2.7107278481012657e-05, |
|
"loss": 0.0, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 114.75, |
|
"learning_rate": 2.7083544303797467e-05, |
|
"loss": 0.0, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 115.0, |
|
"learning_rate": 2.7059810126582276e-05, |
|
"loss": 0.0001, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 115.25, |
|
"learning_rate": 2.7036075949367086e-05, |
|
"loss": 0.0, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 115.5, |
|
"learning_rate": 2.7012341772151895e-05, |
|
"loss": 0.0001, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 115.75, |
|
"learning_rate": 2.6988607594936708e-05, |
|
"loss": 0.0, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 116.0, |
|
"learning_rate": 2.6964873417721517e-05, |
|
"loss": 0.0, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 116.25, |
|
"learning_rate": 2.6941139240506327e-05, |
|
"loss": 0.0, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 116.5, |
|
"learning_rate": 2.6917405063291136e-05, |
|
"loss": 0.0, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 116.75, |
|
"learning_rate": 2.6893670886075946e-05, |
|
"loss": 0.0, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 117.0, |
|
"learning_rate": 2.6869936708860755e-05, |
|
"loss": 0.0, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 117.25, |
|
"learning_rate": 2.6846202531645565e-05, |
|
"loss": 0.0, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 117.5, |
|
"learning_rate": 2.6822468354430378e-05, |
|
"loss": 0.0, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 117.75, |
|
"learning_rate": 2.6798734177215187e-05, |
|
"loss": 0.0, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 118.0, |
|
"learning_rate": 2.6774999999999997e-05, |
|
"loss": 0.0, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 118.25, |
|
"learning_rate": 2.6751265822784806e-05, |
|
"loss": 0.0, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 118.5, |
|
"learning_rate": 2.6727531645569616e-05, |
|
"loss": 0.0, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 118.75, |
|
"learning_rate": 2.6703797468354425e-05, |
|
"loss": 0.0, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 119.0, |
|
"learning_rate": 2.6680063291139235e-05, |
|
"loss": 0.0, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 119.25, |
|
"learning_rate": 2.6656329113924048e-05, |
|
"loss": 0.0, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 119.5, |
|
"learning_rate": 2.6632594936708857e-05, |
|
"loss": 0.0, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 119.75, |
|
"learning_rate": 2.660886075949367e-05, |
|
"loss": 0.0, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"learning_rate": 2.658512658227848e-05, |
|
"loss": 0.0, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"eval_loss": 0.5174862146377563, |
|
"eval_runtime": 80.4061, |
|
"eval_samples_per_second": 158.471, |
|
"eval_steps_per_second": 1.244, |
|
"eval_wer": 19.65862904486478, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 120.25, |
|
"learning_rate": 2.656139240506329e-05, |
|
"loss": 0.0, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 120.5, |
|
"learning_rate": 2.6537658227848102e-05, |
|
"loss": 0.0, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 120.75, |
|
"learning_rate": 2.651392405063291e-05, |
|
"loss": 0.0, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 121.0, |
|
"learning_rate": 2.649018987341772e-05, |
|
"loss": 0.0, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 121.25, |
|
"learning_rate": 2.646645569620253e-05, |
|
"loss": 0.0, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 121.5, |
|
"learning_rate": 2.644272151898734e-05, |
|
"loss": 0.0, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 121.75, |
|
"learning_rate": 2.641898734177215e-05, |
|
"loss": 0.0, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 122.0, |
|
"learning_rate": 2.639525316455696e-05, |
|
"loss": 0.0, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 122.25, |
|
"learning_rate": 2.6371518987341772e-05, |
|
"loss": 0.0, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 122.5, |
|
"learning_rate": 2.634778481012658e-05, |
|
"loss": 0.0, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 122.75, |
|
"learning_rate": 2.632405063291139e-05, |
|
"loss": 0.0, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 123.0, |
|
"learning_rate": 2.63003164556962e-05, |
|
"loss": 0.0, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 123.25, |
|
"learning_rate": 2.627658227848101e-05, |
|
"loss": 0.0, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 123.5, |
|
"learning_rate": 2.625284810126582e-05, |
|
"loss": 0.0, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 123.75, |
|
"learning_rate": 2.622911392405063e-05, |
|
"loss": 0.0, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 124.0, |
|
"learning_rate": 2.6205379746835442e-05, |
|
"loss": 0.0, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 124.25, |
|
"learning_rate": 2.618164556962025e-05, |
|
"loss": 0.0, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 124.5, |
|
"learning_rate": 2.615791139240506e-05, |
|
"loss": 0.0, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 124.75, |
|
"learning_rate": 2.613417721518987e-05, |
|
"loss": 0.0, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"learning_rate": 2.611044303797468e-05, |
|
"loss": 0.0, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 125.25, |
|
"learning_rate": 2.608670886075949e-05, |
|
"loss": 0.0, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 125.5, |
|
"learning_rate": 2.60629746835443e-05, |
|
"loss": 0.0, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 125.75, |
|
"learning_rate": 2.6039240506329112e-05, |
|
"loss": 0.0, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 126.0, |
|
"learning_rate": 2.601550632911392e-05, |
|
"loss": 0.0, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 126.25, |
|
"learning_rate": 2.599177215189873e-05, |
|
"loss": 0.0, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 126.5, |
|
"learning_rate": 2.596803797468354e-05, |
|
"loss": 0.0, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 126.75, |
|
"learning_rate": 2.594430379746835e-05, |
|
"loss": 0.0, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 127.0, |
|
"learning_rate": 2.592056962025316e-05, |
|
"loss": 0.0, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 127.25, |
|
"learning_rate": 2.5896835443037972e-05, |
|
"loss": 0.0, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 127.5, |
|
"learning_rate": 2.5873101265822782e-05, |
|
"loss": 0.0, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 127.75, |
|
"learning_rate": 2.584936708860759e-05, |
|
"loss": 0.0, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 128.0, |
|
"learning_rate": 2.5825632911392404e-05, |
|
"loss": 0.0, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 128.25, |
|
"learning_rate": 2.5801898734177214e-05, |
|
"loss": 0.0, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 128.5, |
|
"learning_rate": 2.5778164556962023e-05, |
|
"loss": 0.0, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 128.75, |
|
"learning_rate": 2.5754430379746836e-05, |
|
"loss": 0.0, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 129.0, |
|
"learning_rate": 2.5730696202531646e-05, |
|
"loss": 0.0, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 129.25, |
|
"learning_rate": 2.5706962025316455e-05, |
|
"loss": 0.0, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 129.5, |
|
"learning_rate": 2.5683227848101265e-05, |
|
"loss": 0.0, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 129.75, |
|
"learning_rate": 2.5659493670886074e-05, |
|
"loss": 0.0, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 130.0, |
|
"learning_rate": 2.5635759493670884e-05, |
|
"loss": 0.0, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 130.0, |
|
"eval_loss": 0.5348048806190491, |
|
"eval_runtime": 80.5003, |
|
"eval_samples_per_second": 158.285, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.77294044083164, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 130.25, |
|
"learning_rate": 2.5612025316455693e-05, |
|
"loss": 0.0, |
|
"step": 13025 |
|
}, |
|
{ |
|
"epoch": 130.5, |
|
"learning_rate": 2.5588291139240506e-05, |
|
"loss": 0.0, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 130.75, |
|
"learning_rate": 2.5564556962025316e-05, |
|
"loss": 0.0, |
|
"step": 13075 |
|
}, |
|
{ |
|
"epoch": 131.0, |
|
"learning_rate": 2.5540822784810125e-05, |
|
"loss": 0.0, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 131.25, |
|
"learning_rate": 2.5517088607594935e-05, |
|
"loss": 0.0, |
|
"step": 13125 |
|
}, |
|
{ |
|
"epoch": 131.5, |
|
"learning_rate": 2.5493354430379744e-05, |
|
"loss": 0.0, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 131.75, |
|
"learning_rate": 2.5469620253164554e-05, |
|
"loss": 0.0, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 132.0, |
|
"learning_rate": 2.5445886075949367e-05, |
|
"loss": 0.0, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 132.25, |
|
"learning_rate": 2.5422151898734176e-05, |
|
"loss": 0.0, |
|
"step": 13225 |
|
}, |
|
{ |
|
"epoch": 132.5, |
|
"learning_rate": 2.5398417721518986e-05, |
|
"loss": 0.0, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 132.75, |
|
"learning_rate": 2.5374683544303795e-05, |
|
"loss": 0.0, |
|
"step": 13275 |
|
}, |
|
{ |
|
"epoch": 133.0, |
|
"learning_rate": 2.5350949367088605e-05, |
|
"loss": 0.0, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 133.25, |
|
"learning_rate": 2.5327215189873414e-05, |
|
"loss": 0.0, |
|
"step": 13325 |
|
}, |
|
{ |
|
"epoch": 133.5, |
|
"learning_rate": 2.5303481012658224e-05, |
|
"loss": 0.0, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 133.75, |
|
"learning_rate": 2.5279746835443036e-05, |
|
"loss": 0.0, |
|
"step": 13375 |
|
}, |
|
{ |
|
"epoch": 134.0, |
|
"learning_rate": 2.5256012658227846e-05, |
|
"loss": 0.0, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 134.25, |
|
"learning_rate": 2.5232278481012656e-05, |
|
"loss": 0.0, |
|
"step": 13425 |
|
}, |
|
{ |
|
"epoch": 134.5, |
|
"learning_rate": 2.5208544303797465e-05, |
|
"loss": 0.0, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 134.75, |
|
"learning_rate": 2.5184810126582275e-05, |
|
"loss": 0.0, |
|
"step": 13475 |
|
}, |
|
{ |
|
"epoch": 135.0, |
|
"learning_rate": 2.5161075949367084e-05, |
|
"loss": 0.0, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 135.25, |
|
"learning_rate": 2.5137341772151894e-05, |
|
"loss": 0.0, |
|
"step": 13525 |
|
}, |
|
{ |
|
"epoch": 135.5, |
|
"learning_rate": 2.5113607594936706e-05, |
|
"loss": 0.0, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 135.75, |
|
"learning_rate": 2.5089873417721516e-05, |
|
"loss": 0.0, |
|
"step": 13575 |
|
}, |
|
{ |
|
"epoch": 136.0, |
|
"learning_rate": 2.5066139240506325e-05, |
|
"loss": 0.0, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 136.25, |
|
"learning_rate": 2.504240506329114e-05, |
|
"loss": 0.0, |
|
"step": 13625 |
|
}, |
|
{ |
|
"epoch": 136.5, |
|
"learning_rate": 2.5018670886075948e-05, |
|
"loss": 0.0, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 136.75, |
|
"learning_rate": 2.499493670886076e-05, |
|
"loss": 0.0, |
|
"step": 13675 |
|
}, |
|
{ |
|
"epoch": 137.0, |
|
"learning_rate": 2.497120253164557e-05, |
|
"loss": 0.0, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 137.25, |
|
"learning_rate": 2.494746835443038e-05, |
|
"loss": 0.0, |
|
"step": 13725 |
|
}, |
|
{ |
|
"epoch": 137.5, |
|
"learning_rate": 2.492373417721519e-05, |
|
"loss": 0.0, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 137.75, |
|
"learning_rate": 2.49e-05, |
|
"loss": 0.0, |
|
"step": 13775 |
|
}, |
|
{ |
|
"epoch": 138.0, |
|
"learning_rate": 2.4876265822784808e-05, |
|
"loss": 0.0, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 138.25, |
|
"learning_rate": 2.4852531645569618e-05, |
|
"loss": 0.0, |
|
"step": 13825 |
|
}, |
|
{ |
|
"epoch": 138.5, |
|
"learning_rate": 2.482879746835443e-05, |
|
"loss": 0.0, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 138.75, |
|
"learning_rate": 2.480506329113924e-05, |
|
"loss": 0.0, |
|
"step": 13875 |
|
}, |
|
{ |
|
"epoch": 139.0, |
|
"learning_rate": 2.478132911392405e-05, |
|
"loss": 0.0, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 139.25, |
|
"learning_rate": 2.475759493670886e-05, |
|
"loss": 0.0, |
|
"step": 13925 |
|
}, |
|
{ |
|
"epoch": 139.5, |
|
"learning_rate": 2.473386075949367e-05, |
|
"loss": 0.0, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 139.75, |
|
"learning_rate": 2.4710126582278478e-05, |
|
"loss": 0.0, |
|
"step": 13975 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"learning_rate": 2.4686392405063288e-05, |
|
"loss": 0.0, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"eval_loss": 0.553089439868927, |
|
"eval_runtime": 80.4755, |
|
"eval_samples_per_second": 158.334, |
|
"eval_steps_per_second": 1.243, |
|
"eval_wer": 19.784664686571833, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 140.25, |
|
"learning_rate": 2.46626582278481e-05, |
|
"loss": 0.0, |
|
"step": 14025 |
|
}, |
|
{ |
|
"epoch": 140.5, |
|
"learning_rate": 2.463892405063291e-05, |
|
"loss": 0.0, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 140.75, |
|
"learning_rate": 2.461518987341772e-05, |
|
"loss": 0.0, |
|
"step": 14075 |
|
}, |
|
{ |
|
"epoch": 141.0, |
|
"learning_rate": 2.459145569620253e-05, |
|
"loss": 0.0, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 141.25, |
|
"learning_rate": 2.456772151898734e-05, |
|
"loss": 0.0, |
|
"step": 14125 |
|
}, |
|
{ |
|
"epoch": 141.5, |
|
"learning_rate": 2.4543987341772148e-05, |
|
"loss": 0.0, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 141.75, |
|
"learning_rate": 2.4520253164556958e-05, |
|
"loss": 0.0, |
|
"step": 14175 |
|
}, |
|
{ |
|
"epoch": 142.0, |
|
"learning_rate": 2.449651898734177e-05, |
|
"loss": 0.0, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 142.25, |
|
"learning_rate": 2.447278481012658e-05, |
|
"loss": 0.0, |
|
"step": 14225 |
|
}, |
|
{ |
|
"epoch": 142.5, |
|
"learning_rate": 2.444905063291139e-05, |
|
"loss": 0.0, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 142.75, |
|
"learning_rate": 2.44253164556962e-05, |
|
"loss": 0.0, |
|
"step": 14275 |
|
}, |
|
{ |
|
"epoch": 143.0, |
|
"learning_rate": 2.440158227848101e-05, |
|
"loss": 0.0, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 143.25, |
|
"learning_rate": 2.4377848101265818e-05, |
|
"loss": 0.0, |
|
"step": 14325 |
|
}, |
|
{ |
|
"epoch": 143.5, |
|
"learning_rate": 2.4354113924050628e-05, |
|
"loss": 0.0, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 143.75, |
|
"learning_rate": 2.4333227848101262e-05, |
|
"loss": 0.0107, |
|
"step": 14375 |
|
}, |
|
{ |
|
"epoch": 144.0, |
|
"learning_rate": 2.4309493670886075e-05, |
|
"loss": 0.0398, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 144.25, |
|
"learning_rate": 2.4285759493670885e-05, |
|
"loss": 0.0325, |
|
"step": 14425 |
|
}, |
|
{ |
|
"epoch": 144.5, |
|
"learning_rate": 2.4262025316455694e-05, |
|
"loss": 0.024, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 144.75, |
|
"learning_rate": 2.4238291139240504e-05, |
|
"loss": 0.0209, |
|
"step": 14475 |
|
}, |
|
{ |
|
"epoch": 145.0, |
|
"learning_rate": 2.4214556962025313e-05, |
|
"loss": 0.0178, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 145.25, |
|
"learning_rate": 2.4190822784810123e-05, |
|
"loss": 0.0087, |
|
"step": 14525 |
|
}, |
|
{ |
|
"epoch": 145.5, |
|
"learning_rate": 2.4167088607594932e-05, |
|
"loss": 0.0102, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 145.75, |
|
"learning_rate": 2.4143354430379745e-05, |
|
"loss": 0.0085, |
|
"step": 14575 |
|
}, |
|
{ |
|
"epoch": 146.0, |
|
"learning_rate": 2.4119620253164555e-05, |
|
"loss": 0.007, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 146.25, |
|
"learning_rate": 2.4095886075949364e-05, |
|
"loss": 0.0046, |
|
"step": 14625 |
|
}, |
|
{ |
|
"epoch": 146.5, |
|
"learning_rate": 2.4072151898734174e-05, |
|
"loss": 0.005, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 146.75, |
|
"learning_rate": 2.4048417721518983e-05, |
|
"loss": 0.0035, |
|
"step": 14675 |
|
}, |
|
{ |
|
"epoch": 147.0, |
|
"learning_rate": 2.4024683544303793e-05, |
|
"loss": 0.0035, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 147.25, |
|
"learning_rate": 2.4000949367088602e-05, |
|
"loss": 0.0021, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 147.5, |
|
"learning_rate": 2.3977215189873415e-05, |
|
"loss": 0.002, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 147.75, |
|
"learning_rate": 2.3953481012658225e-05, |
|
"loss": 0.0016, |
|
"step": 14775 |
|
}, |
|
{ |
|
"epoch": 148.0, |
|
"learning_rate": 2.3929746835443034e-05, |
|
"loss": 0.0015, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 148.25, |
|
"learning_rate": 2.3906012658227847e-05, |
|
"loss": 0.0006, |
|
"step": 14825 |
|
}, |
|
{ |
|
"epoch": 148.5, |
|
"learning_rate": 2.3882278481012657e-05, |
|
"loss": 0.0005, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 148.75, |
|
"learning_rate": 2.385854430379747e-05, |
|
"loss": 0.0004, |
|
"step": 14875 |
|
}, |
|
{ |
|
"epoch": 149.0, |
|
"learning_rate": 2.383481012658228e-05, |
|
"loss": 0.0006, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 149.25, |
|
"learning_rate": 2.381107594936709e-05, |
|
"loss": 0.0003, |
|
"step": 14925 |
|
}, |
|
{ |
|
"epoch": 149.5, |
|
"learning_rate": 2.3787341772151898e-05, |
|
"loss": 0.0003, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 149.75, |
|
"learning_rate": 2.3763607594936707e-05, |
|
"loss": 0.0003, |
|
"step": 14975 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"learning_rate": 2.3739873417721517e-05, |
|
"loss": 0.0002, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"eval_loss": 0.4625629484653473, |
|
"eval_runtime": 81.6484, |
|
"eval_samples_per_second": 156.059, |
|
"eval_steps_per_second": 1.225, |
|
"eval_wer": 19.472995153978427, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 150.25, |
|
"learning_rate": 2.3716139240506327e-05, |
|
"loss": 0.0002, |
|
"step": 15025 |
|
}, |
|
{ |
|
"epoch": 150.5, |
|
"learning_rate": 2.369240506329114e-05, |
|
"loss": 0.0002, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 150.75, |
|
"learning_rate": 2.366867088607595e-05, |
|
"loss": 0.0002, |
|
"step": 15075 |
|
}, |
|
{ |
|
"epoch": 151.0, |
|
"learning_rate": 2.364493670886076e-05, |
|
"loss": 0.0002, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 151.25, |
|
"learning_rate": 2.3621202531645568e-05, |
|
"loss": 0.0001, |
|
"step": 15125 |
|
}, |
|
{ |
|
"epoch": 151.5, |
|
"learning_rate": 2.3597468354430377e-05, |
|
"loss": 0.0001, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 151.75, |
|
"learning_rate": 2.3573734177215187e-05, |
|
"loss": 0.0001, |
|
"step": 15175 |
|
}, |
|
{ |
|
"epoch": 152.0, |
|
"learning_rate": 2.3549999999999996e-05, |
|
"loss": 0.0002, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 152.25, |
|
"learning_rate": 2.352626582278481e-05, |
|
"loss": 0.0001, |
|
"step": 15225 |
|
}, |
|
{ |
|
"epoch": 152.5, |
|
"learning_rate": 2.350253164556962e-05, |
|
"loss": 0.0001, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 152.75, |
|
"learning_rate": 2.347879746835443e-05, |
|
"loss": 0.0001, |
|
"step": 15275 |
|
}, |
|
{ |
|
"epoch": 153.0, |
|
"learning_rate": 2.3455063291139238e-05, |
|
"loss": 0.0001, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 153.25, |
|
"learning_rate": 2.3431329113924047e-05, |
|
"loss": 0.0001, |
|
"step": 15325 |
|
}, |
|
{ |
|
"epoch": 153.5, |
|
"learning_rate": 2.3407594936708857e-05, |
|
"loss": 0.0001, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 153.75, |
|
"learning_rate": 2.3383860759493666e-05, |
|
"loss": 0.0001, |
|
"step": 15375 |
|
}, |
|
{ |
|
"epoch": 154.0, |
|
"learning_rate": 2.336012658227848e-05, |
|
"loss": 0.0001, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 154.25, |
|
"learning_rate": 2.333639240506329e-05, |
|
"loss": 0.0001, |
|
"step": 15425 |
|
}, |
|
{ |
|
"epoch": 154.5, |
|
"learning_rate": 2.3312658227848098e-05, |
|
"loss": 0.0001, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 154.75, |
|
"learning_rate": 2.3288924050632908e-05, |
|
"loss": 0.0001, |
|
"step": 15475 |
|
}, |
|
{ |
|
"epoch": 155.0, |
|
"learning_rate": 2.3265189873417717e-05, |
|
"loss": 0.0001, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 155.25, |
|
"learning_rate": 2.3241455696202527e-05, |
|
"loss": 0.0001, |
|
"step": 15525 |
|
}, |
|
{ |
|
"epoch": 155.5, |
|
"learning_rate": 2.321772151898734e-05, |
|
"loss": 0.0001, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 155.75, |
|
"learning_rate": 2.319398734177215e-05, |
|
"loss": 0.0001, |
|
"step": 15575 |
|
}, |
|
{ |
|
"epoch": 156.0, |
|
"learning_rate": 2.317025316455696e-05, |
|
"loss": 0.0001, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 156.25, |
|
"learning_rate": 2.3146518987341768e-05, |
|
"loss": 0.0001, |
|
"step": 15625 |
|
}, |
|
{ |
|
"epoch": 156.5, |
|
"learning_rate": 2.312278481012658e-05, |
|
"loss": 0.0001, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 156.75, |
|
"learning_rate": 2.309905063291139e-05, |
|
"loss": 0.0001, |
|
"step": 15675 |
|
}, |
|
{ |
|
"epoch": 157.0, |
|
"learning_rate": 2.3075316455696204e-05, |
|
"loss": 0.0001, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 157.25, |
|
"learning_rate": 2.3051582278481013e-05, |
|
"loss": 0.0001, |
|
"step": 15725 |
|
}, |
|
{ |
|
"epoch": 157.5, |
|
"learning_rate": 2.3027848101265823e-05, |
|
"loss": 0.0001, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 157.75, |
|
"learning_rate": 2.3004113924050632e-05, |
|
"loss": 0.0001, |
|
"step": 15775 |
|
}, |
|
{ |
|
"epoch": 158.0, |
|
"learning_rate": 2.298037974683544e-05, |
|
"loss": 0.0001, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 158.25, |
|
"learning_rate": 2.295664556962025e-05, |
|
"loss": 0.0001, |
|
"step": 15825 |
|
}, |
|
{ |
|
"epoch": 158.5, |
|
"learning_rate": 2.293291139240506e-05, |
|
"loss": 0.0001, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 158.75, |
|
"learning_rate": 2.2909177215189873e-05, |
|
"loss": 0.0001, |
|
"step": 15875 |
|
}, |
|
{ |
|
"epoch": 159.0, |
|
"learning_rate": 2.2885443037974683e-05, |
|
"loss": 0.0001, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 159.25, |
|
"learning_rate": 2.2861708860759492e-05, |
|
"loss": 0.0001, |
|
"step": 15925 |
|
}, |
|
{ |
|
"epoch": 159.5, |
|
"learning_rate": 2.2837974683544302e-05, |
|
"loss": 0.0001, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 159.75, |
|
"learning_rate": 2.281424050632911e-05, |
|
"loss": 0.0001, |
|
"step": 15975 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"learning_rate": 2.279050632911392e-05, |
|
"loss": 0.0001, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"eval_loss": 0.4813273549079895, |
|
"eval_runtime": 80.403, |
|
"eval_samples_per_second": 158.477, |
|
"eval_steps_per_second": 1.244, |
|
"eval_wer": 19.21994685008598, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 160.25, |
|
"learning_rate": 2.2766772151898734e-05, |
|
"loss": 0.0001, |
|
"step": 16025 |
|
}, |
|
{ |
|
"epoch": 160.5, |
|
"learning_rate": 2.2743037974683543e-05, |
|
"loss": 0.0001, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 160.75, |
|
"learning_rate": 2.2719303797468353e-05, |
|
"loss": 0.0001, |
|
"step": 16075 |
|
}, |
|
{ |
|
"epoch": 161.0, |
|
"learning_rate": 2.2695569620253162e-05, |
|
"loss": 0.0001, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 161.25, |
|
"learning_rate": 2.2671835443037972e-05, |
|
"loss": 0.0001, |
|
"step": 16125 |
|
}, |
|
{ |
|
"epoch": 161.5, |
|
"learning_rate": 2.264810126582278e-05, |
|
"loss": 0.0001, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 161.75, |
|
"learning_rate": 2.262436708860759e-05, |
|
"loss": 0.0001, |
|
"step": 16175 |
|
}, |
|
{ |
|
"epoch": 162.0, |
|
"learning_rate": 2.2600632911392404e-05, |
|
"loss": 0.0001, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 162.25, |
|
"learning_rate": 2.2576898734177213e-05, |
|
"loss": 0.0001, |
|
"step": 16225 |
|
}, |
|
{ |
|
"epoch": 162.5, |
|
"learning_rate": 2.2553164556962023e-05, |
|
"loss": 0.0001, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 162.75, |
|
"learning_rate": 2.2529430379746832e-05, |
|
"loss": 0.0001, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 163.0, |
|
"learning_rate": 2.2505696202531642e-05, |
|
"loss": 0.0001, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 163.25, |
|
"learning_rate": 2.248196202531645e-05, |
|
"loss": 0.0001, |
|
"step": 16325 |
|
}, |
|
{ |
|
"epoch": 163.5, |
|
"learning_rate": 2.245822784810126e-05, |
|
"loss": 0.0001, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 163.75, |
|
"learning_rate": 2.2434493670886074e-05, |
|
"loss": 0.0001, |
|
"step": 16375 |
|
}, |
|
{ |
|
"epoch": 164.0, |
|
"learning_rate": 2.2410759493670883e-05, |
|
"loss": 0.0001, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 164.25, |
|
"learning_rate": 2.2387025316455693e-05, |
|
"loss": 0.0001, |
|
"step": 16425 |
|
}, |
|
{ |
|
"epoch": 164.5, |
|
"learning_rate": 2.2363291139240502e-05, |
|
"loss": 0.0001, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 164.75, |
|
"learning_rate": 2.2339556962025315e-05, |
|
"loss": 0.0001, |
|
"step": 16475 |
|
}, |
|
{ |
|
"epoch": 165.0, |
|
"learning_rate": 2.2315822784810128e-05, |
|
"loss": 0.0001, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 165.25, |
|
"learning_rate": 2.2292088607594938e-05, |
|
"loss": 0.0001, |
|
"step": 16525 |
|
}, |
|
{ |
|
"epoch": 165.5, |
|
"learning_rate": 2.2268354430379747e-05, |
|
"loss": 0.0001, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 165.75, |
|
"learning_rate": 2.2244620253164557e-05, |
|
"loss": 0.0001, |
|
"step": 16575 |
|
}, |
|
{ |
|
"epoch": 166.0, |
|
"learning_rate": 2.2220886075949366e-05, |
|
"loss": 0.0001, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 166.25, |
|
"learning_rate": 2.2197151898734176e-05, |
|
"loss": 0.0001, |
|
"step": 16625 |
|
}, |
|
{ |
|
"epoch": 166.5, |
|
"learning_rate": 2.2173417721518985e-05, |
|
"loss": 0.0001, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 166.75, |
|
"learning_rate": 2.2149683544303798e-05, |
|
"loss": 0.0001, |
|
"step": 16675 |
|
}, |
|
{ |
|
"epoch": 167.0, |
|
"learning_rate": 2.2125949367088608e-05, |
|
"loss": 0.0001, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 167.25, |
|
"learning_rate": 2.2102215189873417e-05, |
|
"loss": 0.0001, |
|
"step": 16725 |
|
}, |
|
{ |
|
"epoch": 167.5, |
|
"learning_rate": 2.2078481012658227e-05, |
|
"loss": 0.0001, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 167.75, |
|
"learning_rate": 2.2054746835443036e-05, |
|
"loss": 0.0001, |
|
"step": 16775 |
|
}, |
|
{ |
|
"epoch": 168.0, |
|
"learning_rate": 2.2031012658227846e-05, |
|
"loss": 0.0001, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 168.25, |
|
"learning_rate": 2.2007278481012655e-05, |
|
"loss": 0.0, |
|
"step": 16825 |
|
}, |
|
{ |
|
"epoch": 168.5, |
|
"learning_rate": 2.1983544303797468e-05, |
|
"loss": 0.0, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 168.75, |
|
"learning_rate": 2.1959810126582277e-05, |
|
"loss": 0.0, |
|
"step": 16875 |
|
}, |
|
{ |
|
"epoch": 169.0, |
|
"learning_rate": 2.1936075949367087e-05, |
|
"loss": 0.0001, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 169.25, |
|
"learning_rate": 2.1912341772151896e-05, |
|
"loss": 0.0, |
|
"step": 16925 |
|
}, |
|
{ |
|
"epoch": 169.5, |
|
"learning_rate": 2.1888607594936706e-05, |
|
"loss": 0.0001, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 169.75, |
|
"learning_rate": 2.1864873417721515e-05, |
|
"loss": 0.0, |
|
"step": 16975 |
|
}, |
|
{ |
|
"epoch": 170.0, |
|
"learning_rate": 2.1841139240506325e-05, |
|
"loss": 0.0, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 170.0, |
|
"eval_loss": 0.49322769045829773, |
|
"eval_runtime": 80.3155, |
|
"eval_samples_per_second": 158.649, |
|
"eval_steps_per_second": 1.245, |
|
"eval_wer": 19.169141785211817, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 170.25, |
|
"learning_rate": 2.1817405063291138e-05, |
|
"loss": 0.0, |
|
"step": 17025 |
|
}, |
|
{ |
|
"epoch": 170.5, |
|
"learning_rate": 2.1793670886075947e-05, |
|
"loss": 0.0, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 170.75, |
|
"learning_rate": 2.1769936708860757e-05, |
|
"loss": 0.0, |
|
"step": 17075 |
|
}, |
|
{ |
|
"epoch": 171.0, |
|
"learning_rate": 2.1746202531645566e-05, |
|
"loss": 0.0001, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 171.25, |
|
"learning_rate": 2.1722468354430376e-05, |
|
"loss": 0.0, |
|
"step": 17125 |
|
}, |
|
{ |
|
"epoch": 171.5, |
|
"learning_rate": 2.1698734177215185e-05, |
|
"loss": 0.0001, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 171.75, |
|
"learning_rate": 2.1674999999999995e-05, |
|
"loss": 0.0, |
|
"step": 17175 |
|
}, |
|
{ |
|
"epoch": 172.0, |
|
"learning_rate": 2.1651265822784808e-05, |
|
"loss": 0.0, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 172.25, |
|
"learning_rate": 2.1627531645569617e-05, |
|
"loss": 0.0, |
|
"step": 17225 |
|
}, |
|
{ |
|
"epoch": 172.5, |
|
"learning_rate": 2.1603797468354427e-05, |
|
"loss": 0.0, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 172.75, |
|
"learning_rate": 2.1580063291139236e-05, |
|
"loss": 0.0, |
|
"step": 17275 |
|
}, |
|
{ |
|
"epoch": 173.0, |
|
"learning_rate": 2.155632911392405e-05, |
|
"loss": 0.0, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 173.25, |
|
"learning_rate": 2.1532594936708862e-05, |
|
"loss": 0.0, |
|
"step": 17325 |
|
}, |
|
{ |
|
"epoch": 173.5, |
|
"learning_rate": 2.150886075949367e-05, |
|
"loss": 0.0, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 173.75, |
|
"learning_rate": 2.148512658227848e-05, |
|
"loss": 0.0, |
|
"step": 17375 |
|
}, |
|
{ |
|
"epoch": 174.0, |
|
"learning_rate": 2.146139240506329e-05, |
|
"loss": 0.0, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 174.25, |
|
"learning_rate": 2.14376582278481e-05, |
|
"loss": 0.0, |
|
"step": 17425 |
|
}, |
|
{ |
|
"epoch": 174.5, |
|
"learning_rate": 2.141392405063291e-05, |
|
"loss": 0.0, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 174.75, |
|
"learning_rate": 2.139018987341772e-05, |
|
"loss": 0.0, |
|
"step": 17475 |
|
}, |
|
{ |
|
"epoch": 175.0, |
|
"learning_rate": 2.1366455696202532e-05, |
|
"loss": 0.0, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 175.25, |
|
"learning_rate": 2.134272151898734e-05, |
|
"loss": 0.0, |
|
"step": 17525 |
|
}, |
|
{ |
|
"epoch": 175.5, |
|
"learning_rate": 2.131898734177215e-05, |
|
"loss": 0.0, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 175.75, |
|
"learning_rate": 2.129525316455696e-05, |
|
"loss": 0.0, |
|
"step": 17575 |
|
}, |
|
{ |
|
"epoch": 176.0, |
|
"learning_rate": 2.127151898734177e-05, |
|
"loss": 0.0, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 176.25, |
|
"learning_rate": 2.124778481012658e-05, |
|
"loss": 0.0, |
|
"step": 17625 |
|
}, |
|
{ |
|
"epoch": 176.5, |
|
"learning_rate": 2.122405063291139e-05, |
|
"loss": 0.0, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 176.75, |
|
"learning_rate": 2.1200316455696202e-05, |
|
"loss": 0.0, |
|
"step": 17675 |
|
}, |
|
{ |
|
"epoch": 177.0, |
|
"learning_rate": 2.117658227848101e-05, |
|
"loss": 0.0, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 177.25, |
|
"learning_rate": 2.115284810126582e-05, |
|
"loss": 0.0, |
|
"step": 17725 |
|
}, |
|
{ |
|
"epoch": 177.5, |
|
"learning_rate": 2.112911392405063e-05, |
|
"loss": 0.0, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 177.75, |
|
"learning_rate": 2.110537974683544e-05, |
|
"loss": 0.0, |
|
"step": 17775 |
|
}, |
|
{ |
|
"epoch": 178.0, |
|
"learning_rate": 2.108164556962025e-05, |
|
"loss": 0.0, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 178.25, |
|
"learning_rate": 2.1057911392405062e-05, |
|
"loss": 0.0, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 178.5, |
|
"learning_rate": 2.1034177215189872e-05, |
|
"loss": 0.0, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 178.75, |
|
"learning_rate": 2.101044303797468e-05, |
|
"loss": 0.0, |
|
"step": 17875 |
|
}, |
|
{ |
|
"epoch": 179.0, |
|
"learning_rate": 2.098670886075949e-05, |
|
"loss": 0.0, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 179.25, |
|
"learning_rate": 2.09629746835443e-05, |
|
"loss": 0.0, |
|
"step": 17925 |
|
}, |
|
{ |
|
"epoch": 179.5, |
|
"learning_rate": 2.093924050632911e-05, |
|
"loss": 0.0, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 179.75, |
|
"learning_rate": 2.091550632911392e-05, |
|
"loss": 0.0, |
|
"step": 17975 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"learning_rate": 2.0891772151898732e-05, |
|
"loss": 0.0, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"eval_loss": 0.5040577054023743, |
|
"eval_runtime": 80.2662, |
|
"eval_samples_per_second": 158.747, |
|
"eval_steps_per_second": 1.246, |
|
"eval_wer": 19.1290839455995, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 180.25, |
|
"learning_rate": 2.0868037974683542e-05, |
|
"loss": 0.0, |
|
"step": 18025 |
|
}, |
|
{ |
|
"epoch": 180.5, |
|
"learning_rate": 2.084430379746835e-05, |
|
"loss": 0.0, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 180.75, |
|
"learning_rate": 2.082056962025316e-05, |
|
"loss": 0.0, |
|
"step": 18075 |
|
}, |
|
{ |
|
"epoch": 181.0, |
|
"learning_rate": 2.0796835443037974e-05, |
|
"loss": 0.0, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 181.25, |
|
"learning_rate": 2.0773101265822783e-05, |
|
"loss": 0.0, |
|
"step": 18125 |
|
}, |
|
{ |
|
"epoch": 181.5, |
|
"learning_rate": 2.0749367088607596e-05, |
|
"loss": 0.0, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 181.75, |
|
"learning_rate": 2.0725632911392406e-05, |
|
"loss": 0.0, |
|
"step": 18175 |
|
}, |
|
{ |
|
"epoch": 182.0, |
|
"learning_rate": 2.0701898734177215e-05, |
|
"loss": 0.0, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 182.25, |
|
"learning_rate": 2.0678164556962025e-05, |
|
"loss": 0.0, |
|
"step": 18225 |
|
}, |
|
{ |
|
"epoch": 182.5, |
|
"learning_rate": 2.0654430379746834e-05, |
|
"loss": 0.0, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 182.75, |
|
"learning_rate": 2.0630696202531644e-05, |
|
"loss": 0.0, |
|
"step": 18275 |
|
}, |
|
{ |
|
"epoch": 183.0, |
|
"learning_rate": 2.0606962025316457e-05, |
|
"loss": 0.0, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 183.25, |
|
"learning_rate": 2.0583227848101266e-05, |
|
"loss": 0.0, |
|
"step": 18325 |
|
}, |
|
{ |
|
"epoch": 183.5, |
|
"learning_rate": 2.0559493670886076e-05, |
|
"loss": 0.0, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 183.75, |
|
"learning_rate": 2.0535759493670885e-05, |
|
"loss": 0.0, |
|
"step": 18375 |
|
}, |
|
{ |
|
"epoch": 184.0, |
|
"learning_rate": 2.0512025316455695e-05, |
|
"loss": 0.0, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 184.25, |
|
"learning_rate": 2.0488291139240504e-05, |
|
"loss": 0.0, |
|
"step": 18425 |
|
}, |
|
{ |
|
"epoch": 184.5, |
|
"learning_rate": 2.0464556962025314e-05, |
|
"loss": 0.0, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 184.75, |
|
"learning_rate": 2.0440822784810127e-05, |
|
"loss": 0.0, |
|
"step": 18475 |
|
}, |
|
{ |
|
"epoch": 185.0, |
|
"learning_rate": 2.0417088607594936e-05, |
|
"loss": 0.0, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 185.25, |
|
"learning_rate": 2.0393354430379746e-05, |
|
"loss": 0.0, |
|
"step": 18525 |
|
}, |
|
{ |
|
"epoch": 185.5, |
|
"learning_rate": 2.0369620253164555e-05, |
|
"loss": 0.0, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 185.75, |
|
"learning_rate": 2.0345886075949365e-05, |
|
"loss": 0.0, |
|
"step": 18575 |
|
}, |
|
{ |
|
"epoch": 186.0, |
|
"learning_rate": 2.0322151898734174e-05, |
|
"loss": 0.0, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 186.25, |
|
"learning_rate": 2.0298417721518984e-05, |
|
"loss": 0.0, |
|
"step": 18625 |
|
}, |
|
{ |
|
"epoch": 186.5, |
|
"learning_rate": 2.0274683544303796e-05, |
|
"loss": 0.0, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 186.75, |
|
"learning_rate": 2.0250949367088606e-05, |
|
"loss": 0.0, |
|
"step": 18675 |
|
}, |
|
{ |
|
"epoch": 187.0, |
|
"learning_rate": 2.0227215189873415e-05, |
|
"loss": 0.0, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 187.25, |
|
"learning_rate": 2.0203481012658225e-05, |
|
"loss": 0.0, |
|
"step": 18725 |
|
}, |
|
{ |
|
"epoch": 187.5, |
|
"learning_rate": 2.0179746835443035e-05, |
|
"loss": 0.0, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 187.75, |
|
"learning_rate": 2.0156012658227844e-05, |
|
"loss": 0.0, |
|
"step": 18775 |
|
}, |
|
{ |
|
"epoch": 188.0, |
|
"learning_rate": 2.0132278481012654e-05, |
|
"loss": 0.0, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 188.25, |
|
"learning_rate": 2.0108544303797466e-05, |
|
"loss": 0.0, |
|
"step": 18825 |
|
}, |
|
{ |
|
"epoch": 188.5, |
|
"learning_rate": 2.0084810126582276e-05, |
|
"loss": 0.0, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 188.75, |
|
"learning_rate": 2.0061075949367085e-05, |
|
"loss": 0.0, |
|
"step": 18875 |
|
}, |
|
{ |
|
"epoch": 189.0, |
|
"learning_rate": 2.0037341772151895e-05, |
|
"loss": 0.0, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 189.25, |
|
"learning_rate": 2.0013607594936708e-05, |
|
"loss": 0.0, |
|
"step": 18925 |
|
}, |
|
{ |
|
"epoch": 189.5, |
|
"learning_rate": 1.998987341772152e-05, |
|
"loss": 0.0, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 189.75, |
|
"learning_rate": 1.996613924050633e-05, |
|
"loss": 0.0, |
|
"step": 18975 |
|
}, |
|
{ |
|
"epoch": 190.0, |
|
"learning_rate": 1.994240506329114e-05, |
|
"loss": 0.0, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 190.0, |
|
"eval_loss": 0.5146323442459106, |
|
"eval_runtime": 80.5291, |
|
"eval_samples_per_second": 158.229, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.094888228857275, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 190.25, |
|
"learning_rate": 1.991867088607595e-05, |
|
"loss": 0.0, |
|
"step": 19025 |
|
}, |
|
{ |
|
"epoch": 190.5, |
|
"learning_rate": 1.989493670886076e-05, |
|
"loss": 0.0, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 190.75, |
|
"learning_rate": 1.9871202531645568e-05, |
|
"loss": 0.0, |
|
"step": 19075 |
|
}, |
|
{ |
|
"epoch": 191.0, |
|
"learning_rate": 1.9847468354430378e-05, |
|
"loss": 0.0, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 191.25, |
|
"learning_rate": 1.982373417721519e-05, |
|
"loss": 0.0, |
|
"step": 19125 |
|
}, |
|
{ |
|
"epoch": 191.5, |
|
"learning_rate": 1.98e-05, |
|
"loss": 0.0, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 191.75, |
|
"learning_rate": 1.977626582278481e-05, |
|
"loss": 0.0, |
|
"step": 19175 |
|
}, |
|
{ |
|
"epoch": 192.0, |
|
"learning_rate": 1.975253164556962e-05, |
|
"loss": 0.0, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 192.25, |
|
"learning_rate": 1.972879746835443e-05, |
|
"loss": 0.0, |
|
"step": 19225 |
|
}, |
|
{ |
|
"epoch": 192.5, |
|
"learning_rate": 1.9705063291139238e-05, |
|
"loss": 0.0, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 192.75, |
|
"learning_rate": 1.9681329113924048e-05, |
|
"loss": 0.0, |
|
"step": 19275 |
|
}, |
|
{ |
|
"epoch": 193.0, |
|
"learning_rate": 1.965759493670886e-05, |
|
"loss": 0.0, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 193.25, |
|
"learning_rate": 1.963386075949367e-05, |
|
"loss": 0.0, |
|
"step": 19325 |
|
}, |
|
{ |
|
"epoch": 193.5, |
|
"learning_rate": 1.961012658227848e-05, |
|
"loss": 0.0, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 193.75, |
|
"learning_rate": 1.958639240506329e-05, |
|
"loss": 0.0, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 194.0, |
|
"learning_rate": 1.95626582278481e-05, |
|
"loss": 0.0, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 194.25, |
|
"learning_rate": 1.9538924050632908e-05, |
|
"loss": 0.0, |
|
"step": 19425 |
|
}, |
|
{ |
|
"epoch": 194.5, |
|
"learning_rate": 1.9515189873417718e-05, |
|
"loss": 0.0, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 194.75, |
|
"learning_rate": 1.949145569620253e-05, |
|
"loss": 0.0, |
|
"step": 19475 |
|
}, |
|
{ |
|
"epoch": 195.0, |
|
"learning_rate": 1.946772151898734e-05, |
|
"loss": 0.0, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 195.25, |
|
"learning_rate": 1.944398734177215e-05, |
|
"loss": 0.0, |
|
"step": 19525 |
|
}, |
|
{ |
|
"epoch": 195.5, |
|
"learning_rate": 1.942025316455696e-05, |
|
"loss": 0.0, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 195.75, |
|
"learning_rate": 1.939651898734177e-05, |
|
"loss": 0.0, |
|
"step": 19575 |
|
}, |
|
{ |
|
"epoch": 196.0, |
|
"learning_rate": 1.9372784810126578e-05, |
|
"loss": 0.0, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 196.25, |
|
"learning_rate": 1.9349050632911388e-05, |
|
"loss": 0.0, |
|
"step": 19625 |
|
}, |
|
{ |
|
"epoch": 196.5, |
|
"learning_rate": 1.93253164556962e-05, |
|
"loss": 0.0, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 196.75, |
|
"learning_rate": 1.930158227848101e-05, |
|
"loss": 0.0, |
|
"step": 19675 |
|
}, |
|
{ |
|
"epoch": 197.0, |
|
"learning_rate": 1.927784810126582e-05, |
|
"loss": 0.0, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 197.25, |
|
"learning_rate": 1.925411392405063e-05, |
|
"loss": 0.0, |
|
"step": 19725 |
|
}, |
|
{ |
|
"epoch": 197.5, |
|
"learning_rate": 1.9230379746835442e-05, |
|
"loss": 0.0, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 197.75, |
|
"learning_rate": 1.9206645569620255e-05, |
|
"loss": 0.0, |
|
"step": 19775 |
|
}, |
|
{ |
|
"epoch": 198.0, |
|
"learning_rate": 1.9182911392405064e-05, |
|
"loss": 0.0, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 198.25, |
|
"learning_rate": 1.9159177215189874e-05, |
|
"loss": 0.0, |
|
"step": 19825 |
|
}, |
|
{ |
|
"epoch": 198.5, |
|
"learning_rate": 1.9135443037974683e-05, |
|
"loss": 0.0, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 198.75, |
|
"learning_rate": 1.9111708860759493e-05, |
|
"loss": 0.0, |
|
"step": 19875 |
|
}, |
|
{ |
|
"epoch": 199.0, |
|
"learning_rate": 1.9087974683544302e-05, |
|
"loss": 0.0, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 199.25, |
|
"learning_rate": 1.9064240506329112e-05, |
|
"loss": 0.0, |
|
"step": 19925 |
|
}, |
|
{ |
|
"epoch": 199.5, |
|
"learning_rate": 1.9040506329113925e-05, |
|
"loss": 0.0, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 199.75, |
|
"learning_rate": 1.9016772151898734e-05, |
|
"loss": 0.0, |
|
"step": 19975 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"learning_rate": 1.8993037974683544e-05, |
|
"loss": 0.0, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"eval_loss": 0.5253657698631287, |
|
"eval_runtime": 80.492, |
|
"eval_samples_per_second": 158.301, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.123221822729406, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 200.25, |
|
"learning_rate": 1.8969303797468353e-05, |
|
"loss": 0.0, |
|
"step": 20025 |
|
}, |
|
{ |
|
"epoch": 200.5, |
|
"learning_rate": 1.8945569620253163e-05, |
|
"loss": 0.0, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 200.75, |
|
"learning_rate": 1.8921835443037972e-05, |
|
"loss": 0.0, |
|
"step": 20075 |
|
}, |
|
{ |
|
"epoch": 201.0, |
|
"learning_rate": 1.8898101265822785e-05, |
|
"loss": 0.0, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 201.25, |
|
"learning_rate": 1.8874367088607595e-05, |
|
"loss": 0.0, |
|
"step": 20125 |
|
}, |
|
{ |
|
"epoch": 201.5, |
|
"learning_rate": 1.8850632911392404e-05, |
|
"loss": 0.0, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 201.75, |
|
"learning_rate": 1.8826898734177214e-05, |
|
"loss": 0.0, |
|
"step": 20175 |
|
}, |
|
{ |
|
"epoch": 202.0, |
|
"learning_rate": 1.8803164556962023e-05, |
|
"loss": 0.0, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 202.25, |
|
"learning_rate": 1.8779430379746833e-05, |
|
"loss": 0.0, |
|
"step": 20225 |
|
}, |
|
{ |
|
"epoch": 202.5, |
|
"learning_rate": 1.8755696202531642e-05, |
|
"loss": 0.0, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 202.75, |
|
"learning_rate": 1.8731962025316455e-05, |
|
"loss": 0.0, |
|
"step": 20275 |
|
}, |
|
{ |
|
"epoch": 203.0, |
|
"learning_rate": 1.8708227848101265e-05, |
|
"loss": 0.0, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 203.25, |
|
"learning_rate": 1.8684493670886074e-05, |
|
"loss": 0.0, |
|
"step": 20325 |
|
}, |
|
{ |
|
"epoch": 203.5, |
|
"learning_rate": 1.8660759493670884e-05, |
|
"loss": 0.0, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 203.75, |
|
"learning_rate": 1.8637025316455696e-05, |
|
"loss": 0.0, |
|
"step": 20375 |
|
}, |
|
{ |
|
"epoch": 204.0, |
|
"learning_rate": 1.8613291139240506e-05, |
|
"loss": 0.0, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 204.25, |
|
"learning_rate": 1.8589556962025316e-05, |
|
"loss": 0.0, |
|
"step": 20425 |
|
}, |
|
{ |
|
"epoch": 204.5, |
|
"learning_rate": 1.8565822784810125e-05, |
|
"loss": 0.0, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 204.75, |
|
"learning_rate": 1.8542088607594935e-05, |
|
"loss": 0.0, |
|
"step": 20475 |
|
}, |
|
{ |
|
"epoch": 205.0, |
|
"learning_rate": 1.8518354430379744e-05, |
|
"loss": 0.0, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 205.25, |
|
"learning_rate": 1.8494620253164557e-05, |
|
"loss": 0.0, |
|
"step": 20525 |
|
}, |
|
{ |
|
"epoch": 205.5, |
|
"learning_rate": 1.8470886075949366e-05, |
|
"loss": 0.0, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 205.75, |
|
"learning_rate": 1.8447151898734176e-05, |
|
"loss": 0.0, |
|
"step": 20575 |
|
}, |
|
{ |
|
"epoch": 206.0, |
|
"learning_rate": 1.8423417721518985e-05, |
|
"loss": 0.0, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 206.25, |
|
"learning_rate": 1.8399683544303795e-05, |
|
"loss": 0.0, |
|
"step": 20625 |
|
}, |
|
{ |
|
"epoch": 206.5, |
|
"learning_rate": 1.8375949367088604e-05, |
|
"loss": 0.0, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 206.75, |
|
"learning_rate": 1.8352215189873414e-05, |
|
"loss": 0.0, |
|
"step": 20675 |
|
}, |
|
{ |
|
"epoch": 207.0, |
|
"learning_rate": 1.8328481012658227e-05, |
|
"loss": 0.0, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 207.25, |
|
"learning_rate": 1.8304746835443036e-05, |
|
"loss": 0.0, |
|
"step": 20725 |
|
}, |
|
{ |
|
"epoch": 207.5, |
|
"learning_rate": 1.8281012658227846e-05, |
|
"loss": 0.0, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 207.75, |
|
"learning_rate": 1.825727848101266e-05, |
|
"loss": 0.0, |
|
"step": 20775 |
|
}, |
|
{ |
|
"epoch": 208.0, |
|
"learning_rate": 1.8233544303797468e-05, |
|
"loss": 0.0, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 208.25, |
|
"learning_rate": 1.8209810126582278e-05, |
|
"loss": 0.0, |
|
"step": 20825 |
|
}, |
|
{ |
|
"epoch": 208.5, |
|
"learning_rate": 1.8186075949367087e-05, |
|
"loss": 0.0, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 208.75, |
|
"learning_rate": 1.8162341772151897e-05, |
|
"loss": 0.0, |
|
"step": 20875 |
|
}, |
|
{ |
|
"epoch": 209.0, |
|
"learning_rate": 1.8138607594936706e-05, |
|
"loss": 0.0, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 209.25, |
|
"learning_rate": 1.811487341772152e-05, |
|
"loss": 0.0, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 209.5, |
|
"learning_rate": 1.809113924050633e-05, |
|
"loss": 0.0, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 209.75, |
|
"learning_rate": 1.8067405063291138e-05, |
|
"loss": 0.0, |
|
"step": 20975 |
|
}, |
|
{ |
|
"epoch": 210.0, |
|
"learning_rate": 1.8043670886075948e-05, |
|
"loss": 0.0, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 210.0, |
|
"eval_loss": 0.5368518233299255, |
|
"eval_runtime": 80.6917, |
|
"eval_samples_per_second": 157.91, |
|
"eval_steps_per_second": 1.239, |
|
"eval_wer": 19.136900109426293, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 210.25, |
|
"learning_rate": 1.8019936708860757e-05, |
|
"loss": 0.0, |
|
"step": 21025 |
|
}, |
|
{ |
|
"epoch": 210.5, |
|
"learning_rate": 1.7996202531645567e-05, |
|
"loss": 0.0, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 210.75, |
|
"learning_rate": 1.7972468354430376e-05, |
|
"loss": 0.0, |
|
"step": 21075 |
|
}, |
|
{ |
|
"epoch": 211.0, |
|
"learning_rate": 1.794873417721519e-05, |
|
"loss": 0.0, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 211.25, |
|
"learning_rate": 1.7925e-05, |
|
"loss": 0.0, |
|
"step": 21125 |
|
}, |
|
{ |
|
"epoch": 211.5, |
|
"learning_rate": 1.7901265822784808e-05, |
|
"loss": 0.0, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 211.75, |
|
"learning_rate": 1.7877531645569618e-05, |
|
"loss": 0.0, |
|
"step": 21175 |
|
}, |
|
{ |
|
"epoch": 212.0, |
|
"learning_rate": 1.785379746835443e-05, |
|
"loss": 0.0, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 212.25, |
|
"learning_rate": 1.783006329113924e-05, |
|
"loss": 0.0, |
|
"step": 21225 |
|
}, |
|
{ |
|
"epoch": 212.5, |
|
"learning_rate": 1.780632911392405e-05, |
|
"loss": 0.0, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 212.75, |
|
"learning_rate": 1.778259493670886e-05, |
|
"loss": 0.0, |
|
"step": 21275 |
|
}, |
|
{ |
|
"epoch": 213.0, |
|
"learning_rate": 1.775886075949367e-05, |
|
"loss": 0.0, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 213.25, |
|
"learning_rate": 1.773512658227848e-05, |
|
"loss": 0.0, |
|
"step": 21325 |
|
}, |
|
{ |
|
"epoch": 213.5, |
|
"learning_rate": 1.771139240506329e-05, |
|
"loss": 0.0, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 213.75, |
|
"learning_rate": 1.76876582278481e-05, |
|
"loss": 0.0, |
|
"step": 21375 |
|
}, |
|
{ |
|
"epoch": 214.0, |
|
"learning_rate": 1.766392405063291e-05, |
|
"loss": 0.0, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 214.25, |
|
"learning_rate": 1.764018987341772e-05, |
|
"loss": 0.0, |
|
"step": 21425 |
|
}, |
|
{ |
|
"epoch": 214.5, |
|
"learning_rate": 1.761645569620253e-05, |
|
"loss": 0.0, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 214.75, |
|
"learning_rate": 1.759272151898734e-05, |
|
"loss": 0.0, |
|
"step": 21475 |
|
}, |
|
{ |
|
"epoch": 215.0, |
|
"learning_rate": 1.756898734177215e-05, |
|
"loss": 0.0, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 215.25, |
|
"learning_rate": 1.754525316455696e-05, |
|
"loss": 0.0, |
|
"step": 21525 |
|
}, |
|
{ |
|
"epoch": 215.5, |
|
"learning_rate": 1.752151898734177e-05, |
|
"loss": 0.0, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 215.75, |
|
"learning_rate": 1.749778481012658e-05, |
|
"loss": 0.0, |
|
"step": 21575 |
|
}, |
|
{ |
|
"epoch": 216.0, |
|
"learning_rate": 1.7474050632911393e-05, |
|
"loss": 0.0, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 216.25, |
|
"learning_rate": 1.7450316455696202e-05, |
|
"loss": 0.0, |
|
"step": 21625 |
|
}, |
|
{ |
|
"epoch": 216.5, |
|
"learning_rate": 1.7426582278481012e-05, |
|
"loss": 0.0, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 216.75, |
|
"learning_rate": 1.740284810126582e-05, |
|
"loss": 0.0, |
|
"step": 21675 |
|
}, |
|
{ |
|
"epoch": 217.0, |
|
"learning_rate": 1.737911392405063e-05, |
|
"loss": 0.0, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 217.25, |
|
"learning_rate": 1.735537974683544e-05, |
|
"loss": 0.0, |
|
"step": 21725 |
|
}, |
|
{ |
|
"epoch": 217.5, |
|
"learning_rate": 1.7331645569620253e-05, |
|
"loss": 0.0, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 217.75, |
|
"learning_rate": 1.7307911392405063e-05, |
|
"loss": 0.0, |
|
"step": 21775 |
|
}, |
|
{ |
|
"epoch": 218.0, |
|
"learning_rate": 1.7284177215189872e-05, |
|
"loss": 0.0, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 218.25, |
|
"learning_rate": 1.7260443037974682e-05, |
|
"loss": 0.0, |
|
"step": 21825 |
|
}, |
|
{ |
|
"epoch": 218.5, |
|
"learning_rate": 1.723670886075949e-05, |
|
"loss": 0.0, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 218.75, |
|
"learning_rate": 1.72129746835443e-05, |
|
"loss": 0.0, |
|
"step": 21875 |
|
}, |
|
{ |
|
"epoch": 219.0, |
|
"learning_rate": 1.718924050632911e-05, |
|
"loss": 0.0, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 219.25, |
|
"learning_rate": 1.7165506329113923e-05, |
|
"loss": 0.0, |
|
"step": 21925 |
|
}, |
|
{ |
|
"epoch": 219.5, |
|
"learning_rate": 1.7141772151898733e-05, |
|
"loss": 0.0, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 219.75, |
|
"learning_rate": 1.7118037974683542e-05, |
|
"loss": 0.0, |
|
"step": 21975 |
|
}, |
|
{ |
|
"epoch": 220.0, |
|
"learning_rate": 1.7094303797468352e-05, |
|
"loss": 0.0, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 220.0, |
|
"eval_loss": 0.5483614802360535, |
|
"eval_runtime": 80.6444, |
|
"eval_samples_per_second": 158.002, |
|
"eval_steps_per_second": 1.24, |
|
"eval_wer": 19.112474597467564, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 220.25, |
|
"learning_rate": 1.7070569620253165e-05, |
|
"loss": 0.0, |
|
"step": 22025 |
|
}, |
|
{ |
|
"epoch": 220.5, |
|
"learning_rate": 1.7046835443037974e-05, |
|
"loss": 0.0, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 220.75, |
|
"learning_rate": 1.7023101265822784e-05, |
|
"loss": 0.0, |
|
"step": 22075 |
|
}, |
|
{ |
|
"epoch": 221.0, |
|
"learning_rate": 1.6999367088607593e-05, |
|
"loss": 0.0, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 221.25, |
|
"learning_rate": 1.6975632911392403e-05, |
|
"loss": 0.0, |
|
"step": 22125 |
|
}, |
|
{ |
|
"epoch": 221.5, |
|
"learning_rate": 1.6951898734177216e-05, |
|
"loss": 0.0, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 221.75, |
|
"learning_rate": 1.6928164556962025e-05, |
|
"loss": 0.0, |
|
"step": 22175 |
|
}, |
|
{ |
|
"epoch": 222.0, |
|
"learning_rate": 1.6904430379746835e-05, |
|
"loss": 0.0, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 222.25, |
|
"learning_rate": 1.6880696202531644e-05, |
|
"loss": 0.0, |
|
"step": 22225 |
|
}, |
|
{ |
|
"epoch": 222.5, |
|
"learning_rate": 1.6856962025316454e-05, |
|
"loss": 0.0, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 222.75, |
|
"learning_rate": 1.6833227848101263e-05, |
|
"loss": 0.0, |
|
"step": 22275 |
|
}, |
|
{ |
|
"epoch": 223.0, |
|
"learning_rate": 1.6809493670886073e-05, |
|
"loss": 0.0, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 223.25, |
|
"learning_rate": 1.6785759493670885e-05, |
|
"loss": 0.0, |
|
"step": 22325 |
|
}, |
|
{ |
|
"epoch": 223.5, |
|
"learning_rate": 1.6762025316455695e-05, |
|
"loss": 0.0, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 223.75, |
|
"learning_rate": 1.6738291139240504e-05, |
|
"loss": 0.0, |
|
"step": 22375 |
|
}, |
|
{ |
|
"epoch": 224.0, |
|
"learning_rate": 1.6714556962025314e-05, |
|
"loss": 0.0, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 224.25, |
|
"learning_rate": 1.6690822784810127e-05, |
|
"loss": 0.0, |
|
"step": 22425 |
|
}, |
|
{ |
|
"epoch": 224.5, |
|
"learning_rate": 1.6667088607594936e-05, |
|
"loss": 0.0, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 224.75, |
|
"learning_rate": 1.6643354430379746e-05, |
|
"loss": 0.0, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 225.0, |
|
"learning_rate": 1.6619620253164555e-05, |
|
"loss": 0.0, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 225.25, |
|
"learning_rate": 1.6595886075949365e-05, |
|
"loss": 0.0, |
|
"step": 22525 |
|
}, |
|
{ |
|
"epoch": 225.5, |
|
"learning_rate": 1.6572151898734178e-05, |
|
"loss": 0.0, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 225.75, |
|
"learning_rate": 1.6548417721518987e-05, |
|
"loss": 0.0, |
|
"step": 22575 |
|
}, |
|
{ |
|
"epoch": 226.0, |
|
"learning_rate": 1.6524683544303797e-05, |
|
"loss": 0.0, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 226.25, |
|
"learning_rate": 1.6500949367088606e-05, |
|
"loss": 0.0, |
|
"step": 22625 |
|
}, |
|
{ |
|
"epoch": 226.5, |
|
"learning_rate": 1.6477215189873416e-05, |
|
"loss": 0.0, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 226.75, |
|
"learning_rate": 1.6453481012658225e-05, |
|
"loss": 0.0, |
|
"step": 22675 |
|
}, |
|
{ |
|
"epoch": 227.0, |
|
"learning_rate": 1.6429746835443035e-05, |
|
"loss": 0.0, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 227.25, |
|
"learning_rate": 1.6406012658227848e-05, |
|
"loss": 0.0, |
|
"step": 22725 |
|
}, |
|
{ |
|
"epoch": 227.5, |
|
"learning_rate": 1.6382278481012657e-05, |
|
"loss": 0.0, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 227.75, |
|
"learning_rate": 1.6358544303797467e-05, |
|
"loss": 0.0, |
|
"step": 22775 |
|
}, |
|
{ |
|
"epoch": 228.0, |
|
"learning_rate": 1.6334810126582276e-05, |
|
"loss": 0.0, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 228.25, |
|
"learning_rate": 1.6311075949367086e-05, |
|
"loss": 0.0, |
|
"step": 22825 |
|
}, |
|
{ |
|
"epoch": 228.5, |
|
"learning_rate": 1.62873417721519e-05, |
|
"loss": 0.0, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 228.75, |
|
"learning_rate": 1.6263607594936708e-05, |
|
"loss": 0.0, |
|
"step": 22875 |
|
}, |
|
{ |
|
"epoch": 229.0, |
|
"learning_rate": 1.6239873417721518e-05, |
|
"loss": 0.0, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 229.25, |
|
"learning_rate": 1.6216139240506327e-05, |
|
"loss": 0.0, |
|
"step": 22925 |
|
}, |
|
{ |
|
"epoch": 229.5, |
|
"learning_rate": 1.6192405063291137e-05, |
|
"loss": 0.0, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 229.75, |
|
"learning_rate": 1.616867088607595e-05, |
|
"loss": 0.0, |
|
"step": 22975 |
|
}, |
|
{ |
|
"epoch": 230.0, |
|
"learning_rate": 1.614493670886076e-05, |
|
"loss": 0.0, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 230.0, |
|
"eval_loss": 0.5606202483177185, |
|
"eval_runtime": 80.6021, |
|
"eval_samples_per_second": 158.085, |
|
"eval_steps_per_second": 1.241, |
|
"eval_wer": 19.132992027512895, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 230.25, |
|
"learning_rate": 1.612120253164557e-05, |
|
"loss": 0.0, |
|
"step": 23025 |
|
}, |
|
{ |
|
"epoch": 230.5, |
|
"learning_rate": 1.6097468354430378e-05, |
|
"loss": 0.0, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 230.75, |
|
"learning_rate": 1.6073734177215188e-05, |
|
"loss": 0.0, |
|
"step": 23075 |
|
}, |
|
{ |
|
"epoch": 231.0, |
|
"learning_rate": 1.6049999999999997e-05, |
|
"loss": 0.0, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 231.25, |
|
"learning_rate": 1.6026265822784807e-05, |
|
"loss": 0.0, |
|
"step": 23125 |
|
}, |
|
{ |
|
"epoch": 231.5, |
|
"learning_rate": 1.600253164556962e-05, |
|
"loss": 0.0, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 231.75, |
|
"learning_rate": 1.597879746835443e-05, |
|
"loss": 0.0, |
|
"step": 23175 |
|
}, |
|
{ |
|
"epoch": 232.0, |
|
"learning_rate": 1.595506329113924e-05, |
|
"loss": 0.0, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 232.25, |
|
"learning_rate": 1.5931329113924048e-05, |
|
"loss": 0.0, |
|
"step": 23225 |
|
}, |
|
{ |
|
"epoch": 232.5, |
|
"learning_rate": 1.590759493670886e-05, |
|
"loss": 0.0, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 232.75, |
|
"learning_rate": 1.588386075949367e-05, |
|
"loss": 0.0, |
|
"step": 23275 |
|
}, |
|
{ |
|
"epoch": 233.0, |
|
"learning_rate": 1.586012658227848e-05, |
|
"loss": 0.0, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 233.25, |
|
"learning_rate": 1.583639240506329e-05, |
|
"loss": 0.0, |
|
"step": 23325 |
|
}, |
|
{ |
|
"epoch": 233.5, |
|
"learning_rate": 1.58126582278481e-05, |
|
"loss": 0.0, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 233.75, |
|
"learning_rate": 1.5788924050632912e-05, |
|
"loss": 0.0, |
|
"step": 23375 |
|
}, |
|
{ |
|
"epoch": 234.0, |
|
"learning_rate": 1.576518987341772e-05, |
|
"loss": 0.0, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 234.25, |
|
"learning_rate": 1.574145569620253e-05, |
|
"loss": 0.0, |
|
"step": 23425 |
|
}, |
|
{ |
|
"epoch": 234.5, |
|
"learning_rate": 1.571772151898734e-05, |
|
"loss": 0.0, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 234.75, |
|
"learning_rate": 1.569398734177215e-05, |
|
"loss": 0.0, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 235.0, |
|
"learning_rate": 1.567025316455696e-05, |
|
"loss": 0.0, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 235.25, |
|
"learning_rate": 1.564651898734177e-05, |
|
"loss": 0.0, |
|
"step": 23525 |
|
}, |
|
{ |
|
"epoch": 235.5, |
|
"learning_rate": 1.5622784810126582e-05, |
|
"loss": 0.0, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 235.75, |
|
"learning_rate": 1.559905063291139e-05, |
|
"loss": 0.0, |
|
"step": 23575 |
|
}, |
|
{ |
|
"epoch": 236.0, |
|
"learning_rate": 1.55753164556962e-05, |
|
"loss": 0.0, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 236.25, |
|
"learning_rate": 1.555158227848101e-05, |
|
"loss": 0.0, |
|
"step": 23625 |
|
}, |
|
{ |
|
"epoch": 236.5, |
|
"learning_rate": 1.552784810126582e-05, |
|
"loss": 0.0, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 236.75, |
|
"learning_rate": 1.5504113924050633e-05, |
|
"loss": 0.0, |
|
"step": 23675 |
|
}, |
|
{ |
|
"epoch": 237.0, |
|
"learning_rate": 1.5480379746835442e-05, |
|
"loss": 0.0, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 237.25, |
|
"learning_rate": 1.5456645569620252e-05, |
|
"loss": 0.0, |
|
"step": 23725 |
|
}, |
|
{ |
|
"epoch": 237.5, |
|
"learning_rate": 1.543291139240506e-05, |
|
"loss": 0.0, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 237.75, |
|
"learning_rate": 1.5409177215189874e-05, |
|
"loss": 0.0, |
|
"step": 23775 |
|
}, |
|
{ |
|
"epoch": 238.0, |
|
"learning_rate": 1.5385443037974684e-05, |
|
"loss": 0.0, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 238.25, |
|
"learning_rate": 1.5361708860759493e-05, |
|
"loss": 0.0, |
|
"step": 23825 |
|
}, |
|
{ |
|
"epoch": 238.5, |
|
"learning_rate": 1.5337974683544303e-05, |
|
"loss": 0.0, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 238.75, |
|
"learning_rate": 1.5314240506329112e-05, |
|
"loss": 0.0, |
|
"step": 23875 |
|
}, |
|
{ |
|
"epoch": 239.0, |
|
"learning_rate": 1.529050632911392e-05, |
|
"loss": 0.0, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 239.25, |
|
"learning_rate": 1.526677215189873e-05, |
|
"loss": 0.0, |
|
"step": 23925 |
|
}, |
|
{ |
|
"epoch": 239.5, |
|
"learning_rate": 1.5243037974683542e-05, |
|
"loss": 0.0, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 239.75, |
|
"learning_rate": 1.5219303797468352e-05, |
|
"loss": 0.0, |
|
"step": 23975 |
|
}, |
|
{ |
|
"epoch": 240.0, |
|
"learning_rate": 1.5195569620253163e-05, |
|
"loss": 0.0, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 240.0, |
|
"eval_loss": 0.5731757879257202, |
|
"eval_runtime": 80.5462, |
|
"eval_samples_per_second": 158.195, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.196498358605595, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 240.25, |
|
"learning_rate": 1.5171835443037973e-05, |
|
"loss": 0.0, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 240.5, |
|
"learning_rate": 1.5148101265822782e-05, |
|
"loss": 0.0, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 240.75, |
|
"learning_rate": 1.5124367088607595e-05, |
|
"loss": 0.0, |
|
"step": 24075 |
|
}, |
|
{ |
|
"epoch": 241.0, |
|
"learning_rate": 1.5100632911392404e-05, |
|
"loss": 0.0, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 241.25, |
|
"learning_rate": 1.5076898734177214e-05, |
|
"loss": 0.0, |
|
"step": 24125 |
|
}, |
|
{ |
|
"epoch": 241.5, |
|
"learning_rate": 1.5053164556962025e-05, |
|
"loss": 0.0, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 241.75, |
|
"learning_rate": 1.5029430379746835e-05, |
|
"loss": 0.0, |
|
"step": 24175 |
|
}, |
|
{ |
|
"epoch": 242.0, |
|
"learning_rate": 1.5005696202531644e-05, |
|
"loss": 0.0, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 242.25, |
|
"learning_rate": 1.4981962025316455e-05, |
|
"loss": 0.0, |
|
"step": 24225 |
|
}, |
|
{ |
|
"epoch": 242.5, |
|
"learning_rate": 1.4958227848101265e-05, |
|
"loss": 0.0, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 242.75, |
|
"learning_rate": 1.4934493670886074e-05, |
|
"loss": 0.0, |
|
"step": 24275 |
|
}, |
|
{ |
|
"epoch": 243.0, |
|
"learning_rate": 1.4910759493670884e-05, |
|
"loss": 0.0, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 243.25, |
|
"learning_rate": 1.4887025316455695e-05, |
|
"loss": 0.0, |
|
"step": 24325 |
|
}, |
|
{ |
|
"epoch": 243.5, |
|
"learning_rate": 1.4863291139240505e-05, |
|
"loss": 0.0, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 243.75, |
|
"learning_rate": 1.4839556962025314e-05, |
|
"loss": 0.0, |
|
"step": 24375 |
|
}, |
|
{ |
|
"epoch": 244.0, |
|
"learning_rate": 1.4816772151898732e-05, |
|
"loss": 0.0, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 244.25, |
|
"learning_rate": 1.4793037974683542e-05, |
|
"loss": 0.0, |
|
"step": 24425 |
|
}, |
|
{ |
|
"epoch": 244.5, |
|
"learning_rate": 1.4769303797468353e-05, |
|
"loss": 0.0, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 244.75, |
|
"learning_rate": 1.4745569620253164e-05, |
|
"loss": 0.0, |
|
"step": 24475 |
|
}, |
|
{ |
|
"epoch": 245.0, |
|
"learning_rate": 1.4721835443037973e-05, |
|
"loss": 0.0, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 245.25, |
|
"learning_rate": 1.4698101265822785e-05, |
|
"loss": 0.0, |
|
"step": 24525 |
|
}, |
|
{ |
|
"epoch": 245.5, |
|
"learning_rate": 1.4674367088607594e-05, |
|
"loss": 0.0, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 245.75, |
|
"learning_rate": 1.4650632911392404e-05, |
|
"loss": 0.0, |
|
"step": 24575 |
|
}, |
|
{ |
|
"epoch": 246.0, |
|
"learning_rate": 1.4626898734177215e-05, |
|
"loss": 0.0, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 246.25, |
|
"learning_rate": 1.4603164556962024e-05, |
|
"loss": 0.0, |
|
"step": 24625 |
|
}, |
|
{ |
|
"epoch": 246.5, |
|
"learning_rate": 1.4579430379746834e-05, |
|
"loss": 0.0, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 246.75, |
|
"learning_rate": 1.4555696202531643e-05, |
|
"loss": 0.0, |
|
"step": 24675 |
|
}, |
|
{ |
|
"epoch": 247.0, |
|
"learning_rate": 1.4531962025316455e-05, |
|
"loss": 0.0, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 247.25, |
|
"learning_rate": 1.4508227848101264e-05, |
|
"loss": 0.0, |
|
"step": 24725 |
|
}, |
|
{ |
|
"epoch": 247.5, |
|
"learning_rate": 1.4484493670886074e-05, |
|
"loss": 0.0, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 247.75, |
|
"learning_rate": 1.4460759493670885e-05, |
|
"loss": 0.0, |
|
"step": 24775 |
|
}, |
|
{ |
|
"epoch": 248.0, |
|
"learning_rate": 1.4437025316455694e-05, |
|
"loss": 0.0, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 248.25, |
|
"learning_rate": 1.4413291139240504e-05, |
|
"loss": 0.0, |
|
"step": 24825 |
|
}, |
|
{ |
|
"epoch": 248.5, |
|
"learning_rate": 1.4389556962025315e-05, |
|
"loss": 0.0, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 248.75, |
|
"learning_rate": 1.4365822784810126e-05, |
|
"loss": 0.0, |
|
"step": 24875 |
|
}, |
|
{ |
|
"epoch": 249.0, |
|
"learning_rate": 1.4342088607594936e-05, |
|
"loss": 0.0, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 249.25, |
|
"learning_rate": 1.4318354430379747e-05, |
|
"loss": 0.0, |
|
"step": 24925 |
|
}, |
|
{ |
|
"epoch": 249.5, |
|
"learning_rate": 1.4294620253164556e-05, |
|
"loss": 0.0, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 249.75, |
|
"learning_rate": 1.4270886075949366e-05, |
|
"loss": 0.0, |
|
"step": 24975 |
|
}, |
|
{ |
|
"epoch": 250.0, |
|
"learning_rate": 1.4247151898734177e-05, |
|
"loss": 0.0, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 250.0, |
|
"eval_loss": 0.5863945484161377, |
|
"eval_runtime": 80.5107, |
|
"eval_samples_per_second": 158.265, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.221900891042676, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 250.25, |
|
"learning_rate": 1.4223417721518987e-05, |
|
"loss": 0.0, |
|
"step": 25025 |
|
}, |
|
{ |
|
"epoch": 250.5, |
|
"learning_rate": 1.4199683544303796e-05, |
|
"loss": 0.0, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 250.75, |
|
"learning_rate": 1.4175949367088606e-05, |
|
"loss": 0.0, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 251.0, |
|
"learning_rate": 1.4152215189873417e-05, |
|
"loss": 0.0, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 251.25, |
|
"learning_rate": 1.4128481012658226e-05, |
|
"loss": 0.0, |
|
"step": 25125 |
|
}, |
|
{ |
|
"epoch": 251.5, |
|
"learning_rate": 1.4104746835443036e-05, |
|
"loss": 0.0, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 251.75, |
|
"learning_rate": 1.4081012658227847e-05, |
|
"loss": 0.0, |
|
"step": 25175 |
|
}, |
|
{ |
|
"epoch": 252.0, |
|
"learning_rate": 1.4057278481012657e-05, |
|
"loss": 0.0, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 252.25, |
|
"learning_rate": 1.4033544303797466e-05, |
|
"loss": 0.0, |
|
"step": 25225 |
|
}, |
|
{ |
|
"epoch": 252.5, |
|
"learning_rate": 1.4009810126582276e-05, |
|
"loss": 0.0, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 252.75, |
|
"learning_rate": 1.3986075949367087e-05, |
|
"loss": 0.0, |
|
"step": 25275 |
|
}, |
|
{ |
|
"epoch": 253.0, |
|
"learning_rate": 1.3962341772151898e-05, |
|
"loss": 0.0, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 253.25, |
|
"learning_rate": 1.393860759493671e-05, |
|
"loss": 0.0, |
|
"step": 25325 |
|
}, |
|
{ |
|
"epoch": 253.5, |
|
"learning_rate": 1.3914873417721519e-05, |
|
"loss": 0.0, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 253.75, |
|
"learning_rate": 1.3891139240506328e-05, |
|
"loss": 0.0, |
|
"step": 25375 |
|
}, |
|
{ |
|
"epoch": 254.0, |
|
"learning_rate": 1.3867405063291138e-05, |
|
"loss": 0.0, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 254.25, |
|
"learning_rate": 1.3843670886075949e-05, |
|
"loss": 0.0, |
|
"step": 25425 |
|
}, |
|
{ |
|
"epoch": 254.5, |
|
"learning_rate": 1.3819936708860758e-05, |
|
"loss": 0.0, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 254.75, |
|
"learning_rate": 1.3796202531645568e-05, |
|
"loss": 0.0, |
|
"step": 25475 |
|
}, |
|
{ |
|
"epoch": 255.0, |
|
"learning_rate": 1.377246835443038e-05, |
|
"loss": 0.0, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 255.25, |
|
"learning_rate": 1.3748734177215189e-05, |
|
"loss": 0.0, |
|
"step": 25525 |
|
}, |
|
{ |
|
"epoch": 255.5, |
|
"learning_rate": 1.3724999999999998e-05, |
|
"loss": 0.0, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 255.75, |
|
"learning_rate": 1.3701265822784808e-05, |
|
"loss": 0.0, |
|
"step": 25575 |
|
}, |
|
{ |
|
"epoch": 256.0, |
|
"learning_rate": 1.3677531645569619e-05, |
|
"loss": 0.0, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 256.25, |
|
"learning_rate": 1.3653797468354428e-05, |
|
"loss": 0.0, |
|
"step": 25625 |
|
}, |
|
{ |
|
"epoch": 256.5, |
|
"learning_rate": 1.3630063291139238e-05, |
|
"loss": 0.0, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 256.75, |
|
"learning_rate": 1.3606329113924049e-05, |
|
"loss": 0.0, |
|
"step": 25675 |
|
}, |
|
{ |
|
"epoch": 257.0, |
|
"learning_rate": 1.358259493670886e-05, |
|
"loss": 0.0, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 257.25, |
|
"learning_rate": 1.355886075949367e-05, |
|
"loss": 0.0, |
|
"step": 25725 |
|
}, |
|
{ |
|
"epoch": 257.5, |
|
"learning_rate": 1.3535126582278481e-05, |
|
"loss": 0.0, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 257.75, |
|
"learning_rate": 1.351139240506329e-05, |
|
"loss": 0.0, |
|
"step": 25775 |
|
}, |
|
{ |
|
"epoch": 258.0, |
|
"learning_rate": 1.34876582278481e-05, |
|
"loss": 0.0, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 258.25, |
|
"learning_rate": 1.3463924050632911e-05, |
|
"loss": 0.0, |
|
"step": 25825 |
|
}, |
|
{ |
|
"epoch": 258.5, |
|
"learning_rate": 1.344018987341772e-05, |
|
"loss": 0.0, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 258.75, |
|
"learning_rate": 1.341645569620253e-05, |
|
"loss": 0.0, |
|
"step": 25875 |
|
}, |
|
{ |
|
"epoch": 259.0, |
|
"learning_rate": 1.3392721518987341e-05, |
|
"loss": 0.0, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 259.25, |
|
"learning_rate": 1.3368987341772151e-05, |
|
"loss": 0.0, |
|
"step": 25925 |
|
}, |
|
{ |
|
"epoch": 259.5, |
|
"learning_rate": 1.334525316455696e-05, |
|
"loss": 0.0, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 259.75, |
|
"learning_rate": 1.332151898734177e-05, |
|
"loss": 0.0, |
|
"step": 25975 |
|
}, |
|
{ |
|
"epoch": 260.0, |
|
"learning_rate": 1.3297784810126581e-05, |
|
"loss": 0.0, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 260.0, |
|
"eval_loss": 0.6003339886665344, |
|
"eval_runtime": 80.5514, |
|
"eval_samples_per_second": 158.185, |
|
"eval_steps_per_second": 1.241, |
|
"eval_wer": 19.310809754572457, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 260.25, |
|
"learning_rate": 1.327405063291139e-05, |
|
"loss": 0.0, |
|
"step": 26025 |
|
}, |
|
{ |
|
"epoch": 260.5, |
|
"learning_rate": 1.32503164556962e-05, |
|
"loss": 0.0, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 260.75, |
|
"learning_rate": 1.3226582278481011e-05, |
|
"loss": 0.0, |
|
"step": 26075 |
|
}, |
|
{ |
|
"epoch": 261.0, |
|
"learning_rate": 1.3202848101265821e-05, |
|
"loss": 0.0, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 261.25, |
|
"learning_rate": 1.3179113924050632e-05, |
|
"loss": 0.0, |
|
"step": 26125 |
|
}, |
|
{ |
|
"epoch": 261.5, |
|
"learning_rate": 1.3155379746835443e-05, |
|
"loss": 0.0, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 261.75, |
|
"learning_rate": 1.3131645569620253e-05, |
|
"loss": 0.0, |
|
"step": 26175 |
|
}, |
|
{ |
|
"epoch": 262.0, |
|
"learning_rate": 1.3107911392405062e-05, |
|
"loss": 0.0, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 262.25, |
|
"learning_rate": 1.3084177215189874e-05, |
|
"loss": 0.0, |
|
"step": 26225 |
|
}, |
|
{ |
|
"epoch": 262.5, |
|
"learning_rate": 1.3060443037974683e-05, |
|
"loss": 0.0, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 262.75, |
|
"learning_rate": 1.3036708860759493e-05, |
|
"loss": 0.0, |
|
"step": 26275 |
|
}, |
|
{ |
|
"epoch": 263.0, |
|
"learning_rate": 1.3012974683544302e-05, |
|
"loss": 0.0, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 263.25, |
|
"learning_rate": 1.2989240506329113e-05, |
|
"loss": 0.0, |
|
"step": 26325 |
|
}, |
|
{ |
|
"epoch": 263.5, |
|
"learning_rate": 1.2965506329113923e-05, |
|
"loss": 0.0, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 263.75, |
|
"learning_rate": 1.2941772151898732e-05, |
|
"loss": 0.0, |
|
"step": 26375 |
|
}, |
|
{ |
|
"epoch": 264.0, |
|
"learning_rate": 1.2918037974683543e-05, |
|
"loss": 0.0, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 264.25, |
|
"learning_rate": 1.2894303797468353e-05, |
|
"loss": 0.0, |
|
"step": 26425 |
|
}, |
|
{ |
|
"epoch": 264.5, |
|
"learning_rate": 1.2870569620253162e-05, |
|
"loss": 0.0, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 264.75, |
|
"learning_rate": 1.284778481012658e-05, |
|
"loss": 0.0, |
|
"step": 26475 |
|
}, |
|
{ |
|
"epoch": 265.0, |
|
"learning_rate": 1.282405063291139e-05, |
|
"loss": 0.0, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 265.25, |
|
"learning_rate": 1.2800316455696203e-05, |
|
"loss": 0.0, |
|
"step": 26525 |
|
}, |
|
{ |
|
"epoch": 265.5, |
|
"learning_rate": 1.2776582278481012e-05, |
|
"loss": 0.0, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 265.75, |
|
"learning_rate": 1.2752848101265822e-05, |
|
"loss": 0.0, |
|
"step": 26575 |
|
}, |
|
{ |
|
"epoch": 266.0, |
|
"learning_rate": 1.2729113924050633e-05, |
|
"loss": 0.0, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 266.25, |
|
"learning_rate": 1.2705379746835443e-05, |
|
"loss": 0.0, |
|
"step": 26625 |
|
}, |
|
{ |
|
"epoch": 266.5, |
|
"learning_rate": 1.2681645569620252e-05, |
|
"loss": 0.0, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 266.75, |
|
"learning_rate": 1.2657911392405062e-05, |
|
"loss": 0.0, |
|
"step": 26675 |
|
}, |
|
{ |
|
"epoch": 267.0, |
|
"learning_rate": 1.2634177215189873e-05, |
|
"loss": 0.0, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 267.25, |
|
"learning_rate": 1.2610443037974682e-05, |
|
"loss": 0.0, |
|
"step": 26725 |
|
}, |
|
{ |
|
"epoch": 267.5, |
|
"learning_rate": 1.2586708860759492e-05, |
|
"loss": 0.0, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 267.75, |
|
"learning_rate": 1.2562974683544303e-05, |
|
"loss": 0.0, |
|
"step": 26775 |
|
}, |
|
{ |
|
"epoch": 268.0, |
|
"learning_rate": 1.2539240506329112e-05, |
|
"loss": 0.0, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 268.25, |
|
"learning_rate": 1.2515506329113922e-05, |
|
"loss": 0.0, |
|
"step": 26825 |
|
}, |
|
{ |
|
"epoch": 268.5, |
|
"learning_rate": 1.2491772151898733e-05, |
|
"loss": 0.0, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 268.75, |
|
"learning_rate": 1.2468037974683543e-05, |
|
"loss": 0.0, |
|
"step": 26875 |
|
}, |
|
{ |
|
"epoch": 269.0, |
|
"learning_rate": 1.2444303797468352e-05, |
|
"loss": 0.0, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 269.25, |
|
"learning_rate": 1.2420569620253165e-05, |
|
"loss": 0.0, |
|
"step": 26925 |
|
}, |
|
{ |
|
"epoch": 269.5, |
|
"learning_rate": 1.2396835443037975e-05, |
|
"loss": 0.0, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 269.75, |
|
"learning_rate": 1.2373101265822784e-05, |
|
"loss": 0.0, |
|
"step": 26975 |
|
}, |
|
{ |
|
"epoch": 270.0, |
|
"learning_rate": 1.2349367088607595e-05, |
|
"loss": 0.0, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 270.0, |
|
"eval_loss": 0.6140475273132324, |
|
"eval_runtime": 80.6529, |
|
"eval_samples_per_second": 157.986, |
|
"eval_steps_per_second": 1.24, |
|
"eval_wer": 19.371385024230108, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 270.25, |
|
"learning_rate": 1.2325632911392405e-05, |
|
"loss": 0.0, |
|
"step": 27025 |
|
}, |
|
{ |
|
"epoch": 270.5, |
|
"learning_rate": 1.2301898734177214e-05, |
|
"loss": 0.0, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 270.75, |
|
"learning_rate": 1.2278164556962024e-05, |
|
"loss": 0.0, |
|
"step": 27075 |
|
}, |
|
{ |
|
"epoch": 271.0, |
|
"learning_rate": 1.2254430379746835e-05, |
|
"loss": 0.0, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 271.25, |
|
"learning_rate": 1.2230696202531645e-05, |
|
"loss": 0.0, |
|
"step": 27125 |
|
}, |
|
{ |
|
"epoch": 271.5, |
|
"learning_rate": 1.2206962025316454e-05, |
|
"loss": 0.0, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 271.75, |
|
"learning_rate": 1.2183227848101265e-05, |
|
"loss": 0.0, |
|
"step": 27175 |
|
}, |
|
{ |
|
"epoch": 272.0, |
|
"learning_rate": 1.2159493670886075e-05, |
|
"loss": 0.0, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 272.25, |
|
"learning_rate": 1.2135759493670884e-05, |
|
"loss": 0.0, |
|
"step": 27225 |
|
}, |
|
{ |
|
"epoch": 272.5, |
|
"learning_rate": 1.2112025316455694e-05, |
|
"loss": 0.0, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 272.75, |
|
"learning_rate": 1.2088291139240505e-05, |
|
"loss": 0.0, |
|
"step": 27275 |
|
}, |
|
{ |
|
"epoch": 273.0, |
|
"learning_rate": 1.2064556962025314e-05, |
|
"loss": 0.0, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 273.25, |
|
"learning_rate": 1.2040822784810124e-05, |
|
"loss": 0.0, |
|
"step": 27325 |
|
}, |
|
{ |
|
"epoch": 273.5, |
|
"learning_rate": 1.2017088607594937e-05, |
|
"loss": 0.0, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 273.75, |
|
"learning_rate": 1.1993354430379746e-05, |
|
"loss": 0.0, |
|
"step": 27375 |
|
}, |
|
{ |
|
"epoch": 274.0, |
|
"learning_rate": 1.1969620253164556e-05, |
|
"loss": 0.0, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 274.25, |
|
"learning_rate": 1.1945886075949367e-05, |
|
"loss": 0.0, |
|
"step": 27425 |
|
}, |
|
{ |
|
"epoch": 274.5, |
|
"learning_rate": 1.1922151898734177e-05, |
|
"loss": 0.0, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 274.75, |
|
"learning_rate": 1.1898417721518986e-05, |
|
"loss": 0.0, |
|
"step": 27475 |
|
}, |
|
{ |
|
"epoch": 275.0, |
|
"learning_rate": 1.1874683544303797e-05, |
|
"loss": 0.0, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 275.25, |
|
"learning_rate": 1.1850949367088607e-05, |
|
"loss": 0.0, |
|
"step": 27525 |
|
}, |
|
{ |
|
"epoch": 275.5, |
|
"learning_rate": 1.1827215189873416e-05, |
|
"loss": 0.0, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 275.75, |
|
"learning_rate": 1.1803481012658226e-05, |
|
"loss": 0.0, |
|
"step": 27575 |
|
}, |
|
{ |
|
"epoch": 276.0, |
|
"learning_rate": 1.1779746835443037e-05, |
|
"loss": 0.0, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 276.25, |
|
"learning_rate": 1.1756012658227847e-05, |
|
"loss": 0.0, |
|
"step": 27625 |
|
}, |
|
{ |
|
"epoch": 276.5, |
|
"learning_rate": 1.1732278481012656e-05, |
|
"loss": 0.0, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 276.75, |
|
"learning_rate": 1.1708544303797467e-05, |
|
"loss": 0.0, |
|
"step": 27675 |
|
}, |
|
{ |
|
"epoch": 277.0, |
|
"learning_rate": 1.1684810126582277e-05, |
|
"loss": 0.0, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 277.25, |
|
"learning_rate": 1.1661075949367086e-05, |
|
"loss": 0.0, |
|
"step": 27725 |
|
}, |
|
{ |
|
"epoch": 277.5, |
|
"learning_rate": 1.1637341772151899e-05, |
|
"loss": 0.0, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 277.75, |
|
"learning_rate": 1.1613607594936709e-05, |
|
"loss": 0.0, |
|
"step": 27775 |
|
}, |
|
{ |
|
"epoch": 278.0, |
|
"learning_rate": 1.1589873417721518e-05, |
|
"loss": 0.0, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 278.25, |
|
"learning_rate": 1.156613924050633e-05, |
|
"loss": 0.0, |
|
"step": 27825 |
|
}, |
|
{ |
|
"epoch": 278.5, |
|
"learning_rate": 1.1542405063291139e-05, |
|
"loss": 0.0, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 278.75, |
|
"learning_rate": 1.1518670886075948e-05, |
|
"loss": 0.0, |
|
"step": 27875 |
|
}, |
|
{ |
|
"epoch": 279.0, |
|
"learning_rate": 1.1496835443037973e-05, |
|
"loss": 0.0033, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 279.25, |
|
"learning_rate": 1.1474050632911391e-05, |
|
"loss": 0.0072, |
|
"step": 27925 |
|
}, |
|
{ |
|
"epoch": 279.5, |
|
"learning_rate": 1.14503164556962e-05, |
|
"loss": 0.0052, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 279.75, |
|
"learning_rate": 1.1426582278481012e-05, |
|
"loss": 0.004, |
|
"step": 27975 |
|
}, |
|
{ |
|
"epoch": 280.0, |
|
"learning_rate": 1.1402848101265821e-05, |
|
"loss": 0.0034, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 280.0, |
|
"eval_loss": 0.5535955429077148, |
|
"eval_runtime": 81.9868, |
|
"eval_samples_per_second": 155.415, |
|
"eval_steps_per_second": 1.22, |
|
"eval_wer": 20.663006096607788, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 280.25, |
|
"learning_rate": 1.137911392405063e-05, |
|
"loss": 0.0018, |
|
"step": 28025 |
|
}, |
|
{ |
|
"epoch": 280.5, |
|
"learning_rate": 1.1355379746835442e-05, |
|
"loss": 0.0018, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 280.75, |
|
"learning_rate": 1.1331645569620251e-05, |
|
"loss": 0.0015, |
|
"step": 28075 |
|
}, |
|
{ |
|
"epoch": 281.0, |
|
"learning_rate": 1.1307911392405061e-05, |
|
"loss": 0.0015, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 281.25, |
|
"learning_rate": 1.1284177215189874e-05, |
|
"loss": 0.0003, |
|
"step": 28125 |
|
}, |
|
{ |
|
"epoch": 281.5, |
|
"learning_rate": 1.1260443037974683e-05, |
|
"loss": 0.0004, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 281.75, |
|
"learning_rate": 1.1236708860759493e-05, |
|
"loss": 0.0004, |
|
"step": 28175 |
|
}, |
|
{ |
|
"epoch": 282.0, |
|
"learning_rate": 1.1212974683544304e-05, |
|
"loss": 0.0003, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 282.25, |
|
"learning_rate": 1.1189240506329114e-05, |
|
"loss": 0.0001, |
|
"step": 28225 |
|
}, |
|
{ |
|
"epoch": 282.5, |
|
"learning_rate": 1.1165506329113923e-05, |
|
"loss": 0.0001, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 282.75, |
|
"learning_rate": 1.1141772151898733e-05, |
|
"loss": 0.0001, |
|
"step": 28275 |
|
}, |
|
{ |
|
"epoch": 283.0, |
|
"learning_rate": 1.1118037974683544e-05, |
|
"loss": 0.0001, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 283.25, |
|
"learning_rate": 1.1094303797468353e-05, |
|
"loss": 0.0001, |
|
"step": 28325 |
|
}, |
|
{ |
|
"epoch": 283.5, |
|
"learning_rate": 1.1070569620253163e-05, |
|
"loss": 0.0001, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 283.75, |
|
"learning_rate": 1.1046835443037974e-05, |
|
"loss": 0.0, |
|
"step": 28375 |
|
}, |
|
{ |
|
"epoch": 284.0, |
|
"learning_rate": 1.1023101265822783e-05, |
|
"loss": 0.0003, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 284.25, |
|
"learning_rate": 1.0999367088607593e-05, |
|
"loss": 0.0, |
|
"step": 28425 |
|
}, |
|
{ |
|
"epoch": 284.5, |
|
"learning_rate": 1.0975632911392404e-05, |
|
"loss": 0.0, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 284.75, |
|
"learning_rate": 1.0951898734177214e-05, |
|
"loss": 0.0, |
|
"step": 28475 |
|
}, |
|
{ |
|
"epoch": 285.0, |
|
"learning_rate": 1.0928164556962023e-05, |
|
"loss": 0.0, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 285.25, |
|
"learning_rate": 1.0904430379746836e-05, |
|
"loss": 0.0, |
|
"step": 28525 |
|
}, |
|
{ |
|
"epoch": 285.5, |
|
"learning_rate": 1.0880696202531646e-05, |
|
"loss": 0.0, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 285.75, |
|
"learning_rate": 1.0856962025316455e-05, |
|
"loss": 0.0, |
|
"step": 28575 |
|
}, |
|
{ |
|
"epoch": 286.0, |
|
"learning_rate": 1.0833227848101266e-05, |
|
"loss": 0.0, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 286.25, |
|
"learning_rate": 1.0809493670886076e-05, |
|
"loss": 0.0, |
|
"step": 28625 |
|
}, |
|
{ |
|
"epoch": 286.5, |
|
"learning_rate": 1.0785759493670885e-05, |
|
"loss": 0.0, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 286.75, |
|
"learning_rate": 1.0762025316455695e-05, |
|
"loss": 0.0, |
|
"step": 28675 |
|
}, |
|
{ |
|
"epoch": 287.0, |
|
"learning_rate": 1.0738291139240506e-05, |
|
"loss": 0.0, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 287.25, |
|
"learning_rate": 1.0714556962025316e-05, |
|
"loss": 0.0, |
|
"step": 28725 |
|
}, |
|
{ |
|
"epoch": 287.5, |
|
"learning_rate": 1.0690822784810125e-05, |
|
"loss": 0.0, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 287.75, |
|
"learning_rate": 1.0667088607594936e-05, |
|
"loss": 0.0, |
|
"step": 28775 |
|
}, |
|
{ |
|
"epoch": 288.0, |
|
"learning_rate": 1.0643354430379746e-05, |
|
"loss": 0.0, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 288.25, |
|
"learning_rate": 1.0619620253164555e-05, |
|
"loss": 0.0, |
|
"step": 28825 |
|
}, |
|
{ |
|
"epoch": 288.5, |
|
"learning_rate": 1.0595886075949365e-05, |
|
"loss": 0.0, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 288.75, |
|
"learning_rate": 1.0572151898734176e-05, |
|
"loss": 0.0, |
|
"step": 28875 |
|
}, |
|
{ |
|
"epoch": 289.0, |
|
"learning_rate": 1.0548417721518985e-05, |
|
"loss": 0.0, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 289.25, |
|
"learning_rate": 1.0524683544303795e-05, |
|
"loss": 0.0, |
|
"step": 28925 |
|
}, |
|
{ |
|
"epoch": 289.5, |
|
"learning_rate": 1.0500949367088608e-05, |
|
"loss": 0.0, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 289.75, |
|
"learning_rate": 1.0477215189873417e-05, |
|
"loss": 0.0, |
|
"step": 28975 |
|
}, |
|
{ |
|
"epoch": 290.0, |
|
"learning_rate": 1.0453481012658227e-05, |
|
"loss": 0.0, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 290.0, |
|
"eval_loss": 0.5485877394676208, |
|
"eval_runtime": 80.6015, |
|
"eval_samples_per_second": 158.086, |
|
"eval_steps_per_second": 1.241, |
|
"eval_wer": 19.339143348444583, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 290.25, |
|
"learning_rate": 1.0429746835443038e-05, |
|
"loss": 0.0, |
|
"step": 29025 |
|
}, |
|
{ |
|
"epoch": 290.5, |
|
"learning_rate": 1.0406012658227848e-05, |
|
"loss": 0.0, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 290.75, |
|
"learning_rate": 1.0382278481012657e-05, |
|
"loss": 0.0, |
|
"step": 29075 |
|
}, |
|
{ |
|
"epoch": 291.0, |
|
"learning_rate": 1.0358544303797468e-05, |
|
"loss": 0.0, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 291.25, |
|
"learning_rate": 1.0334810126582278e-05, |
|
"loss": 0.0, |
|
"step": 29125 |
|
}, |
|
{ |
|
"epoch": 291.5, |
|
"learning_rate": 1.0311075949367087e-05, |
|
"loss": 0.0, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 291.75, |
|
"learning_rate": 1.0287341772151897e-05, |
|
"loss": 0.0, |
|
"step": 29175 |
|
}, |
|
{ |
|
"epoch": 292.0, |
|
"learning_rate": 1.0263607594936708e-05, |
|
"loss": 0.0, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 292.25, |
|
"learning_rate": 1.0239873417721518e-05, |
|
"loss": 0.0, |
|
"step": 29225 |
|
}, |
|
{ |
|
"epoch": 292.5, |
|
"learning_rate": 1.0216139240506327e-05, |
|
"loss": 0.0, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 292.75, |
|
"learning_rate": 1.0192405063291138e-05, |
|
"loss": 0.0, |
|
"step": 29275 |
|
}, |
|
{ |
|
"epoch": 293.0, |
|
"learning_rate": 1.0168670886075948e-05, |
|
"loss": 0.0, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 293.25, |
|
"learning_rate": 1.0144936708860757e-05, |
|
"loss": 0.0, |
|
"step": 29325 |
|
}, |
|
{ |
|
"epoch": 293.5, |
|
"learning_rate": 1.012120253164557e-05, |
|
"loss": 0.0, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 293.75, |
|
"learning_rate": 1.009746835443038e-05, |
|
"loss": 0.0, |
|
"step": 29375 |
|
}, |
|
{ |
|
"epoch": 294.0, |
|
"learning_rate": 1.0073734177215189e-05, |
|
"loss": 0.0, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 294.25, |
|
"learning_rate": 1.005e-05, |
|
"loss": 0.0, |
|
"step": 29425 |
|
}, |
|
{ |
|
"epoch": 294.5, |
|
"learning_rate": 1.002626582278481e-05, |
|
"loss": 0.0, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 294.75, |
|
"learning_rate": 1.000253164556962e-05, |
|
"loss": 0.0, |
|
"step": 29475 |
|
}, |
|
{ |
|
"epoch": 295.0, |
|
"learning_rate": 9.97879746835443e-06, |
|
"loss": 0.0, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 295.25, |
|
"learning_rate": 9.95506329113924e-06, |
|
"loss": 0.0, |
|
"step": 29525 |
|
}, |
|
{ |
|
"epoch": 295.5, |
|
"learning_rate": 9.93132911392405e-06, |
|
"loss": 0.0, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 295.75, |
|
"learning_rate": 9.907594936708859e-06, |
|
"loss": 0.0, |
|
"step": 29575 |
|
}, |
|
{ |
|
"epoch": 296.0, |
|
"learning_rate": 9.88386075949367e-06, |
|
"loss": 0.0, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 296.25, |
|
"learning_rate": 9.86012658227848e-06, |
|
"loss": 0.0, |
|
"step": 29625 |
|
}, |
|
{ |
|
"epoch": 296.5, |
|
"learning_rate": 9.83639240506329e-06, |
|
"loss": 0.0, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 296.75, |
|
"learning_rate": 9.8126582278481e-06, |
|
"loss": 0.0, |
|
"step": 29675 |
|
}, |
|
{ |
|
"epoch": 297.0, |
|
"learning_rate": 9.78892405063291e-06, |
|
"loss": 0.0, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 297.25, |
|
"learning_rate": 9.76518987341772e-06, |
|
"loss": 0.0, |
|
"step": 29725 |
|
}, |
|
{ |
|
"epoch": 297.5, |
|
"learning_rate": 9.741455696202532e-06, |
|
"loss": 0.0, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 297.75, |
|
"learning_rate": 9.717721518987342e-06, |
|
"loss": 0.0, |
|
"step": 29775 |
|
}, |
|
{ |
|
"epoch": 298.0, |
|
"learning_rate": 9.693987341772151e-06, |
|
"loss": 0.0, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 298.25, |
|
"learning_rate": 9.670253164556963e-06, |
|
"loss": 0.0, |
|
"step": 29825 |
|
}, |
|
{ |
|
"epoch": 298.5, |
|
"learning_rate": 9.646518987341772e-06, |
|
"loss": 0.0, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 298.75, |
|
"learning_rate": 9.622784810126582e-06, |
|
"loss": 0.0, |
|
"step": 29875 |
|
}, |
|
{ |
|
"epoch": 299.0, |
|
"learning_rate": 9.599050632911391e-06, |
|
"loss": 0.0, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 299.25, |
|
"learning_rate": 9.575316455696202e-06, |
|
"loss": 0.0, |
|
"step": 29925 |
|
}, |
|
{ |
|
"epoch": 299.5, |
|
"learning_rate": 9.551582278481012e-06, |
|
"loss": 0.0, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 299.75, |
|
"learning_rate": 9.527848101265821e-06, |
|
"loss": 0.0, |
|
"step": 29975 |
|
}, |
|
{ |
|
"epoch": 300.0, |
|
"learning_rate": 9.504113924050633e-06, |
|
"loss": 0.0, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 300.0, |
|
"eval_loss": 0.5590522885322571, |
|
"eval_runtime": 80.5922, |
|
"eval_samples_per_second": 158.105, |
|
"eval_steps_per_second": 1.241, |
|
"eval_wer": 19.30592465218071, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 300.25, |
|
"learning_rate": 9.480379746835442e-06, |
|
"loss": 0.0, |
|
"step": 30025 |
|
}, |
|
{ |
|
"epoch": 300.5, |
|
"learning_rate": 9.456645569620252e-06, |
|
"loss": 0.0, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 300.75, |
|
"learning_rate": 9.432911392405061e-06, |
|
"loss": 0.0, |
|
"step": 30075 |
|
}, |
|
{ |
|
"epoch": 301.0, |
|
"learning_rate": 9.409177215189872e-06, |
|
"loss": 0.0, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 301.25, |
|
"learning_rate": 9.385443037974682e-06, |
|
"loss": 0.0, |
|
"step": 30125 |
|
}, |
|
{ |
|
"epoch": 301.5, |
|
"learning_rate": 9.361708860759493e-06, |
|
"loss": 0.0, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 301.75, |
|
"learning_rate": 9.337974683544302e-06, |
|
"loss": 0.0, |
|
"step": 30175 |
|
}, |
|
{ |
|
"epoch": 302.0, |
|
"learning_rate": 9.314240506329114e-06, |
|
"loss": 0.0, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 302.25, |
|
"learning_rate": 9.290506329113923e-06, |
|
"loss": 0.0, |
|
"step": 30225 |
|
}, |
|
{ |
|
"epoch": 302.5, |
|
"learning_rate": 9.266772151898733e-06, |
|
"loss": 0.0, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 302.75, |
|
"learning_rate": 9.243037974683544e-06, |
|
"loss": 0.0, |
|
"step": 30275 |
|
}, |
|
{ |
|
"epoch": 303.0, |
|
"learning_rate": 9.219303797468353e-06, |
|
"loss": 0.0, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 303.25, |
|
"learning_rate": 9.195569620253165e-06, |
|
"loss": 0.0, |
|
"step": 30325 |
|
}, |
|
{ |
|
"epoch": 303.5, |
|
"learning_rate": 9.171835443037974e-06, |
|
"loss": 0.0, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 303.75, |
|
"learning_rate": 9.148101265822784e-06, |
|
"loss": 0.0, |
|
"step": 30375 |
|
}, |
|
{ |
|
"epoch": 304.0, |
|
"learning_rate": 9.124367088607593e-06, |
|
"loss": 0.0, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 304.25, |
|
"learning_rate": 9.100632911392404e-06, |
|
"loss": 0.0, |
|
"step": 30425 |
|
}, |
|
{ |
|
"epoch": 304.5, |
|
"learning_rate": 9.076898734177214e-06, |
|
"loss": 0.0, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 304.75, |
|
"learning_rate": 9.053164556962025e-06, |
|
"loss": 0.0, |
|
"step": 30475 |
|
}, |
|
{ |
|
"epoch": 305.0, |
|
"learning_rate": 9.029430379746835e-06, |
|
"loss": 0.0, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 305.25, |
|
"learning_rate": 9.005696202531646e-06, |
|
"loss": 0.0, |
|
"step": 30525 |
|
}, |
|
{ |
|
"epoch": 305.5, |
|
"learning_rate": 8.981962025316455e-06, |
|
"loss": 0.0, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 305.75, |
|
"learning_rate": 8.958227848101265e-06, |
|
"loss": 0.0, |
|
"step": 30575 |
|
}, |
|
{ |
|
"epoch": 306.0, |
|
"learning_rate": 8.934493670886074e-06, |
|
"loss": 0.0, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 306.25, |
|
"learning_rate": 8.910759493670885e-06, |
|
"loss": 0.0, |
|
"step": 30625 |
|
}, |
|
{ |
|
"epoch": 306.5, |
|
"learning_rate": 8.887025316455695e-06, |
|
"loss": 0.0, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 306.75, |
|
"learning_rate": 8.863291139240504e-06, |
|
"loss": 0.0, |
|
"step": 30675 |
|
}, |
|
{ |
|
"epoch": 307.0, |
|
"learning_rate": 8.839556962025316e-06, |
|
"loss": 0.0, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 307.25, |
|
"learning_rate": 8.815822784810127e-06, |
|
"loss": 0.0, |
|
"step": 30725 |
|
}, |
|
{ |
|
"epoch": 307.5, |
|
"learning_rate": 8.792088607594936e-06, |
|
"loss": 0.0, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 307.75, |
|
"learning_rate": 8.768354430379746e-06, |
|
"loss": 0.0, |
|
"step": 30775 |
|
}, |
|
{ |
|
"epoch": 308.0, |
|
"learning_rate": 8.744620253164555e-06, |
|
"loss": 0.0, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 308.25, |
|
"learning_rate": 8.720886075949367e-06, |
|
"loss": 0.0, |
|
"step": 30825 |
|
}, |
|
{ |
|
"epoch": 308.5, |
|
"learning_rate": 8.697151898734176e-06, |
|
"loss": 0.0, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 308.75, |
|
"learning_rate": 8.673417721518986e-06, |
|
"loss": 0.0, |
|
"step": 30875 |
|
}, |
|
{ |
|
"epoch": 309.0, |
|
"learning_rate": 8.649683544303797e-06, |
|
"loss": 0.0, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 309.25, |
|
"learning_rate": 8.625949367088606e-06, |
|
"loss": 0.0, |
|
"step": 30925 |
|
}, |
|
{ |
|
"epoch": 309.5, |
|
"learning_rate": 8.602215189873418e-06, |
|
"loss": 0.0, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 309.75, |
|
"learning_rate": 8.578481012658227e-06, |
|
"loss": 0.0, |
|
"step": 30975 |
|
}, |
|
{ |
|
"epoch": 310.0, |
|
"learning_rate": 8.554746835443037e-06, |
|
"loss": 0.0, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 310.0, |
|
"eval_loss": 0.566923975944519, |
|
"eval_runtime": 80.5145, |
|
"eval_samples_per_second": 158.257, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.313740816007506, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 310.25, |
|
"learning_rate": 8.531012658227848e-06, |
|
"loss": 0.0, |
|
"step": 31025 |
|
}, |
|
{ |
|
"epoch": 310.5, |
|
"learning_rate": 8.507278481012657e-06, |
|
"loss": 0.0, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 310.75, |
|
"learning_rate": 8.483544303797467e-06, |
|
"loss": 0.0, |
|
"step": 31075 |
|
}, |
|
{ |
|
"epoch": 311.0, |
|
"learning_rate": 8.459810126582278e-06, |
|
"loss": 0.0, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 311.25, |
|
"learning_rate": 8.436075949367087e-06, |
|
"loss": 0.0, |
|
"step": 31125 |
|
}, |
|
{ |
|
"epoch": 311.5, |
|
"learning_rate": 8.412341772151899e-06, |
|
"loss": 0.0, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 311.75, |
|
"learning_rate": 8.388607594936708e-06, |
|
"loss": 0.0, |
|
"step": 31175 |
|
}, |
|
{ |
|
"epoch": 312.0, |
|
"learning_rate": 8.364873417721518e-06, |
|
"loss": 0.0, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 312.25, |
|
"learning_rate": 8.341139240506329e-06, |
|
"loss": 0.0, |
|
"step": 31225 |
|
}, |
|
{ |
|
"epoch": 312.5, |
|
"learning_rate": 8.317405063291138e-06, |
|
"loss": 0.0, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 312.75, |
|
"learning_rate": 8.293670886075948e-06, |
|
"loss": 0.0, |
|
"step": 31275 |
|
}, |
|
{ |
|
"epoch": 313.0, |
|
"learning_rate": 8.269936708860759e-06, |
|
"loss": 0.0, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 313.25, |
|
"learning_rate": 8.246202531645569e-06, |
|
"loss": 0.0, |
|
"step": 31325 |
|
}, |
|
{ |
|
"epoch": 313.5, |
|
"learning_rate": 8.22246835443038e-06, |
|
"loss": 0.0, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 313.75, |
|
"learning_rate": 8.19873417721519e-06, |
|
"loss": 0.0, |
|
"step": 31375 |
|
}, |
|
{ |
|
"epoch": 314.0, |
|
"learning_rate": 8.174999999999999e-06, |
|
"loss": 0.0, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 314.25, |
|
"learning_rate": 8.15126582278481e-06, |
|
"loss": 0.0, |
|
"step": 31425 |
|
}, |
|
{ |
|
"epoch": 314.5, |
|
"learning_rate": 8.12753164556962e-06, |
|
"loss": 0.0, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 314.75, |
|
"learning_rate": 8.103797468354429e-06, |
|
"loss": 0.0, |
|
"step": 31475 |
|
}, |
|
{ |
|
"epoch": 315.0, |
|
"learning_rate": 8.08006329113924e-06, |
|
"loss": 0.0, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 315.25, |
|
"learning_rate": 8.05632911392405e-06, |
|
"loss": 0.0, |
|
"step": 31525 |
|
}, |
|
{ |
|
"epoch": 315.5, |
|
"learning_rate": 8.032594936708861e-06, |
|
"loss": 0.0, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 315.75, |
|
"learning_rate": 8.00886075949367e-06, |
|
"loss": 0.0, |
|
"step": 31575 |
|
}, |
|
{ |
|
"epoch": 316.0, |
|
"learning_rate": 7.98512658227848e-06, |
|
"loss": 0.0, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 316.25, |
|
"learning_rate": 7.961392405063291e-06, |
|
"loss": 0.0, |
|
"step": 31625 |
|
}, |
|
{ |
|
"epoch": 316.5, |
|
"learning_rate": 7.9376582278481e-06, |
|
"loss": 0.0, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 316.75, |
|
"learning_rate": 7.91392405063291e-06, |
|
"loss": 0.0, |
|
"step": 31675 |
|
}, |
|
{ |
|
"epoch": 317.0, |
|
"learning_rate": 7.89018987341772e-06, |
|
"loss": 0.0, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 317.25, |
|
"learning_rate": 7.866455696202531e-06, |
|
"loss": 0.0, |
|
"step": 31725 |
|
}, |
|
{ |
|
"epoch": 317.5, |
|
"learning_rate": 7.842721518987342e-06, |
|
"loss": 0.0, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 317.75, |
|
"learning_rate": 7.818987341772152e-06, |
|
"loss": 0.0, |
|
"step": 31775 |
|
}, |
|
{ |
|
"epoch": 318.0, |
|
"learning_rate": 7.795253164556961e-06, |
|
"loss": 0.0, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 318.25, |
|
"learning_rate": 7.77151898734177e-06, |
|
"loss": 0.0, |
|
"step": 31825 |
|
}, |
|
{ |
|
"epoch": 318.5, |
|
"learning_rate": 7.747784810126582e-06, |
|
"loss": 0.0, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 318.75, |
|
"learning_rate": 7.724050632911391e-06, |
|
"loss": 0.0, |
|
"step": 31875 |
|
}, |
|
{ |
|
"epoch": 319.0, |
|
"learning_rate": 7.7003164556962e-06, |
|
"loss": 0.0, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 319.25, |
|
"learning_rate": 7.676582278481012e-06, |
|
"loss": 0.0, |
|
"step": 31925 |
|
}, |
|
{ |
|
"epoch": 319.5, |
|
"learning_rate": 7.652848101265823e-06, |
|
"loss": 0.0, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 319.75, |
|
"learning_rate": 7.629113924050633e-06, |
|
"loss": 0.0, |
|
"step": 31975 |
|
}, |
|
{ |
|
"epoch": 320.0, |
|
"learning_rate": 7.605379746835442e-06, |
|
"loss": 0.0, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 320.0, |
|
"eval_loss": 0.573691189289093, |
|
"eval_runtime": 80.5487, |
|
"eval_samples_per_second": 158.19, |
|
"eval_steps_per_second": 1.241, |
|
"eval_wer": 19.322534000312647, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 320.25, |
|
"learning_rate": 7.5816455696202526e-06, |
|
"loss": 0.0, |
|
"step": 32025 |
|
}, |
|
{ |
|
"epoch": 320.5, |
|
"learning_rate": 7.557911392405062e-06, |
|
"loss": 0.0, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 320.75, |
|
"learning_rate": 7.5341772151898724e-06, |
|
"loss": 0.0, |
|
"step": 32075 |
|
}, |
|
{ |
|
"epoch": 321.0, |
|
"learning_rate": 7.510443037974683e-06, |
|
"loss": 0.0, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 321.25, |
|
"learning_rate": 7.486708860759493e-06, |
|
"loss": 0.0, |
|
"step": 32125 |
|
}, |
|
{ |
|
"epoch": 321.5, |
|
"learning_rate": 7.4629746835443035e-06, |
|
"loss": 0.0, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 321.75, |
|
"learning_rate": 7.439240506329114e-06, |
|
"loss": 0.0, |
|
"step": 32175 |
|
}, |
|
{ |
|
"epoch": 322.0, |
|
"learning_rate": 7.415506329113923e-06, |
|
"loss": 0.0, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 322.25, |
|
"learning_rate": 7.391772151898734e-06, |
|
"loss": 0.0, |
|
"step": 32225 |
|
}, |
|
{ |
|
"epoch": 322.5, |
|
"learning_rate": 7.368037974683543e-06, |
|
"loss": 0.0, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 322.75, |
|
"learning_rate": 7.3443037974683536e-06, |
|
"loss": 0.0, |
|
"step": 32275 |
|
}, |
|
{ |
|
"epoch": 323.0, |
|
"learning_rate": 7.320569620253164e-06, |
|
"loss": 0.0, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 323.25, |
|
"learning_rate": 7.296835443037974e-06, |
|
"loss": 0.0, |
|
"step": 32325 |
|
}, |
|
{ |
|
"epoch": 323.5, |
|
"learning_rate": 7.273101265822785e-06, |
|
"loss": 0.0, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 323.75, |
|
"learning_rate": 7.249367088607594e-06, |
|
"loss": 0.0, |
|
"step": 32375 |
|
}, |
|
{ |
|
"epoch": 324.0, |
|
"learning_rate": 7.2256329113924045e-06, |
|
"loss": 0.0, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 324.25, |
|
"learning_rate": 7.201898734177215e-06, |
|
"loss": 0.0, |
|
"step": 32425 |
|
}, |
|
{ |
|
"epoch": 324.5, |
|
"learning_rate": 7.178164556962024e-06, |
|
"loss": 0.0, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 324.75, |
|
"learning_rate": 7.154430379746835e-06, |
|
"loss": 0.0, |
|
"step": 32475 |
|
}, |
|
{ |
|
"epoch": 325.0, |
|
"learning_rate": 7.130696202531644e-06, |
|
"loss": 0.0, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 325.25, |
|
"learning_rate": 7.1069620253164546e-06, |
|
"loss": 0.0, |
|
"step": 32525 |
|
}, |
|
{ |
|
"epoch": 325.5, |
|
"learning_rate": 7.083227848101266e-06, |
|
"loss": 0.0, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 325.75, |
|
"learning_rate": 7.059493670886075e-06, |
|
"loss": 0.0, |
|
"step": 32575 |
|
}, |
|
{ |
|
"epoch": 326.0, |
|
"learning_rate": 7.035759493670886e-06, |
|
"loss": 0.0, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 326.25, |
|
"learning_rate": 7.012025316455696e-06, |
|
"loss": 0.0, |
|
"step": 32625 |
|
}, |
|
{ |
|
"epoch": 326.5, |
|
"learning_rate": 6.9882911392405055e-06, |
|
"loss": 0.0, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 326.75, |
|
"learning_rate": 6.964556962025316e-06, |
|
"loss": 0.0, |
|
"step": 32675 |
|
}, |
|
{ |
|
"epoch": 327.0, |
|
"learning_rate": 6.940822784810125e-06, |
|
"loss": 0.0, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 327.25, |
|
"learning_rate": 6.917088607594936e-06, |
|
"loss": 0.0, |
|
"step": 32725 |
|
}, |
|
{ |
|
"epoch": 327.5, |
|
"learning_rate": 6.893354430379747e-06, |
|
"loss": 0.0, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 327.75, |
|
"learning_rate": 6.869620253164556e-06, |
|
"loss": 0.0, |
|
"step": 32775 |
|
}, |
|
{ |
|
"epoch": 328.0, |
|
"learning_rate": 6.845886075949367e-06, |
|
"loss": 0.0, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 328.25, |
|
"learning_rate": 6.822151898734176e-06, |
|
"loss": 0.0, |
|
"step": 32825 |
|
}, |
|
{ |
|
"epoch": 328.5, |
|
"learning_rate": 6.798417721518987e-06, |
|
"loss": 0.0, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 328.75, |
|
"learning_rate": 6.774683544303797e-06, |
|
"loss": 0.0, |
|
"step": 32875 |
|
}, |
|
{ |
|
"epoch": 329.0, |
|
"learning_rate": 6.7509493670886065e-06, |
|
"loss": 0.0, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 329.25, |
|
"learning_rate": 6.727215189873417e-06, |
|
"loss": 0.0, |
|
"step": 32925 |
|
}, |
|
{ |
|
"epoch": 329.5, |
|
"learning_rate": 6.703481012658228e-06, |
|
"loss": 0.0, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 329.75, |
|
"learning_rate": 6.6797468354430375e-06, |
|
"loss": 0.0, |
|
"step": 32975 |
|
}, |
|
{ |
|
"epoch": 330.0, |
|
"learning_rate": 6.656012658227848e-06, |
|
"loss": 0.0, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 330.0, |
|
"eval_loss": 0.5798343420028687, |
|
"eval_runtime": 80.7786, |
|
"eval_samples_per_second": 157.74, |
|
"eval_steps_per_second": 1.238, |
|
"eval_wer": 19.288338283570425, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 330.25, |
|
"learning_rate": 6.632278481012657e-06, |
|
"loss": 0.0, |
|
"step": 33025 |
|
}, |
|
{ |
|
"epoch": 330.5, |
|
"learning_rate": 6.608544303797468e-06, |
|
"loss": 0.0, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 330.75, |
|
"learning_rate": 6.584810126582278e-06, |
|
"loss": 0.0, |
|
"step": 33075 |
|
}, |
|
{ |
|
"epoch": 331.0, |
|
"learning_rate": 6.561075949367088e-06, |
|
"loss": 0.0, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 331.25, |
|
"learning_rate": 6.537341772151898e-06, |
|
"loss": 0.0, |
|
"step": 33125 |
|
}, |
|
{ |
|
"epoch": 331.5, |
|
"learning_rate": 6.513607594936709e-06, |
|
"loss": 0.0, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 331.75, |
|
"learning_rate": 6.489873417721519e-06, |
|
"loss": 0.0, |
|
"step": 33175 |
|
}, |
|
{ |
|
"epoch": 332.0, |
|
"learning_rate": 6.466139240506329e-06, |
|
"loss": 0.0, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 332.25, |
|
"learning_rate": 6.4424050632911385e-06, |
|
"loss": 0.0, |
|
"step": 33225 |
|
}, |
|
{ |
|
"epoch": 332.5, |
|
"learning_rate": 6.418670886075949e-06, |
|
"loss": 0.0, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 332.75, |
|
"learning_rate": 6.394936708860758e-06, |
|
"loss": 0.0, |
|
"step": 33275 |
|
}, |
|
{ |
|
"epoch": 333.0, |
|
"learning_rate": 6.371202531645569e-06, |
|
"loss": 0.0, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 333.25, |
|
"learning_rate": 6.347468354430379e-06, |
|
"loss": 0.0, |
|
"step": 33325 |
|
}, |
|
{ |
|
"epoch": 333.5, |
|
"learning_rate": 6.3237341772151895e-06, |
|
"loss": 0.0, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 333.75, |
|
"learning_rate": 6.3e-06, |
|
"loss": 0.0, |
|
"step": 33375 |
|
}, |
|
{ |
|
"epoch": 334.0, |
|
"learning_rate": 6.27626582278481e-06, |
|
"loss": 0.0, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 334.25, |
|
"learning_rate": 6.25253164556962e-06, |
|
"loss": 0.0, |
|
"step": 33425 |
|
}, |
|
{ |
|
"epoch": 334.5, |
|
"learning_rate": 6.22879746835443e-06, |
|
"loss": 0.0, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 334.75, |
|
"learning_rate": 6.2050632911392395e-06, |
|
"loss": 0.0, |
|
"step": 33475 |
|
}, |
|
{ |
|
"epoch": 335.0, |
|
"learning_rate": 6.18132911392405e-06, |
|
"loss": 0.0, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 335.25, |
|
"learning_rate": 6.15759493670886e-06, |
|
"loss": 0.0, |
|
"step": 33525 |
|
}, |
|
{ |
|
"epoch": 335.5, |
|
"learning_rate": 6.13386075949367e-06, |
|
"loss": 0.0, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 335.75, |
|
"learning_rate": 6.110126582278481e-06, |
|
"loss": 0.0, |
|
"step": 33575 |
|
}, |
|
{ |
|
"epoch": 336.0, |
|
"learning_rate": 6.086392405063291e-06, |
|
"loss": 0.0, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 336.25, |
|
"learning_rate": 6.062658227848101e-06, |
|
"loss": 0.0, |
|
"step": 33625 |
|
}, |
|
{ |
|
"epoch": 336.5, |
|
"learning_rate": 6.038924050632911e-06, |
|
"loss": 0.0, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 336.75, |
|
"learning_rate": 6.015189873417721e-06, |
|
"loss": 0.0, |
|
"step": 33675 |
|
}, |
|
{ |
|
"epoch": 337.0, |
|
"learning_rate": 5.991455696202531e-06, |
|
"loss": 0.0, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 337.25, |
|
"learning_rate": 5.9677215189873405e-06, |
|
"loss": 0.0, |
|
"step": 33725 |
|
}, |
|
{ |
|
"epoch": 337.5, |
|
"learning_rate": 5.943987341772151e-06, |
|
"loss": 0.0, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 337.75, |
|
"learning_rate": 5.920253164556962e-06, |
|
"loss": 0.0, |
|
"step": 33775 |
|
}, |
|
{ |
|
"epoch": 338.0, |
|
"learning_rate": 5.896518987341772e-06, |
|
"loss": 0.0, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 338.25, |
|
"learning_rate": 5.872784810126582e-06, |
|
"loss": 0.0, |
|
"step": 33825 |
|
}, |
|
{ |
|
"epoch": 338.5, |
|
"learning_rate": 5.849050632911392e-06, |
|
"loss": 0.0, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 338.75, |
|
"learning_rate": 5.825316455696202e-06, |
|
"loss": 0.0, |
|
"step": 33875 |
|
}, |
|
{ |
|
"epoch": 339.0, |
|
"learning_rate": 5.801582278481012e-06, |
|
"loss": 0.0, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 339.25, |
|
"learning_rate": 5.777848101265822e-06, |
|
"loss": 0.0, |
|
"step": 33925 |
|
}, |
|
{ |
|
"epoch": 339.5, |
|
"learning_rate": 5.754113924050632e-06, |
|
"loss": 0.0, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 339.75, |
|
"learning_rate": 5.730379746835443e-06, |
|
"loss": 0.0, |
|
"step": 33975 |
|
}, |
|
{ |
|
"epoch": 340.0, |
|
"learning_rate": 5.706645569620253e-06, |
|
"loss": 0.0, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 340.0, |
|
"eval_loss": 0.5856410264968872, |
|
"eval_runtime": 80.4348, |
|
"eval_samples_per_second": 158.414, |
|
"eval_steps_per_second": 1.243, |
|
"eval_wer": 19.26684383304674, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 340.25, |
|
"learning_rate": 5.682911392405063e-06, |
|
"loss": 0.0, |
|
"step": 34025 |
|
}, |
|
{ |
|
"epoch": 340.5, |
|
"learning_rate": 5.659177215189873e-06, |
|
"loss": 0.0, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 340.75, |
|
"learning_rate": 5.635443037974683e-06, |
|
"loss": 0.0, |
|
"step": 34075 |
|
}, |
|
{ |
|
"epoch": 341.0, |
|
"learning_rate": 5.611708860759493e-06, |
|
"loss": 0.0, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 341.25, |
|
"learning_rate": 5.587974683544303e-06, |
|
"loss": 0.0, |
|
"step": 34125 |
|
}, |
|
{ |
|
"epoch": 341.5, |
|
"learning_rate": 5.564240506329113e-06, |
|
"loss": 0.0, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 341.75, |
|
"learning_rate": 5.540506329113924e-06, |
|
"loss": 0.0, |
|
"step": 34175 |
|
}, |
|
{ |
|
"epoch": 342.0, |
|
"learning_rate": 5.516772151898734e-06, |
|
"loss": 0.0, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 342.25, |
|
"learning_rate": 5.493037974683544e-06, |
|
"loss": 0.0, |
|
"step": 34225 |
|
}, |
|
{ |
|
"epoch": 342.5, |
|
"learning_rate": 5.469303797468354e-06, |
|
"loss": 0.0, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 342.75, |
|
"learning_rate": 5.445569620253164e-06, |
|
"loss": 0.0, |
|
"step": 34275 |
|
}, |
|
{ |
|
"epoch": 343.0, |
|
"learning_rate": 5.4218354430379744e-06, |
|
"loss": 0.0, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 343.25, |
|
"learning_rate": 5.398101265822784e-06, |
|
"loss": 0.0, |
|
"step": 34325 |
|
}, |
|
{ |
|
"epoch": 343.5, |
|
"learning_rate": 5.374367088607594e-06, |
|
"loss": 0.0, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 343.75, |
|
"learning_rate": 5.350632911392404e-06, |
|
"loss": 0.0, |
|
"step": 34375 |
|
}, |
|
{ |
|
"epoch": 344.0, |
|
"learning_rate": 5.326898734177215e-06, |
|
"loss": 0.0, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 344.25, |
|
"learning_rate": 5.303164556962025e-06, |
|
"loss": 0.0, |
|
"step": 34425 |
|
}, |
|
{ |
|
"epoch": 344.5, |
|
"learning_rate": 5.279430379746835e-06, |
|
"loss": 0.0, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 344.75, |
|
"learning_rate": 5.255696202531645e-06, |
|
"loss": 0.0, |
|
"step": 34475 |
|
}, |
|
{ |
|
"epoch": 345.0, |
|
"learning_rate": 5.231962025316455e-06, |
|
"loss": 0.0, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 345.25, |
|
"learning_rate": 5.208227848101265e-06, |
|
"loss": 0.0, |
|
"step": 34525 |
|
}, |
|
{ |
|
"epoch": 345.5, |
|
"learning_rate": 5.1844936708860754e-06, |
|
"loss": 0.0, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 345.75, |
|
"learning_rate": 5.160759493670885e-06, |
|
"loss": 0.0, |
|
"step": 34575 |
|
}, |
|
{ |
|
"epoch": 346.0, |
|
"learning_rate": 5.137025316455696e-06, |
|
"loss": 0.0, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 346.25, |
|
"learning_rate": 5.1132911392405065e-06, |
|
"loss": 0.0, |
|
"step": 34625 |
|
}, |
|
{ |
|
"epoch": 346.5, |
|
"learning_rate": 5.089556962025316e-06, |
|
"loss": 0.0, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 346.75, |
|
"learning_rate": 5.065822784810126e-06, |
|
"loss": 0.0, |
|
"step": 34675 |
|
}, |
|
{ |
|
"epoch": 347.0, |
|
"learning_rate": 5.042088607594936e-06, |
|
"loss": 0.0, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 347.25, |
|
"learning_rate": 5.018354430379746e-06, |
|
"loss": 0.0, |
|
"step": 34725 |
|
}, |
|
{ |
|
"epoch": 347.5, |
|
"learning_rate": 4.9946202531645566e-06, |
|
"loss": 0.0, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 347.75, |
|
"learning_rate": 4.970886075949366e-06, |
|
"loss": 0.0, |
|
"step": 34775 |
|
}, |
|
{ |
|
"epoch": 348.0, |
|
"learning_rate": 4.947151898734177e-06, |
|
"loss": 0.0, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 348.25, |
|
"learning_rate": 4.923417721518988e-06, |
|
"loss": 0.0, |
|
"step": 34825 |
|
}, |
|
{ |
|
"epoch": 348.5, |
|
"learning_rate": 4.899683544303797e-06, |
|
"loss": 0.0, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 348.75, |
|
"learning_rate": 4.8759493670886075e-06, |
|
"loss": 0.0, |
|
"step": 34875 |
|
}, |
|
{ |
|
"epoch": 349.0, |
|
"learning_rate": 4.852215189873417e-06, |
|
"loss": 0.0, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 349.25, |
|
"learning_rate": 4.828481012658227e-06, |
|
"loss": 0.0, |
|
"step": 34925 |
|
}, |
|
{ |
|
"epoch": 349.5, |
|
"learning_rate": 4.804746835443037e-06, |
|
"loss": 0.0, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 349.75, |
|
"learning_rate": 4.781012658227847e-06, |
|
"loss": 0.0, |
|
"step": 34975 |
|
}, |
|
{ |
|
"epoch": 350.0, |
|
"learning_rate": 4.757278481012658e-06, |
|
"loss": 0.0, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 350.0, |
|
"eval_loss": 0.5910990238189697, |
|
"eval_runtime": 80.5464, |
|
"eval_samples_per_second": 158.195, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.234602157261214, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 350.25, |
|
"learning_rate": 4.733544303797468e-06, |
|
"loss": 0.0, |
|
"step": 35025 |
|
}, |
|
{ |
|
"epoch": 350.5, |
|
"learning_rate": 4.709810126582278e-06, |
|
"loss": 0.0, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 350.75, |
|
"learning_rate": 4.686075949367089e-06, |
|
"loss": 0.0, |
|
"step": 35075 |
|
}, |
|
{ |
|
"epoch": 351.0, |
|
"learning_rate": 4.662341772151898e-06, |
|
"loss": 0.0, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 351.25, |
|
"learning_rate": 4.6386075949367085e-06, |
|
"loss": 0.0, |
|
"step": 35125 |
|
}, |
|
{ |
|
"epoch": 351.5, |
|
"learning_rate": 4.614873417721519e-06, |
|
"loss": 0.0, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 351.75, |
|
"learning_rate": 4.591139240506329e-06, |
|
"loss": 0.0, |
|
"step": 35175 |
|
}, |
|
{ |
|
"epoch": 352.0, |
|
"learning_rate": 4.567405063291139e-06, |
|
"loss": 0.0, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 352.25, |
|
"learning_rate": 4.543670886075949e-06, |
|
"loss": 0.0, |
|
"step": 35225 |
|
}, |
|
{ |
|
"epoch": 352.5, |
|
"learning_rate": 4.5199367088607586e-06, |
|
"loss": 0.0, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 352.75, |
|
"learning_rate": 4.49620253164557e-06, |
|
"loss": 0.0, |
|
"step": 35275 |
|
}, |
|
{ |
|
"epoch": 353.0, |
|
"learning_rate": 4.472468354430379e-06, |
|
"loss": 0.0, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 353.25, |
|
"learning_rate": 4.44873417721519e-06, |
|
"loss": 0.0, |
|
"step": 35325 |
|
}, |
|
{ |
|
"epoch": 353.5, |
|
"learning_rate": 4.424999999999999e-06, |
|
"loss": 0.0, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 353.75, |
|
"learning_rate": 4.4012658227848095e-06, |
|
"loss": 0.0, |
|
"step": 35375 |
|
}, |
|
{ |
|
"epoch": 354.0, |
|
"learning_rate": 4.37753164556962e-06, |
|
"loss": 0.0, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 354.25, |
|
"learning_rate": 4.35379746835443e-06, |
|
"loss": 0.0, |
|
"step": 35425 |
|
}, |
|
{ |
|
"epoch": 354.5, |
|
"learning_rate": 4.33006329113924e-06, |
|
"loss": 0.0, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 354.75, |
|
"learning_rate": 4.30632911392405e-06, |
|
"loss": 0.0, |
|
"step": 35475 |
|
}, |
|
{ |
|
"epoch": 355.0, |
|
"learning_rate": 4.28259493670886e-06, |
|
"loss": 0.0, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 355.25, |
|
"learning_rate": 4.258860759493671e-06, |
|
"loss": 0.0, |
|
"step": 35525 |
|
}, |
|
{ |
|
"epoch": 355.5, |
|
"learning_rate": 4.23512658227848e-06, |
|
"loss": 0.0, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 355.75, |
|
"learning_rate": 4.211392405063291e-06, |
|
"loss": 0.0, |
|
"step": 35575 |
|
}, |
|
{ |
|
"epoch": 356.0, |
|
"learning_rate": 4.187658227848101e-06, |
|
"loss": 0.0, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 356.25, |
|
"learning_rate": 4.163924050632911e-06, |
|
"loss": 0.0, |
|
"step": 35625 |
|
}, |
|
{ |
|
"epoch": 356.5, |
|
"learning_rate": 4.140189873417721e-06, |
|
"loss": 0.0, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 356.75, |
|
"learning_rate": 4.116455696202531e-06, |
|
"loss": 0.0, |
|
"step": 35675 |
|
}, |
|
{ |
|
"epoch": 357.0, |
|
"learning_rate": 4.0927215189873416e-06, |
|
"loss": 0.0, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 357.25, |
|
"learning_rate": 4.068987341772152e-06, |
|
"loss": 0.0, |
|
"step": 35725 |
|
}, |
|
{ |
|
"epoch": 357.5, |
|
"learning_rate": 4.045253164556961e-06, |
|
"loss": 0.0, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 357.75, |
|
"learning_rate": 4.021518987341772e-06, |
|
"loss": 0.0, |
|
"step": 35775 |
|
}, |
|
{ |
|
"epoch": 358.0, |
|
"learning_rate": 3.997784810126582e-06, |
|
"loss": 0.0, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 358.25, |
|
"learning_rate": 3.974050632911392e-06, |
|
"loss": 0.0, |
|
"step": 35825 |
|
}, |
|
{ |
|
"epoch": 358.5, |
|
"learning_rate": 3.950316455696202e-06, |
|
"loss": 0.0, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 358.75, |
|
"learning_rate": 3.926582278481012e-06, |
|
"loss": 0.0, |
|
"step": 35875 |
|
}, |
|
{ |
|
"epoch": 359.0, |
|
"learning_rate": 3.902848101265823e-06, |
|
"loss": 0.0, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 359.25, |
|
"learning_rate": 3.879113924050632e-06, |
|
"loss": 0.0, |
|
"step": 35925 |
|
}, |
|
{ |
|
"epoch": 359.5, |
|
"learning_rate": 3.8553797468354425e-06, |
|
"loss": 0.0, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 359.75, |
|
"learning_rate": 3.831645569620253e-06, |
|
"loss": 0.0, |
|
"step": 35975 |
|
}, |
|
{ |
|
"epoch": 360.0, |
|
"learning_rate": 3.8088607594936704e-06, |
|
"loss": 0.0, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 360.0, |
|
"eval_loss": 0.5961853861808777, |
|
"eval_runtime": 80.7236, |
|
"eval_samples_per_second": 157.847, |
|
"eval_steps_per_second": 1.239, |
|
"eval_wer": 19.22874003439112, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 360.25, |
|
"learning_rate": 3.7851265822784808e-06, |
|
"loss": 0.0, |
|
"step": 36025 |
|
}, |
|
{ |
|
"epoch": 360.5, |
|
"learning_rate": 3.7613924050632907e-06, |
|
"loss": 0.0, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 360.75, |
|
"learning_rate": 3.737658227848101e-06, |
|
"loss": 0.0, |
|
"step": 36075 |
|
}, |
|
{ |
|
"epoch": 361.0, |
|
"learning_rate": 3.713924050632911e-06, |
|
"loss": 0.0, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 361.25, |
|
"learning_rate": 3.690189873417721e-06, |
|
"loss": 0.0, |
|
"step": 36125 |
|
}, |
|
{ |
|
"epoch": 361.5, |
|
"learning_rate": 3.6664556962025313e-06, |
|
"loss": 0.0, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 361.75, |
|
"learning_rate": 3.6427215189873416e-06, |
|
"loss": 0.0, |
|
"step": 36175 |
|
}, |
|
{ |
|
"epoch": 362.0, |
|
"learning_rate": 3.6189873417721516e-06, |
|
"loss": 0.0, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 362.25, |
|
"learning_rate": 3.5952531645569615e-06, |
|
"loss": 0.0, |
|
"step": 36225 |
|
}, |
|
{ |
|
"epoch": 362.5, |
|
"learning_rate": 3.571518987341772e-06, |
|
"loss": 0.0, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 362.75, |
|
"learning_rate": 3.547784810126582e-06, |
|
"loss": 0.0, |
|
"step": 36275 |
|
}, |
|
{ |
|
"epoch": 363.0, |
|
"learning_rate": 3.524050632911392e-06, |
|
"loss": 0.0, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 363.25, |
|
"learning_rate": 3.500316455696202e-06, |
|
"loss": 0.0, |
|
"step": 36325 |
|
}, |
|
{ |
|
"epoch": 363.5, |
|
"learning_rate": 3.476582278481012e-06, |
|
"loss": 0.0, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 363.75, |
|
"learning_rate": 3.4528481012658228e-06, |
|
"loss": 0.0, |
|
"step": 36375 |
|
}, |
|
{ |
|
"epoch": 364.0, |
|
"learning_rate": 3.4291139240506327e-06, |
|
"loss": 0.0, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 364.25, |
|
"learning_rate": 3.4053797468354426e-06, |
|
"loss": 0.0, |
|
"step": 36425 |
|
}, |
|
{ |
|
"epoch": 364.5, |
|
"learning_rate": 3.3816455696202526e-06, |
|
"loss": 0.0, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 364.75, |
|
"learning_rate": 3.3579113924050633e-06, |
|
"loss": 0.0, |
|
"step": 36475 |
|
}, |
|
{ |
|
"epoch": 365.0, |
|
"learning_rate": 3.3341772151898733e-06, |
|
"loss": 0.0, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 365.25, |
|
"learning_rate": 3.310443037974683e-06, |
|
"loss": 0.0, |
|
"step": 36525 |
|
}, |
|
{ |
|
"epoch": 365.5, |
|
"learning_rate": 3.286708860759493e-06, |
|
"loss": 0.0, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 365.75, |
|
"learning_rate": 3.262974683544304e-06, |
|
"loss": 0.0, |
|
"step": 36575 |
|
}, |
|
{ |
|
"epoch": 366.0, |
|
"learning_rate": 3.239240506329114e-06, |
|
"loss": 0.0, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 366.25, |
|
"learning_rate": 3.2155063291139238e-06, |
|
"loss": 0.0, |
|
"step": 36625 |
|
}, |
|
{ |
|
"epoch": 366.5, |
|
"learning_rate": 3.1917721518987337e-06, |
|
"loss": 0.0, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 366.75, |
|
"learning_rate": 3.1680379746835436e-06, |
|
"loss": 0.0, |
|
"step": 36675 |
|
}, |
|
{ |
|
"epoch": 367.0, |
|
"learning_rate": 3.1443037974683544e-06, |
|
"loss": 0.0, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 367.25, |
|
"learning_rate": 3.1205696202531643e-06, |
|
"loss": 0.0, |
|
"step": 36725 |
|
}, |
|
{ |
|
"epoch": 367.5, |
|
"learning_rate": 3.0968354430379743e-06, |
|
"loss": 0.0, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 367.75, |
|
"learning_rate": 3.073101265822784e-06, |
|
"loss": 0.0, |
|
"step": 36775 |
|
}, |
|
{ |
|
"epoch": 368.0, |
|
"learning_rate": 3.049367088607595e-06, |
|
"loss": 0.0, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 368.25, |
|
"learning_rate": 3.025632911392405e-06, |
|
"loss": 0.0, |
|
"step": 36825 |
|
}, |
|
{ |
|
"epoch": 368.5, |
|
"learning_rate": 3.001898734177215e-06, |
|
"loss": 0.0, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 368.75, |
|
"learning_rate": 2.9781645569620248e-06, |
|
"loss": 0.0, |
|
"step": 36875 |
|
}, |
|
{ |
|
"epoch": 369.0, |
|
"learning_rate": 2.9544303797468355e-06, |
|
"loss": 0.0, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 369.25, |
|
"learning_rate": 2.9306962025316455e-06, |
|
"loss": 0.0, |
|
"step": 36925 |
|
}, |
|
{ |
|
"epoch": 369.5, |
|
"learning_rate": 2.9069620253164554e-06, |
|
"loss": 0.0, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 369.75, |
|
"learning_rate": 2.8832278481012653e-06, |
|
"loss": 0.0, |
|
"step": 36975 |
|
}, |
|
{ |
|
"epoch": 370.0, |
|
"learning_rate": 2.8594936708860757e-06, |
|
"loss": 0.0, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 370.0, |
|
"eval_loss": 0.6009894013404846, |
|
"eval_runtime": 80.466, |
|
"eval_samples_per_second": 158.353, |
|
"eval_steps_per_second": 1.243, |
|
"eval_wer": 19.232648116304517, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 370.25, |
|
"learning_rate": 2.835759493670886e-06, |
|
"loss": 0.0, |
|
"step": 37025 |
|
}, |
|
{ |
|
"epoch": 370.5, |
|
"learning_rate": 2.812025316455696e-06, |
|
"loss": 0.0, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 370.75, |
|
"learning_rate": 2.788291139240506e-06, |
|
"loss": 0.0, |
|
"step": 37075 |
|
}, |
|
{ |
|
"epoch": 371.0, |
|
"learning_rate": 2.7645569620253162e-06, |
|
"loss": 0.0, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 371.25, |
|
"learning_rate": 2.7408227848101266e-06, |
|
"loss": 0.0, |
|
"step": 37125 |
|
}, |
|
{ |
|
"epoch": 371.5, |
|
"learning_rate": 2.7170886075949365e-06, |
|
"loss": 0.0, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 371.75, |
|
"learning_rate": 2.6933544303797465e-06, |
|
"loss": 0.0, |
|
"step": 37175 |
|
}, |
|
{ |
|
"epoch": 372.0, |
|
"learning_rate": 2.669620253164557e-06, |
|
"loss": 0.0, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 372.25, |
|
"learning_rate": 2.6458860759493667e-06, |
|
"loss": 0.0, |
|
"step": 37225 |
|
}, |
|
{ |
|
"epoch": 372.5, |
|
"learning_rate": 2.622151898734177e-06, |
|
"loss": 0.0, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 372.75, |
|
"learning_rate": 2.598417721518987e-06, |
|
"loss": 0.0, |
|
"step": 37275 |
|
}, |
|
{ |
|
"epoch": 373.0, |
|
"learning_rate": 2.5746835443037974e-06, |
|
"loss": 0.0, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 373.25, |
|
"learning_rate": 2.5509493670886073e-06, |
|
"loss": 0.0, |
|
"step": 37325 |
|
}, |
|
{ |
|
"epoch": 373.5, |
|
"learning_rate": 2.5272151898734177e-06, |
|
"loss": 0.0, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 373.75, |
|
"learning_rate": 2.5034810126582276e-06, |
|
"loss": 0.0, |
|
"step": 37375 |
|
}, |
|
{ |
|
"epoch": 374.0, |
|
"learning_rate": 2.479746835443038e-06, |
|
"loss": 0.0, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 374.25, |
|
"learning_rate": 2.456012658227848e-06, |
|
"loss": 0.0, |
|
"step": 37425 |
|
}, |
|
{ |
|
"epoch": 374.5, |
|
"learning_rate": 2.432278481012658e-06, |
|
"loss": 0.0, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 374.75, |
|
"learning_rate": 2.408544303797468e-06, |
|
"loss": 0.0, |
|
"step": 37475 |
|
}, |
|
{ |
|
"epoch": 375.0, |
|
"learning_rate": 2.384810126582278e-06, |
|
"loss": 0.0, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 375.25, |
|
"learning_rate": 2.3610759493670884e-06, |
|
"loss": 0.0, |
|
"step": 37525 |
|
}, |
|
{ |
|
"epoch": 375.5, |
|
"learning_rate": 2.3373417721518984e-06, |
|
"loss": 0.0, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 375.75, |
|
"learning_rate": 2.3136075949367087e-06, |
|
"loss": 0.0, |
|
"step": 37575 |
|
}, |
|
{ |
|
"epoch": 376.0, |
|
"learning_rate": 2.2898734177215187e-06, |
|
"loss": 0.0, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 376.25, |
|
"learning_rate": 2.2661392405063286e-06, |
|
"loss": 0.0, |
|
"step": 37625 |
|
}, |
|
{ |
|
"epoch": 376.5, |
|
"learning_rate": 2.242405063291139e-06, |
|
"loss": 0.0, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 376.75, |
|
"learning_rate": 2.218670886075949e-06, |
|
"loss": 0.0, |
|
"step": 37675 |
|
}, |
|
{ |
|
"epoch": 377.0, |
|
"learning_rate": 2.1949367088607592e-06, |
|
"loss": 0.0, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 377.25, |
|
"learning_rate": 2.171202531645569e-06, |
|
"loss": 0.0, |
|
"step": 37725 |
|
}, |
|
{ |
|
"epoch": 377.5, |
|
"learning_rate": 2.1474683544303795e-06, |
|
"loss": 0.0, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 377.75, |
|
"learning_rate": 2.1237341772151894e-06, |
|
"loss": 0.0, |
|
"step": 37775 |
|
}, |
|
{ |
|
"epoch": 378.0, |
|
"learning_rate": 2.1e-06, |
|
"loss": 0.0, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 378.25, |
|
"learning_rate": 2.0762658227848097e-06, |
|
"loss": 0.0, |
|
"step": 37825 |
|
}, |
|
{ |
|
"epoch": 378.5, |
|
"learning_rate": 2.05253164556962e-06, |
|
"loss": 0.0, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 378.75, |
|
"learning_rate": 2.02879746835443e-06, |
|
"loss": 0.0, |
|
"step": 37875 |
|
}, |
|
{ |
|
"epoch": 379.0, |
|
"learning_rate": 2.0050632911392404e-06, |
|
"loss": 0.0, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 379.25, |
|
"learning_rate": 1.9813291139240503e-06, |
|
"loss": 0.0, |
|
"step": 37925 |
|
}, |
|
{ |
|
"epoch": 379.5, |
|
"learning_rate": 1.9575949367088607e-06, |
|
"loss": 0.0, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 379.75, |
|
"learning_rate": 1.9338607594936706e-06, |
|
"loss": 0.0, |
|
"step": 37975 |
|
}, |
|
{ |
|
"epoch": 380.0, |
|
"learning_rate": 1.910126582278481e-06, |
|
"loss": 0.0, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 380.0, |
|
"eval_loss": 0.6050010919570923, |
|
"eval_runtime": 80.5317, |
|
"eval_samples_per_second": 158.223, |
|
"eval_steps_per_second": 1.242, |
|
"eval_wer": 19.22874003439112, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 380.25, |
|
"learning_rate": 1.8863924050632909e-06, |
|
"loss": 0.0, |
|
"step": 38025 |
|
}, |
|
{ |
|
"epoch": 380.5, |
|
"learning_rate": 1.8626582278481012e-06, |
|
"loss": 0.0, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 380.75, |
|
"learning_rate": 1.8389240506329112e-06, |
|
"loss": 0.0, |
|
"step": 38075 |
|
}, |
|
{ |
|
"epoch": 381.0, |
|
"learning_rate": 1.8151898734177215e-06, |
|
"loss": 0.0, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 381.25, |
|
"learning_rate": 1.7914556962025314e-06, |
|
"loss": 0.0, |
|
"step": 38125 |
|
}, |
|
{ |
|
"epoch": 381.5, |
|
"learning_rate": 1.7677215189873418e-06, |
|
"loss": 0.0, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 381.75, |
|
"learning_rate": 1.7439873417721517e-06, |
|
"loss": 0.0, |
|
"step": 38175 |
|
}, |
|
{ |
|
"epoch": 382.0, |
|
"learning_rate": 1.720253164556962e-06, |
|
"loss": 0.0, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 382.25, |
|
"learning_rate": 1.696518987341772e-06, |
|
"loss": 0.0, |
|
"step": 38225 |
|
}, |
|
{ |
|
"epoch": 382.5, |
|
"learning_rate": 1.6727848101265821e-06, |
|
"loss": 0.0, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 382.75, |
|
"learning_rate": 1.6490506329113923e-06, |
|
"loss": 0.0, |
|
"step": 38275 |
|
}, |
|
{ |
|
"epoch": 383.0, |
|
"learning_rate": 1.6253164556962024e-06, |
|
"loss": 0.0, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 383.25, |
|
"learning_rate": 1.6015822784810126e-06, |
|
"loss": 0.0, |
|
"step": 38325 |
|
}, |
|
{ |
|
"epoch": 383.5, |
|
"learning_rate": 1.5778481012658227e-06, |
|
"loss": 0.0, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 383.75, |
|
"learning_rate": 1.5541139240506329e-06, |
|
"loss": 0.0, |
|
"step": 38375 |
|
}, |
|
{ |
|
"epoch": 384.0, |
|
"learning_rate": 1.5303797468354428e-06, |
|
"loss": 0.0, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 384.25, |
|
"learning_rate": 1.5066455696202531e-06, |
|
"loss": 0.0, |
|
"step": 38425 |
|
}, |
|
{ |
|
"epoch": 384.5, |
|
"learning_rate": 1.482911392405063e-06, |
|
"loss": 0.0, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 384.75, |
|
"learning_rate": 1.4591772151898732e-06, |
|
"loss": 0.0, |
|
"step": 38475 |
|
}, |
|
{ |
|
"epoch": 385.0, |
|
"learning_rate": 1.4354430379746834e-06, |
|
"loss": 0.0, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 385.25, |
|
"learning_rate": 1.4117088607594935e-06, |
|
"loss": 0.0, |
|
"step": 38525 |
|
}, |
|
{ |
|
"epoch": 385.5, |
|
"learning_rate": 1.3879746835443036e-06, |
|
"loss": 0.0, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 385.75, |
|
"learning_rate": 1.3642405063291138e-06, |
|
"loss": 0.0, |
|
"step": 38575 |
|
}, |
|
{ |
|
"epoch": 386.0, |
|
"learning_rate": 1.340506329113924e-06, |
|
"loss": 0.0, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 386.25, |
|
"learning_rate": 1.316772151898734e-06, |
|
"loss": 0.0, |
|
"step": 38625 |
|
}, |
|
{ |
|
"epoch": 386.5, |
|
"learning_rate": 1.293037974683544e-06, |
|
"loss": 0.0, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 386.75, |
|
"learning_rate": 1.2693037974683543e-06, |
|
"loss": 0.0, |
|
"step": 38675 |
|
}, |
|
{ |
|
"epoch": 387.0, |
|
"learning_rate": 1.2455696202531643e-06, |
|
"loss": 0.0, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 387.25, |
|
"learning_rate": 1.2218354430379746e-06, |
|
"loss": 0.0, |
|
"step": 38725 |
|
}, |
|
{ |
|
"epoch": 387.5, |
|
"learning_rate": 1.1981012658227846e-06, |
|
"loss": 0.0, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 387.75, |
|
"learning_rate": 1.174367088607595e-06, |
|
"loss": 0.0, |
|
"step": 38775 |
|
}, |
|
{ |
|
"epoch": 388.0, |
|
"learning_rate": 1.150632911392405e-06, |
|
"loss": 0.0, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 388.25, |
|
"learning_rate": 1.1268987341772152e-06, |
|
"loss": 0.0, |
|
"step": 38825 |
|
}, |
|
{ |
|
"epoch": 388.5, |
|
"learning_rate": 1.1031645569620251e-06, |
|
"loss": 0.0, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 388.75, |
|
"learning_rate": 1.0794303797468353e-06, |
|
"loss": 0.0, |
|
"step": 38875 |
|
}, |
|
{ |
|
"epoch": 389.0, |
|
"learning_rate": 1.0556962025316454e-06, |
|
"loss": 0.0, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 389.25, |
|
"learning_rate": 1.0319620253164556e-06, |
|
"loss": 0.0, |
|
"step": 38925 |
|
}, |
|
{ |
|
"epoch": 389.5, |
|
"learning_rate": 1.0082278481012657e-06, |
|
"loss": 0.0, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 389.75, |
|
"learning_rate": 9.844936708860758e-07, |
|
"loss": 0.0, |
|
"step": 38975 |
|
}, |
|
{ |
|
"epoch": 390.0, |
|
"learning_rate": 9.60759493670886e-07, |
|
"loss": 0.0, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 390.0, |
|
"eval_loss": 0.608101487159729, |
|
"eval_runtime": 80.4312, |
|
"eval_samples_per_second": 158.421, |
|
"eval_steps_per_second": 1.243, |
|
"eval_wer": 19.237533218696264, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 390.25, |
|
"learning_rate": 9.370253164556961e-07, |
|
"loss": 0.0, |
|
"step": 39025 |
|
}, |
|
{ |
|
"epoch": 390.5, |
|
"learning_rate": 9.132911392405062e-07, |
|
"loss": 0.0, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 390.75, |
|
"learning_rate": 8.895569620253163e-07, |
|
"loss": 0.0, |
|
"step": 39075 |
|
}, |
|
{ |
|
"epoch": 391.0, |
|
"learning_rate": 8.658227848101264e-07, |
|
"loss": 0.0, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 391.25, |
|
"learning_rate": 8.420886075949366e-07, |
|
"loss": 0.0, |
|
"step": 39125 |
|
}, |
|
{ |
|
"epoch": 391.5, |
|
"learning_rate": 8.183544303797467e-07, |
|
"loss": 0.0, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 391.75, |
|
"learning_rate": 7.946202531645569e-07, |
|
"loss": 0.0, |
|
"step": 39175 |
|
}, |
|
{ |
|
"epoch": 392.0, |
|
"learning_rate": 7.70886075949367e-07, |
|
"loss": 0.0, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 392.25, |
|
"learning_rate": 7.471518987341772e-07, |
|
"loss": 0.0, |
|
"step": 39225 |
|
}, |
|
{ |
|
"epoch": 392.5, |
|
"learning_rate": 7.234177215189873e-07, |
|
"loss": 0.0, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 392.75, |
|
"learning_rate": 6.996835443037974e-07, |
|
"loss": 0.0, |
|
"step": 39275 |
|
}, |
|
{ |
|
"epoch": 393.0, |
|
"learning_rate": 6.759493670886076e-07, |
|
"loss": 0.0, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 393.25, |
|
"learning_rate": 6.522151898734176e-07, |
|
"loss": 0.0, |
|
"step": 39325 |
|
}, |
|
{ |
|
"epoch": 393.5, |
|
"learning_rate": 6.284810126582278e-07, |
|
"loss": 0.0, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 393.75, |
|
"learning_rate": 6.047468354430379e-07, |
|
"loss": 0.0, |
|
"step": 39375 |
|
}, |
|
{ |
|
"epoch": 394.0, |
|
"learning_rate": 5.81012658227848e-07, |
|
"loss": 0.0, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 394.25, |
|
"learning_rate": 5.572784810126582e-07, |
|
"loss": 0.0, |
|
"step": 39425 |
|
}, |
|
{ |
|
"epoch": 394.5, |
|
"learning_rate": 5.335443037974683e-07, |
|
"loss": 0.0, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 394.75, |
|
"learning_rate": 5.098101265822785e-07, |
|
"loss": 0.0, |
|
"step": 39475 |
|
}, |
|
{ |
|
"epoch": 395.0, |
|
"learning_rate": 4.860759493670886e-07, |
|
"loss": 0.0, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 395.25, |
|
"learning_rate": 4.6234177215189865e-07, |
|
"loss": 0.0, |
|
"step": 39525 |
|
}, |
|
{ |
|
"epoch": 395.5, |
|
"learning_rate": 4.386075949367088e-07, |
|
"loss": 0.0, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 395.75, |
|
"learning_rate": 4.1487341772151893e-07, |
|
"loss": 0.0, |
|
"step": 39575 |
|
}, |
|
{ |
|
"epoch": 396.0, |
|
"learning_rate": 3.911392405063291e-07, |
|
"loss": 0.0, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 396.25, |
|
"learning_rate": 3.674050632911392e-07, |
|
"loss": 0.0, |
|
"step": 39625 |
|
}, |
|
{ |
|
"epoch": 396.5, |
|
"learning_rate": 3.436708860759493e-07, |
|
"loss": 0.0, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 396.75, |
|
"learning_rate": 3.1993670886075945e-07, |
|
"loss": 0.0, |
|
"step": 39675 |
|
}, |
|
{ |
|
"epoch": 397.0, |
|
"learning_rate": 2.962025316455696e-07, |
|
"loss": 0.0, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 397.25, |
|
"learning_rate": 2.7246835443037973e-07, |
|
"loss": 0.0, |
|
"step": 39725 |
|
}, |
|
{ |
|
"epoch": 397.5, |
|
"learning_rate": 2.487341772151899e-07, |
|
"loss": 0.0, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 397.75, |
|
"learning_rate": 2.25e-07, |
|
"loss": 0.0, |
|
"step": 39775 |
|
}, |
|
{ |
|
"epoch": 398.0, |
|
"learning_rate": 2.012658227848101e-07, |
|
"loss": 0.0, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 398.25, |
|
"learning_rate": 1.7753164556962025e-07, |
|
"loss": 0.0, |
|
"step": 39825 |
|
}, |
|
{ |
|
"epoch": 398.5, |
|
"learning_rate": 1.537974683544304e-07, |
|
"loss": 0.0, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 398.75, |
|
"learning_rate": 1.300632911392405e-07, |
|
"loss": 0.0, |
|
"step": 39875 |
|
}, |
|
{ |
|
"epoch": 399.0, |
|
"learning_rate": 1.0632911392405063e-07, |
|
"loss": 0.0, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 399.25, |
|
"learning_rate": 8.259493670886075e-08, |
|
"loss": 0.0, |
|
"step": 39925 |
|
}, |
|
{ |
|
"epoch": 399.5, |
|
"learning_rate": 5.886075949367088e-08, |
|
"loss": 0.0, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 399.75, |
|
"learning_rate": 3.5126582278481014e-08, |
|
"loss": 0.0, |
|
"step": 39975 |
|
}, |
|
{ |
|
"epoch": 400.0, |
|
"learning_rate": 1.2341772151898733e-08, |
|
"loss": 0.0, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 400.0, |
|
"eval_loss": 0.6094558238983154, |
|
"eval_runtime": 80.3325, |
|
"eval_samples_per_second": 158.616, |
|
"eval_steps_per_second": 1.245, |
|
"eval_wer": 19.196498358605595, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 400.0, |
|
"step": 40000, |
|
"total_flos": 2.50955067359232e+20, |
|
"train_loss": 0.010605198578223871, |
|
"train_runtime": 40076.3871, |
|
"train_samples_per_second": 255.512, |
|
"train_steps_per_second": 0.998 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 40000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 400, |
|
"save_steps": 1000, |
|
"total_flos": 2.50955067359232e+20, |
|
"train_batch_size": 256, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|