|
{ |
|
"best_metric": 9.788891681326966, |
|
"best_model_checkpoint": "./checkpoint-12000", |
|
"epoch": 0.6, |
|
"global_step": 12000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5.0453611334320685e-06, |
|
"loss": 1.2966, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.229195710491767e-06, |
|
"loss": 0.8695, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.903829450223392e-06, |
|
"loss": 0.7429, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.377725845391017e-06, |
|
"loss": 0.6562, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.743343231239583e-06, |
|
"loss": 0.6217, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.041073861170494e-06, |
|
"loss": 0.5393, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.292222957399574e-06, |
|
"loss": 0.5282, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.509413541357755e-06, |
|
"loss": 0.5062, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.700744577655557e-06, |
|
"loss": 0.4942, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.871723942761204e-06, |
|
"loss": 0.4508, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.026267958246849e-06, |
|
"loss": 0.4466, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.16726106663399e-06, |
|
"loss": 0.4026, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.296889251455016e-06, |
|
"loss": 0.37, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.416848797368692e-06, |
|
"loss": 0.3864, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.528482449516371e-06, |
|
"loss": 0.3801, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.632871309784314e-06, |
|
"loss": 0.3893, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.73089868785391e-06, |
|
"loss": 0.4079, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.823295589572114e-06, |
|
"loss": 0.3554, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.910673836465484e-06, |
|
"loss": 0.3197, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.993550644973805e-06, |
|
"loss": 0.3405, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.98871794871795e-06, |
|
"loss": 0.3328, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.975897435897436e-06, |
|
"loss": 0.3147, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.963076923076925e-06, |
|
"loss": 0.2954, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.950256410256412e-06, |
|
"loss": 0.3034, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.937435897435898e-06, |
|
"loss": 0.2931, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.924615384615385e-06, |
|
"loss": 0.2798, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.911794871794874e-06, |
|
"loss": 0.2619, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.89897435897436e-06, |
|
"loss": 0.3157, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.886153846153846e-06, |
|
"loss": 0.2937, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.873333333333334e-06, |
|
"loss": 0.2606, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.860512820512821e-06, |
|
"loss": 0.2482, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.847692307692308e-06, |
|
"loss": 0.2539, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.834871794871795e-06, |
|
"loss": 0.2501, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.822051282051283e-06, |
|
"loss": 0.241, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.80923076923077e-06, |
|
"loss": 0.2323, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.796410256410257e-06, |
|
"loss": 0.2507, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.783589743589744e-06, |
|
"loss": 0.2157, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.770769230769232e-06, |
|
"loss": 0.2356, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.757948717948719e-06, |
|
"loss": 0.2352, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.745128205128206e-06, |
|
"loss": 0.2181, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 0.2607421875, |
|
"eval_runtime": 228.2189, |
|
"eval_samples_per_second": 4.382, |
|
"eval_steps_per_second": 0.14, |
|
"eval_wer": 18.384016084443328, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.733333333333334e-06, |
|
"loss": 0.2279, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.720512820512822e-06, |
|
"loss": 0.225, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.707692307692308e-06, |
|
"loss": 0.2238, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.694871794871794e-06, |
|
"loss": 0.228, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.682051282051283e-06, |
|
"loss": 0.2261, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.66923076923077e-06, |
|
"loss": 0.196, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.656410256410257e-06, |
|
"loss": 0.1962, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.643589743589743e-06, |
|
"loss": 0.1931, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.630769230769232e-06, |
|
"loss": 0.1827, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.617948717948719e-06, |
|
"loss": 0.1815, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.605128205128206e-06, |
|
"loss": 0.1972, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.592307692307692e-06, |
|
"loss": 0.1681, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.579487179487181e-06, |
|
"loss": 0.151, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.566666666666668e-06, |
|
"loss": 0.1743, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.553846153846155e-06, |
|
"loss": 0.1808, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.541025641025641e-06, |
|
"loss": 0.1984, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.52820512820513e-06, |
|
"loss": 0.2098, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.515384615384617e-06, |
|
"loss": 0.1741, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.502564102564103e-06, |
|
"loss": 0.1489, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.48974358974359e-06, |
|
"loss": 0.1655, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.476923076923079e-06, |
|
"loss": 0.1643, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.464102564102566e-06, |
|
"loss": 0.151, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.451282051282052e-06, |
|
"loss": 0.1431, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.43846153846154e-06, |
|
"loss": 0.1503, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.425641025641026e-06, |
|
"loss": 0.1491, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.412820512820515e-06, |
|
"loss": 0.1408, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.4e-06, |
|
"loss": 0.1294, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.387179487179488e-06, |
|
"loss": 0.1618, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.374358974358975e-06, |
|
"loss": 0.1484, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.361538461538462e-06, |
|
"loss": 0.1298, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.348717948717949e-06, |
|
"loss": 0.1231, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.335897435897436e-06, |
|
"loss": 0.1241, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.323076923076924e-06, |
|
"loss": 0.1215, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.310256410256411e-06, |
|
"loss": 0.1225, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.297435897435898e-06, |
|
"loss": 0.1122, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.284615384615385e-06, |
|
"loss": 0.1191, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.271794871794873e-06, |
|
"loss": 0.0995, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.25897435897436e-06, |
|
"loss": 0.1199, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.246153846153847e-06, |
|
"loss": 0.1243, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.233333333333334e-06, |
|
"loss": 0.1089, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 0.2489013671875, |
|
"eval_runtime": 209.1058, |
|
"eval_samples_per_second": 4.782, |
|
"eval_steps_per_second": 0.153, |
|
"eval_wer": 16.486554410655945, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.221538461538462e-06, |
|
"loss": 0.1231, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.208717948717948e-06, |
|
"loss": 0.1173, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.195897435897437e-06, |
|
"loss": 0.1234, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.183076923076924e-06, |
|
"loss": 0.1244, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.17025641025641e-06, |
|
"loss": 0.1223, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.157435897435897e-06, |
|
"loss": 0.1027, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.144615384615386e-06, |
|
"loss": 0.1022, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.131794871794873e-06, |
|
"loss": 0.1013, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.11897435897436e-06, |
|
"loss": 0.0949, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.106153846153846e-06, |
|
"loss": 0.0971, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.093333333333333e-06, |
|
"loss": 0.1098, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.080512820512822e-06, |
|
"loss": 0.0861, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.067692307692309e-06, |
|
"loss": 0.0765, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.054871794871795e-06, |
|
"loss": 0.0933, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.042051282051282e-06, |
|
"loss": 0.1025, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.02923076923077e-06, |
|
"loss": 0.1138, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.016410256410258e-06, |
|
"loss": 0.1198, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.003589743589744e-06, |
|
"loss": 0.0981, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.990769230769231e-06, |
|
"loss": 0.0807, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.97794871794872e-06, |
|
"loss": 0.0875, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.965128205128207e-06, |
|
"loss": 0.0911, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.952307692307693e-06, |
|
"loss": 0.0796, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.93948717948718e-06, |
|
"loss": 0.0766, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.926666666666669e-06, |
|
"loss": 0.0862, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.913846153846154e-06, |
|
"loss": 0.0822, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.90102564102564e-06, |
|
"loss": 0.0757, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.88820512820513e-06, |
|
"loss": 0.0674, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.875384615384616e-06, |
|
"loss": 0.0913, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.862564102564103e-06, |
|
"loss": 0.0815, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.84974358974359e-06, |
|
"loss": 0.0687, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.836923076923078e-06, |
|
"loss": 0.0649, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.824102564102565e-06, |
|
"loss": 0.0647, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.811282051282052e-06, |
|
"loss": 0.0607, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.798461538461539e-06, |
|
"loss": 0.065, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.785641025641025e-06, |
|
"loss": 0.058, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.772820512820514e-06, |
|
"loss": 0.0618, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.76e-06, |
|
"loss": 0.0491, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.747179487179488e-06, |
|
"loss": 0.0646, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.734358974358974e-06, |
|
"loss": 0.0673, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.721538461538463e-06, |
|
"loss": 0.0564, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 0.261474609375, |
|
"eval_runtime": 212.2501, |
|
"eval_samples_per_second": 4.711, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 16.14727318421714, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.70974358974359e-06, |
|
"loss": 0.1893, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.696923076923078e-06, |
|
"loss": 0.1774, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.684102564102565e-06, |
|
"loss": 0.1866, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.671282051282051e-06, |
|
"loss": 0.1896, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.658461538461538e-06, |
|
"loss": 0.1925, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.645641025641027e-06, |
|
"loss": 0.2126, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.632820512820514e-06, |
|
"loss": 0.2004, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.62e-06, |
|
"loss": 0.1636, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.607179487179487e-06, |
|
"loss": 0.1597, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.594358974358976e-06, |
|
"loss": 0.1768, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.581538461538463e-06, |
|
"loss": 0.1684, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.56871794871795e-06, |
|
"loss": 0.1739, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.555897435897436e-06, |
|
"loss": 0.1706, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.543076923076923e-06, |
|
"loss": 0.1701, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.530256410256412e-06, |
|
"loss": 0.1564, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.517435897435898e-06, |
|
"loss": 0.1627, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.504615384615385e-06, |
|
"loss": 0.1586, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.491794871794872e-06, |
|
"loss": 0.1579, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.47897435897436e-06, |
|
"loss": 0.1676, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.466153846153847e-06, |
|
"loss": 0.1558, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.453333333333334e-06, |
|
"loss": 0.1624, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.440512820512821e-06, |
|
"loss": 0.1459, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.427692307692308e-06, |
|
"loss": 0.151, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.414871794871795e-06, |
|
"loss": 0.1524, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.402051282051282e-06, |
|
"loss": 0.1444, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.38923076923077e-06, |
|
"loss": 0.1218, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.376410256410257e-06, |
|
"loss": 0.1207, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.363589743589744e-06, |
|
"loss": 0.1199, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.35076923076923e-06, |
|
"loss": 0.1154, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.337948717948719e-06, |
|
"loss": 0.1075, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.325128205128206e-06, |
|
"loss": 0.1005, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.312307692307693e-06, |
|
"loss": 0.1053, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.29948717948718e-06, |
|
"loss": 0.1087, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.286666666666668e-06, |
|
"loss": 0.1207, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.273846153846155e-06, |
|
"loss": 0.1099, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.261025641025642e-06, |
|
"loss": 0.1054, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.248205128205129e-06, |
|
"loss": 0.1019, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.235384615384615e-06, |
|
"loss": 0.0974, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.222564102564104e-06, |
|
"loss": 0.0975, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.20974358974359e-06, |
|
"loss": 0.1031, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 0.1995849609375, |
|
"eval_runtime": 214.2177, |
|
"eval_samples_per_second": 4.668, |
|
"eval_steps_per_second": 0.149, |
|
"eval_wer": 14.224679567730586, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.197948717948719e-06, |
|
"loss": 0.1122, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.185128205128206e-06, |
|
"loss": 0.1038, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.172307692307692e-06, |
|
"loss": 0.1172, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.15948717948718e-06, |
|
"loss": 0.1251, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.146666666666668e-06, |
|
"loss": 0.1306, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.133846153846155e-06, |
|
"loss": 0.1043, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.121025641025641e-06, |
|
"loss": 0.1095, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.108205128205128e-06, |
|
"loss": 0.1194, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.095384615384617e-06, |
|
"loss": 0.1209, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.082564102564104e-06, |
|
"loss": 0.1108, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.06974358974359e-06, |
|
"loss": 0.1059, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.056923076923077e-06, |
|
"loss": 0.0923, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.044102564102566e-06, |
|
"loss": 0.1027, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.031282051282053e-06, |
|
"loss": 0.1008, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.01846153846154e-06, |
|
"loss": 0.0885, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.005641025641026e-06, |
|
"loss": 0.0916, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 7.992820512820515e-06, |
|
"loss": 0.1045, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 7.980000000000002e-06, |
|
"loss": 0.0954, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 7.967179487179488e-06, |
|
"loss": 0.1013, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 7.954358974358975e-06, |
|
"loss": 0.0983, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.941538461538462e-06, |
|
"loss": 0.1021, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.928717948717949e-06, |
|
"loss": 0.1024, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.915897435897436e-06, |
|
"loss": 0.1068, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.903076923076922e-06, |
|
"loss": 0.1057, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.890256410256411e-06, |
|
"loss": 0.1144, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.877435897435898e-06, |
|
"loss": 0.1401, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.864615384615385e-06, |
|
"loss": 0.1234, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.851794871794871e-06, |
|
"loss": 0.1364, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.83897435897436e-06, |
|
"loss": 0.1367, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.826153846153847e-06, |
|
"loss": 0.1399, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.813333333333334e-06, |
|
"loss": 0.1387, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.80051282051282e-06, |
|
"loss": 0.1429, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.787692307692309e-06, |
|
"loss": 0.134, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.774871794871796e-06, |
|
"loss": 0.1615, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.762051282051283e-06, |
|
"loss": 0.1502, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.74923076923077e-06, |
|
"loss": 0.1574, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.736410256410258e-06, |
|
"loss": 0.1529, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.723589743589745e-06, |
|
"loss": 0.1406, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.710769230769232e-06, |
|
"loss": 0.1363, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.697948717948718e-06, |
|
"loss": 0.14, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 0.18212890625, |
|
"eval_runtime": 217.0927, |
|
"eval_samples_per_second": 4.606, |
|
"eval_steps_per_second": 0.147, |
|
"eval_wer": 13.005780346820808, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.686153846153846e-06, |
|
"loss": 0.1421, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.673333333333333e-06, |
|
"loss": 0.1507, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.660512820512822e-06, |
|
"loss": 0.1445, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.647692307692309e-06, |
|
"loss": 0.1517, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.634871794871795e-06, |
|
"loss": 0.1363, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.622051282051282e-06, |
|
"loss": 0.1548, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.60923076923077e-06, |
|
"loss": 0.1403, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.596410256410257e-06, |
|
"loss": 0.1543, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.5835897435897444e-06, |
|
"loss": 0.1386, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.570769230769231e-06, |
|
"loss": 0.1194, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.557948717948719e-06, |
|
"loss": 0.1075, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.545128205128206e-06, |
|
"loss": 0.1076, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.5323076923076934e-06, |
|
"loss": 0.1034, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.51948717948718e-06, |
|
"loss": 0.1096, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.506666666666668e-06, |
|
"loss": 0.1051, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.493846153846155e-06, |
|
"loss": 0.0989, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.481025641025642e-06, |
|
"loss": 0.1002, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.468205128205129e-06, |
|
"loss": 0.0961, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.455384615384615e-06, |
|
"loss": 0.1017, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.442564102564103e-06, |
|
"loss": 0.088, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.42974358974359e-06, |
|
"loss": 0.0921, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.416923076923077e-06, |
|
"loss": 0.0921, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.404102564102564e-06, |
|
"loss": 0.0853, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.391282051282052e-06, |
|
"loss": 0.0752, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.378461538461539e-06, |
|
"loss": 0.0705, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.365641025641026e-06, |
|
"loss": 0.0743, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.352820512820513e-06, |
|
"loss": 0.0859, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.340000000000001e-06, |
|
"loss": 0.0759, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.327179487179488e-06, |
|
"loss": 0.0821, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.3143589743589745e-06, |
|
"loss": 0.0824, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.301538461538462e-06, |
|
"loss": 0.0747, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.288717948717949e-06, |
|
"loss": 0.072, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.275897435897437e-06, |
|
"loss": 0.0785, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.2630769230769235e-06, |
|
"loss": 0.0751, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.250256410256411e-06, |
|
"loss": 0.0822, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.237435897435898e-06, |
|
"loss": 0.101, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.224615384615386e-06, |
|
"loss": 0.0896, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.2117948717948725e-06, |
|
"loss": 0.1193, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.19897435897436e-06, |
|
"loss": 0.0841, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.186153846153847e-06, |
|
"loss": 0.0872, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 0.1734619140625, |
|
"eval_runtime": 224.4062, |
|
"eval_samples_per_second": 4.456, |
|
"eval_steps_per_second": 0.143, |
|
"eval_wer": 12.226690123146518, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.174358974358975e-06, |
|
"loss": 0.1113, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.161538461538462e-06, |
|
"loss": 0.1046, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.1487179487179495e-06, |
|
"loss": 0.1319, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.135897435897436e-06, |
|
"loss": 0.1351, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.123076923076924e-06, |
|
"loss": 0.1141, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.110256410256411e-06, |
|
"loss": 0.103, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.0974358974358985e-06, |
|
"loss": 0.0985, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.084615384615385e-06, |
|
"loss": 0.1041, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.071794871794872e-06, |
|
"loss": 0.1041, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.05897435897436e-06, |
|
"loss": 0.0999, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.046153846153847e-06, |
|
"loss": 0.1036, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.033333333333334e-06, |
|
"loss": 0.1013, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.020512820512821e-06, |
|
"loss": 0.1192, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.007692307692309e-06, |
|
"loss": 0.1138, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.994871794871796e-06, |
|
"loss": 0.1082, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.982051282051283e-06, |
|
"loss": 0.1008, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.96923076923077e-06, |
|
"loss": 0.1291, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.956410256410256e-06, |
|
"loss": 0.1273, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.943589743589744e-06, |
|
"loss": 0.1468, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.930769230769231e-06, |
|
"loss": 0.0954, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.917948717948718e-06, |
|
"loss": 0.0813, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.905128205128205e-06, |
|
"loss": 0.0983, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.892307692307693e-06, |
|
"loss": 0.0891, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.87948717948718e-06, |
|
"loss": 0.082, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.866666666666667e-06, |
|
"loss": 0.0896, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.853846153846154e-06, |
|
"loss": 0.097, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.841025641025642e-06, |
|
"loss": 0.086, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.828205128205129e-06, |
|
"loss": 0.0813, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.815384615384616e-06, |
|
"loss": 0.0909, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.802564102564103e-06, |
|
"loss": 0.0981, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.789743589743591e-06, |
|
"loss": 0.0829, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.776923076923078e-06, |
|
"loss": 0.0816, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.764102564102564e-06, |
|
"loss": 0.0854, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.751282051282052e-06, |
|
"loss": 0.0752, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.738461538461539e-06, |
|
"loss": 0.0848, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.725641025641027e-06, |
|
"loss": 0.0751, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.712820512820513e-06, |
|
"loss": 0.0752, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.700000000000001e-06, |
|
"loss": 0.08, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.687179487179488e-06, |
|
"loss": 0.0732, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.6743589743589756e-06, |
|
"loss": 0.0782, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 0.163330078125, |
|
"eval_runtime": 212.6202, |
|
"eval_samples_per_second": 4.703, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 11.221412415179694, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.662564102564103e-06, |
|
"loss": 0.0736, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.64974358974359e-06, |
|
"loss": 0.082, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.636923076923077e-06, |
|
"loss": 0.0735, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.624102564102565e-06, |
|
"loss": 0.0672, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.611282051282052e-06, |
|
"loss": 0.0948, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.598461538461539e-06, |
|
"loss": 0.0995, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.585641025641026e-06, |
|
"loss": 0.0758, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.572820512820514e-06, |
|
"loss": 0.0699, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.560000000000001e-06, |
|
"loss": 0.0644, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.547179487179488e-06, |
|
"loss": 0.0661, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.534358974358975e-06, |
|
"loss": 0.0689, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.521538461538463e-06, |
|
"loss": 0.0739, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.50871794871795e-06, |
|
"loss": 0.0871, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.4958974358974365e-06, |
|
"loss": 0.0785, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.483076923076924e-06, |
|
"loss": 0.073, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.47025641025641e-06, |
|
"loss": 0.0762, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.457435897435898e-06, |
|
"loss": 0.0977, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.444615384615385e-06, |
|
"loss": 0.0786, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.431794871794872e-06, |
|
"loss": 0.0808, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.418974358974359e-06, |
|
"loss": 0.0926, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.406153846153846e-06, |
|
"loss": 0.0825, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.393333333333334e-06, |
|
"loss": 0.0887, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.3805128205128205e-06, |
|
"loss": 0.0924, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.367692307692308e-06, |
|
"loss": 0.0935, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.354871794871795e-06, |
|
"loss": 0.0917, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.342051282051283e-06, |
|
"loss": 0.0755, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.3292307692307695e-06, |
|
"loss": 0.0776, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.316410256410257e-06, |
|
"loss": 0.1006, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.303589743589744e-06, |
|
"loss": 0.0914, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.290769230769232e-06, |
|
"loss": 0.0831, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.2779487179487185e-06, |
|
"loss": 0.0716, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.265128205128206e-06, |
|
"loss": 0.074, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.252307692307693e-06, |
|
"loss": 0.0734, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.239487179487181e-06, |
|
"loss": 0.0797, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.2266666666666675e-06, |
|
"loss": 0.0735, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.213846153846155e-06, |
|
"loss": 0.0783, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.201025641025642e-06, |
|
"loss": 0.0749, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.188205128205129e-06, |
|
"loss": 0.0763, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.1753846153846165e-06, |
|
"loss": 0.0712, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.1625641025641024e-06, |
|
"loss": 0.0718, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 0.1585693359375, |
|
"eval_runtime": 230.5512, |
|
"eval_samples_per_second": 4.337, |
|
"eval_steps_per_second": 0.139, |
|
"eval_wer": 11.862276954008545, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.150769230769231e-06, |
|
"loss": 0.0669, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.137948717948718e-06, |
|
"loss": 0.0877, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.125128205128206e-06, |
|
"loss": 0.0829, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.112307692307693e-06, |
|
"loss": 0.0724, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.09948717948718e-06, |
|
"loss": 0.0613, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.086666666666667e-06, |
|
"loss": 0.0643, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.073846153846155e-06, |
|
"loss": 0.0621, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.061025641025642e-06, |
|
"loss": 0.0529, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.048205128205129e-06, |
|
"loss": 0.0768, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.035384615384616e-06, |
|
"loss": 0.0704, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.022564102564104e-06, |
|
"loss": 0.0572, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.009743589743591e-06, |
|
"loss": 0.0502, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.996923076923078e-06, |
|
"loss": 0.0644, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.984102564102564e-06, |
|
"loss": 0.0729, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.971282051282051e-06, |
|
"loss": 0.0597, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.958461538461539e-06, |
|
"loss": 0.0637, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.9456410256410256e-06, |
|
"loss": 0.0682, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.932820512820513e-06, |
|
"loss": 0.0672, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.92e-06, |
|
"loss": 0.0555, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.907179487179488e-06, |
|
"loss": 0.0579, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.8943589743589746e-06, |
|
"loss": 0.0735, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.881538461538462e-06, |
|
"loss": 0.0653, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.868717948717949e-06, |
|
"loss": 0.0712, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.855897435897437e-06, |
|
"loss": 0.0593, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.8430769230769236e-06, |
|
"loss": 0.0642, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.83025641025641e-06, |
|
"loss": 0.0638, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.817435897435898e-06, |
|
"loss": 0.0586, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.804615384615385e-06, |
|
"loss": 0.0545, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.7917948717948726e-06, |
|
"loss": 0.0599, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.778974358974359e-06, |
|
"loss": 0.0577, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.766153846153847e-06, |
|
"loss": 0.0671, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.753333333333334e-06, |
|
"loss": 0.0673, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.7405128205128215e-06, |
|
"loss": 0.0588, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.727692307692308e-06, |
|
"loss": 0.1067, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.714871794871796e-06, |
|
"loss": 0.0759, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.702051282051283e-06, |
|
"loss": 0.0631, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.6892307692307705e-06, |
|
"loss": 0.0695, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.6764102564102565e-06, |
|
"loss": 0.0923, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.663589743589743e-06, |
|
"loss": 0.0774, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.650769230769231e-06, |
|
"loss": 0.0685, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 0.1539306640625, |
|
"eval_runtime": 210.7067, |
|
"eval_samples_per_second": 4.746, |
|
"eval_steps_per_second": 0.152, |
|
"eval_wer": 10.442322191505403, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.63897435897436e-06, |
|
"loss": 0.0664, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.626153846153847e-06, |
|
"loss": 0.0775, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.613333333333334e-06, |
|
"loss": 0.0562, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.600512820512821e-06, |
|
"loss": 0.0666, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.587692307692308e-06, |
|
"loss": 0.0623, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.574871794871796e-06, |
|
"loss": 0.069, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.5620512820512825e-06, |
|
"loss": 0.0661, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.54923076923077e-06, |
|
"loss": 0.141, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.536410256410257e-06, |
|
"loss": 0.0988, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.523589743589745e-06, |
|
"loss": 0.0718, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.5107692307692315e-06, |
|
"loss": 0.0821, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.4979487179487175e-06, |
|
"loss": 0.0823, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.485128205128205e-06, |
|
"loss": 0.0631, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.472307692307692e-06, |
|
"loss": 0.0911, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.45948717948718e-06, |
|
"loss": 0.064, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.4466666666666665e-06, |
|
"loss": 0.0708, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.433846153846154e-06, |
|
"loss": 0.0676, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.421025641025641e-06, |
|
"loss": 0.0693, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.408205128205129e-06, |
|
"loss": 0.0684, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.3953846153846155e-06, |
|
"loss": 0.0661, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.382564102564103e-06, |
|
"loss": 0.0649, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.36974358974359e-06, |
|
"loss": 0.0593, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.356923076923078e-06, |
|
"loss": 0.0678, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.3441025641025645e-06, |
|
"loss": 0.0789, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.331282051282052e-06, |
|
"loss": 0.0627, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.318461538461539e-06, |
|
"loss": 0.0734, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.305641025641027e-06, |
|
"loss": 0.0721, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.2928205128205134e-06, |
|
"loss": 0.0688, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.28e-06, |
|
"loss": 0.0717, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.267179487179488e-06, |
|
"loss": 0.0588, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.254358974358975e-06, |
|
"loss": 0.0601, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2415384615384624e-06, |
|
"loss": 0.0668, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.228717948717949e-06, |
|
"loss": 0.0611, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.215897435897437e-06, |
|
"loss": 0.0628, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.203076923076924e-06, |
|
"loss": 0.0577, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.19025641025641e-06, |
|
"loss": 0.0583, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.177435897435897e-06, |
|
"loss": 0.0633, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.164615384615384e-06, |
|
"loss": 0.0572, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.151794871794872e-06, |
|
"loss": 0.0505, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.138974358974359e-06, |
|
"loss": 0.0557, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.14794921875, |
|
"eval_runtime": 208.1675, |
|
"eval_samples_per_second": 4.804, |
|
"eval_steps_per_second": 0.154, |
|
"eval_wer": 9.801457652676552, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.127179487179488e-06, |
|
"loss": 0.0446, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.114358974358975e-06, |
|
"loss": 0.0585, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.101538461538462e-06, |
|
"loss": 0.0514, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.08871794871795e-06, |
|
"loss": 0.0605, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.075897435897437e-06, |
|
"loss": 0.05, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.063076923076924e-06, |
|
"loss": 0.0596, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.050256410256411e-06, |
|
"loss": 0.0486, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.037435897435899e-06, |
|
"loss": 0.0439, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.0246153846153856e-06, |
|
"loss": 0.0464, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.0117948717948715e-06, |
|
"loss": 0.0445, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.998974358974359e-06, |
|
"loss": 0.0588, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.986153846153847e-06, |
|
"loss": 0.0514, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.973333333333334e-06, |
|
"loss": 0.0558, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.960512820512821e-06, |
|
"loss": 0.073, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.947692307692308e-06, |
|
"loss": 0.0533, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.934871794871795e-06, |
|
"loss": 0.0507, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.922051282051282e-06, |
|
"loss": 0.0435, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.9092307692307695e-06, |
|
"loss": 0.0445, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.896410256410256e-06, |
|
"loss": 0.0508, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.883589743589744e-06, |
|
"loss": 0.0533, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.870769230769231e-06, |
|
"loss": 0.0607, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.8579487179487185e-06, |
|
"loss": 0.0603, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.845128205128205e-06, |
|
"loss": 0.0867, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.832307692307693e-06, |
|
"loss": 0.0616, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.81948717948718e-06, |
|
"loss": 0.0797, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.8066666666666675e-06, |
|
"loss": 0.0707, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.793846153846154e-06, |
|
"loss": 0.0677, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.781025641025641e-06, |
|
"loss": 0.0722, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.768205128205128e-06, |
|
"loss": 0.0798, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.755384615384616e-06, |
|
"loss": 0.0804, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.7425641025641025e-06, |
|
"loss": 0.0689, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.72974358974359e-06, |
|
"loss": 0.0639, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.716923076923077e-06, |
|
"loss": 0.072, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.704102564102565e-06, |
|
"loss": 0.0674, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.6912820512820515e-06, |
|
"loss": 0.0583, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.678461538461539e-06, |
|
"loss": 0.0598, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.665641025641026e-06, |
|
"loss": 0.0461, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.652820512820514e-06, |
|
"loss": 0.0656, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.6400000000000005e-06, |
|
"loss": 0.0441, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.627179487179487e-06, |
|
"loss": 0.0528, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"eval_loss": 0.1502685546875, |
|
"eval_runtime": 211.4155, |
|
"eval_samples_per_second": 4.73, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 10.040211108318672, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.615384615384616e-06, |
|
"loss": 0.0608, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.602564102564103e-06, |
|
"loss": 0.0585, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.58974358974359e-06, |
|
"loss": 0.0528, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.5769230769230775e-06, |
|
"loss": 0.0402, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.564102564102564e-06, |
|
"loss": 0.0412, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.551282051282052e-06, |
|
"loss": 0.0428, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.538461538461539e-06, |
|
"loss": 0.0398, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.525641025641026e-06, |
|
"loss": 0.04, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.512820512820513e-06, |
|
"loss": 0.0443, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.0425, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.487179487179488e-06, |
|
"loss": 0.0555, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.474358974358975e-06, |
|
"loss": 0.0677, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.461538461538462e-06, |
|
"loss": 0.0718, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.448717948717949e-06, |
|
"loss": 0.0592, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.435897435897436e-06, |
|
"loss": 0.0585, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.423076923076924e-06, |
|
"loss": 0.0687, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.4102564102564104e-06, |
|
"loss": 0.073, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.397435897435898e-06, |
|
"loss": 0.0585, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.384615384615385e-06, |
|
"loss": 0.0584, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.371794871794872e-06, |
|
"loss": 0.0609, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.358974358974359e-06, |
|
"loss": 0.0647, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.346153846153846e-06, |
|
"loss": 0.0612, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.333333333333334e-06, |
|
"loss": 0.0523, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.320512820512821e-06, |
|
"loss": 0.0542, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.307692307692308e-06, |
|
"loss": 0.0536, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.294871794871795e-06, |
|
"loss": 0.0529, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.282051282051282e-06, |
|
"loss": 0.0471, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.26923076923077e-06, |
|
"loss": 0.0567, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.2564102564102566e-06, |
|
"loss": 0.0753, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.243589743589744e-06, |
|
"loss": 0.0535, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.230769230769231e-06, |
|
"loss": 0.0549, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.217948717948718e-06, |
|
"loss": 0.0526, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.2051282051282055e-06, |
|
"loss": 0.0497, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.192307692307692e-06, |
|
"loss": 0.0537, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.17948717948718e-06, |
|
"loss": 0.0691, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.166666666666667e-06, |
|
"loss": 0.0696, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.1538461538461545e-06, |
|
"loss": 0.0623, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.141025641025641e-06, |
|
"loss": 0.0538, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.128205128205128e-06, |
|
"loss": 0.0582, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.115384615384616e-06, |
|
"loss": 0.0733, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"eval_loss": 0.1458740234375, |
|
"eval_runtime": 211.5744, |
|
"eval_samples_per_second": 4.726, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 9.788891681326966, |
|
"step": 12000 |
|
} |
|
], |
|
"max_steps": 20000, |
|
"num_train_epochs": 9223372036854775807, |
|
"total_flos": 7.838245771089543e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|