|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 24.999766967123723, |
|
"eval_steps": 500, |
|
"global_step": 335250, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.992647278150634e-05, |
|
"loss": 5.3605, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.985190156599553e-05, |
|
"loss": 3.4845, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.9777330350484714e-05, |
|
"loss": 2.9795, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.97027591349739e-05, |
|
"loss": 2.6753, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.9628187919463085e-05, |
|
"loss": 2.485, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.9553616703952274e-05, |
|
"loss": 2.3458, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.947904548844147e-05, |
|
"loss": 2.2333, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.940447427293065e-05, |
|
"loss": 2.1338, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.932990305741984e-05, |
|
"loss": 2.0561, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.925533184190902e-05, |
|
"loss": 2.0012, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.918076062639822e-05, |
|
"loss": 1.9287, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.91061894108874e-05, |
|
"loss": 1.8906, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.903161819537659e-05, |
|
"loss": 1.8421, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.895704697986577e-05, |
|
"loss": 1.8052, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.8882475764354966e-05, |
|
"loss": 1.7832, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.880790454884415e-05, |
|
"loss": 1.7501, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.8733333333333337e-05, |
|
"loss": 1.7011, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.865876211782252e-05, |
|
"loss": 1.6742, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.858419090231171e-05, |
|
"loss": 1.6682, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.8509619686800896e-05, |
|
"loss": 1.6518, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.8435048471290085e-05, |
|
"loss": 1.6028, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.8360477255779274e-05, |
|
"loss": 1.5983, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.8285906040268456e-05, |
|
"loss": 1.5778, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.821133482475765e-05, |
|
"loss": 1.5532, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.813691275167786e-05, |
|
"loss": 1.5414, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.8062341536167046e-05, |
|
"loss": 1.5212, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.798791946308725e-05, |
|
"loss": 1.5105, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.7913497390007464e-05, |
|
"loss": 1.4837, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.7838926174496646e-05, |
|
"loss": 1.4661, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.7764354958985834e-05, |
|
"loss": 1.4747, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 4.7689783743475016e-05, |
|
"loss": 1.4458, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 4.761521252796421e-05, |
|
"loss": 1.4461, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 4.7540641312453394e-05, |
|
"loss": 1.4218, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.746607009694258e-05, |
|
"loss": 1.3968, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 4.7391498881431765e-05, |
|
"loss": 1.3959, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 4.731692766592096e-05, |
|
"loss": 1.3846, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.724235645041014e-05, |
|
"loss": 1.3882, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 4.716778523489933e-05, |
|
"loss": 1.3611, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 4.709321401938852e-05, |
|
"loss": 1.3609, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 4.70186428038777e-05, |
|
"loss": 1.3447, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.69440715883669e-05, |
|
"loss": 1.3426, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.68696495152871e-05, |
|
"loss": 1.3421, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.679507829977629e-05, |
|
"loss": 1.3309, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.6720507084265474e-05, |
|
"loss": 1.2993, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.6646085011185686e-05, |
|
"loss": 1.3166, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.657151379567487e-05, |
|
"loss": 1.3053, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 4.649694258016406e-05, |
|
"loss": 1.3013, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.6422371364653246e-05, |
|
"loss": 1.2956, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.6347800149142434e-05, |
|
"loss": 1.2732, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.6273228933631616e-05, |
|
"loss": 1.2753, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.6198657718120805e-05, |
|
"loss": 1.2672, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.6124086502609994e-05, |
|
"loss": 1.261, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.604951528709918e-05, |
|
"loss": 1.2585, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.597509321401939e-05, |
|
"loss": 1.2504, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.590052199850858e-05, |
|
"loss": 1.2506, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.5825950782997766e-05, |
|
"loss": 1.2363, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.5751379567486955e-05, |
|
"loss": 1.222, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 4.5676808351976143e-05, |
|
"loss": 1.2189, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.560238627889635e-05, |
|
"loss": 1.2014, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.552781506338554e-05, |
|
"loss": 1.2192, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.545324384787472e-05, |
|
"loss": 1.1989, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.537867263236391e-05, |
|
"loss": 1.1974, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.53041014168531e-05, |
|
"loss": 1.2056, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.5229530201342286e-05, |
|
"loss": 1.1954, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.515510812826249e-05, |
|
"loss": 1.1942, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.508053691275168e-05, |
|
"loss": 1.1946, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.500596569724086e-05, |
|
"loss": 1.174, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 4.493139448173005e-05, |
|
"loss": 1.1633, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.485682326621924e-05, |
|
"loss": 1.1667, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.478240119313945e-05, |
|
"loss": 1.1628, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 4.470797912005966e-05, |
|
"loss": 1.1526, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 4.463340790454885e-05, |
|
"loss": 1.1463, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.4558836689038036e-05, |
|
"loss": 1.1465, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.448426547352722e-05, |
|
"loss": 1.1445, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.440984340044743e-05, |
|
"loss": 1.1454, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 4.433527218493662e-05, |
|
"loss": 1.1362, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.426070096942581e-05, |
|
"loss": 1.1308, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 4.418612975391499e-05, |
|
"loss": 1.1211, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 4.411155853840418e-05, |
|
"loss": 1.1337, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 4.403698732289336e-05, |
|
"loss": 1.1272, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 4.396256524981357e-05, |
|
"loss": 1.1187, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 4.388799403430276e-05, |
|
"loss": 1.1082, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 4.381342281879195e-05, |
|
"loss": 1.1056, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 4.373885160328113e-05, |
|
"loss": 1.1042, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 4.366428038777032e-05, |
|
"loss": 1.1029, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 4.358970917225951e-05, |
|
"loss": 1.1082, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 4.35151379567487e-05, |
|
"loss": 1.1002, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 4.344056674123789e-05, |
|
"loss": 1.0933, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.336599552572707e-05, |
|
"loss": 1.0962, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 4.329142431021626e-05, |
|
"loss": 1.0848, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 4.321685309470545e-05, |
|
"loss": 1.0868, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 4.3142281879194636e-05, |
|
"loss": 1.0955, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 4.306771066368382e-05, |
|
"loss": 1.075, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 4.299328859060403e-05, |
|
"loss": 1.0769, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 4.291871737509321e-05, |
|
"loss": 1.0804, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 4.284414615958241e-05, |
|
"loss": 1.0595, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 4.2769724086502613e-05, |
|
"loss": 1.0628, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 4.26951528709918e-05, |
|
"loss": 1.0656, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 4.2620581655480984e-05, |
|
"loss": 1.0656, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 4.254601043997017e-05, |
|
"loss": 1.0523, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 4.247143922445936e-05, |
|
"loss": 1.0538, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 4.239686800894855e-05, |
|
"loss": 1.0577, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 4.232229679343773e-05, |
|
"loss": 1.0548, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 4.2247874720357945e-05, |
|
"loss": 1.0555, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 4.2173303504847134e-05, |
|
"loss": 1.0373, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 4.2098732289336316e-05, |
|
"loss": 1.0473, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 4.2024161073825504e-05, |
|
"loss": 1.0423, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 4.194973900074572e-05, |
|
"loss": 1.0484, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 4.1875167785234906e-05, |
|
"loss": 1.0332, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 4.180059656972409e-05, |
|
"loss": 1.0329, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 4.1726025354213276e-05, |
|
"loss": 1.0273, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 4.165160328113348e-05, |
|
"loss": 1.0288, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 4.157703206562267e-05, |
|
"loss": 1.0206, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 4.150246085011186e-05, |
|
"loss": 1.023, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 4.142788963460105e-05, |
|
"loss": 1.0277, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 4.135331841909023e-05, |
|
"loss": 1.0213, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 4.127874720357942e-05, |
|
"loss": 1.017, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 4.120417598806861e-05, |
|
"loss": 1.0195, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 4.1129604772557797e-05, |
|
"loss": 1.0132, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 4.1055182699478e-05, |
|
"loss": 1.0123, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 4.098061148396719e-05, |
|
"loss": 0.9978, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 4.090604026845638e-05, |
|
"loss": 1.0106, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 4.083146905294556e-05, |
|
"loss": 0.9945, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 4.075689783743476e-05, |
|
"loss": 1.0011, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 4.068232662192394e-05, |
|
"loss": 0.9933, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 4.060790454884415e-05, |
|
"loss": 1.0049, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 4.053348247576436e-05, |
|
"loss": 0.9991, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 4.0458911260253546e-05, |
|
"loss": 0.9994, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 4.038434004474273e-05, |
|
"loss": 0.9937, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 4.030976882923192e-05, |
|
"loss": 0.9941, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 4.0235197613721106e-05, |
|
"loss": 0.9905, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 4.0160626398210294e-05, |
|
"loss": 1.0044, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 4.0086055182699476e-05, |
|
"loss": 0.9899, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 4.0011483967188665e-05, |
|
"loss": 0.9891, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 3.9936912751677854e-05, |
|
"loss": 0.9818, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 3.9862490678598066e-05, |
|
"loss": 0.9858, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 3.978791946308725e-05, |
|
"loss": 0.9752, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 3.971334824757644e-05, |
|
"loss": 0.9735, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 3.9638777032065626e-05, |
|
"loss": 0.9742, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 3.956420581655481e-05, |
|
"loss": 0.9745, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 3.9489634601044003e-05, |
|
"loss": 0.9675, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 3.9415063385533185e-05, |
|
"loss": 0.9649, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 3.93406413124534e-05, |
|
"loss": 0.983, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 3.926607009694258e-05, |
|
"loss": 0.9622, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 3.919149888143177e-05, |
|
"loss": 0.9659, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 3.911692766592096e-05, |
|
"loss": 0.9717, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 3.9042356450410146e-05, |
|
"loss": 0.9503, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 3.896793437733035e-05, |
|
"loss": 0.9699, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 3.889336316181954e-05, |
|
"loss": 0.962, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 3.881879194630872e-05, |
|
"loss": 0.9561, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 3.874422073079791e-05, |
|
"loss": 0.9587, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 3.86696495152871e-05, |
|
"loss": 0.9576, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 3.859507829977629e-05, |
|
"loss": 0.9509, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 3.8520656226696495e-05, |
|
"loss": 0.9565, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 3.844608501118568e-05, |
|
"loss": 0.9555, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 3.837151379567487e-05, |
|
"loss": 0.9573, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 3.829694258016406e-05, |
|
"loss": 0.9459, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 3.822237136465325e-05, |
|
"loss": 0.9394, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 3.814780014914243e-05, |
|
"loss": 0.9417, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 3.8073378076062644e-05, |
|
"loss": 0.9459, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 3.7998806860551826e-05, |
|
"loss": 0.9329, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 3.7924235645041015e-05, |
|
"loss": 0.9297, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"learning_rate": 3.7849664429530204e-05, |
|
"loss": 0.9379, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 3.777509321401939e-05, |
|
"loss": 0.9363, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 3.77006711409396e-05, |
|
"loss": 0.9273, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 3.762609992542879e-05, |
|
"loss": 0.9367, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 3.755152870991797e-05, |
|
"loss": 0.9395, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 3.7476957494407164e-05, |
|
"loss": 0.9302, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 3.7402386278896346e-05, |
|
"loss": 0.9259, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 3.732796420581656e-05, |
|
"loss": 0.9315, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 3.725339299030574e-05, |
|
"loss": 0.9245, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 3.717882177479493e-05, |
|
"loss": 0.9289, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 3.710425055928412e-05, |
|
"loss": 0.9199, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 3.702967934377331e-05, |
|
"loss": 0.9177, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 3.6955108128262496e-05, |
|
"loss": 0.9263, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 3.68806860551827e-05, |
|
"loss": 0.9239, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 3.680611483967189e-05, |
|
"loss": 0.924, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 3.673154362416107e-05, |
|
"loss": 0.918, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 3.665697240865026e-05, |
|
"loss": 0.9203, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 3.658269947800149e-05, |
|
"loss": 0.9224, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 3.650812826249068e-05, |
|
"loss": 0.9149, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 3.643355704697987e-05, |
|
"loss": 0.9134, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 3.635898583146906e-05, |
|
"loss": 0.9185, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 3.628441461595824e-05, |
|
"loss": 0.9132, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 3.620984340044743e-05, |
|
"loss": 0.9159, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 3.6135272184936616e-05, |
|
"loss": 0.9081, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 3.6060700969425805e-05, |
|
"loss": 0.9047, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 3.5986129753914994e-05, |
|
"loss": 0.9107, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 3.5911558538404176e-05, |
|
"loss": 0.9004, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 3.5836987322893364e-05, |
|
"loss": 0.8956, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 3.576256524981357e-05, |
|
"loss": 0.9009, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 3.5687994034302766e-05, |
|
"loss": 0.9023, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 3.561342281879195e-05, |
|
"loss": 0.8972, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 3.5538851603281136e-05, |
|
"loss": 0.8967, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 3.546428038777032e-05, |
|
"loss": 0.8867, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 3.538985831469053e-05, |
|
"loss": 0.9029, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 3.5315436241610737e-05, |
|
"loss": 0.901, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 3.5240865026099925e-05, |
|
"loss": 0.889, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 3.5166293810589114e-05, |
|
"loss": 0.8931, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 3.50917225950783e-05, |
|
"loss": 0.892, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 3.5017151379567485e-05, |
|
"loss": 0.8955, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 3.4942580164056674e-05, |
|
"loss": 0.891, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 3.486800894854586e-05, |
|
"loss": 0.8959, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 3.4793586875466075e-05, |
|
"loss": 0.8927, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 3.471901565995526e-05, |
|
"loss": 0.886, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 3.4644444444444446e-05, |
|
"loss": 0.8811, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 3.4569873228933634e-05, |
|
"loss": 0.8788, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 3.449530201342282e-05, |
|
"loss": 0.8892, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 3.442073079791201e-05, |
|
"loss": 0.8826, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 3.4346159582401194e-05, |
|
"loss": 0.8872, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 3.427158836689038e-05, |
|
"loss": 0.8826, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 3.4197017151379565e-05, |
|
"loss": 0.8824, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 3.412244593586876e-05, |
|
"loss": 0.8805, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 3.404787472035794e-05, |
|
"loss": 0.8835, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 3.397330350484713e-05, |
|
"loss": 0.8835, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 3.3898881431767337e-05, |
|
"loss": 0.877, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 3.3824310216256525e-05, |
|
"loss": 0.8645, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 3.3749739000745714e-05, |
|
"loss": 0.8742, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 3.36751677852349e-05, |
|
"loss": 0.8655, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 3.3600596569724085e-05, |
|
"loss": 0.8711, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 3.352632363907532e-05, |
|
"loss": 0.8715, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 3.34517524235645e-05, |
|
"loss": 0.8717, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 3.337718120805369e-05, |
|
"loss": 0.8632, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 3.330260999254288e-05, |
|
"loss": 0.8666, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 3.3228187919463086e-05, |
|
"loss": 0.8669, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 3.3153616703952275e-05, |
|
"loss": 0.8653, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 3.3079045488441464e-05, |
|
"loss": 0.8602, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 3.300447427293065e-05, |
|
"loss": 0.8666, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 3.2929903057419834e-05, |
|
"loss": 0.8565, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 8.58, |
|
"learning_rate": 3.285533184190902e-05, |
|
"loss": 0.8628, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 3.278076062639821e-05, |
|
"loss": 0.8666, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 3.27061894108874e-05, |
|
"loss": 0.8649, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 3.2631767337807606e-05, |
|
"loss": 0.8605, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 3.2557196122296795e-05, |
|
"loss": 0.8594, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 3.248262490678598e-05, |
|
"loss": 0.8624, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 3.240805369127517e-05, |
|
"loss": 0.8604, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 3.2333482475764355e-05, |
|
"loss": 0.8592, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 3.225906040268457e-05, |
|
"loss": 0.8643, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 3.218448918717375e-05, |
|
"loss": 0.8638, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 3.210991797166294e-05, |
|
"loss": 0.8563, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 3.203534675615213e-05, |
|
"loss": 0.8497, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 3.1960775540641315e-05, |
|
"loss": 0.8588, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 3.188635346756153e-05, |
|
"loss": 0.8427, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 3.181178225205071e-05, |
|
"loss": 0.8571, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 3.17372110365399e-05, |
|
"loss": 0.8445, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 3.166263982102908e-05, |
|
"loss": 0.8491, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 3.1588068605518276e-05, |
|
"loss": 0.8394, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 3.151364653243848e-05, |
|
"loss": 0.8485, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 3.143907531692767e-05, |
|
"loss": 0.8426, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"learning_rate": 3.136450410141685e-05, |
|
"loss": 0.847, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 3.128993288590604e-05, |
|
"loss": 0.8453, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 3.121536167039523e-05, |
|
"loss": 0.8464, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 3.114079045488442e-05, |
|
"loss": 0.8477, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 3.1066368381804625e-05, |
|
"loss": 0.84, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 3.099179716629381e-05, |
|
"loss": 0.8442, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 3.091737509321402e-05, |
|
"loss": 0.8431, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 3.084280387770321e-05, |
|
"loss": 0.842, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 3.0768232662192396e-05, |
|
"loss": 0.8428, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 3.0693661446681585e-05, |
|
"loss": 0.8397, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 3.0619090231170774e-05, |
|
"loss": 0.8439, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 3.0544519015659956e-05, |
|
"loss": 0.8376, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 3.047009694258017e-05, |
|
"loss": 0.835, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 3.039552572706935e-05, |
|
"loss": 0.8321, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 3.0320954511558543e-05, |
|
"loss": 0.8346, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 3.0246383296047725e-05, |
|
"loss": 0.8362, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 3.0171961222967937e-05, |
|
"loss": 0.8297, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 3.0097539149888143e-05, |
|
"loss": 0.841, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"learning_rate": 3.002296793437733e-05, |
|
"loss": 0.8359, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 2.9948396718866517e-05, |
|
"loss": 0.8275, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 2.9873825503355706e-05, |
|
"loss": 0.8226, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 2.979925428784489e-05, |
|
"loss": 0.8322, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 2.972468307233408e-05, |
|
"loss": 0.8213, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 10.18, |
|
"learning_rate": 2.9650111856823265e-05, |
|
"loss": 0.8237, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"learning_rate": 2.9575540641312454e-05, |
|
"loss": 0.8273, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 10.25, |
|
"learning_rate": 2.9500969425801646e-05, |
|
"loss": 0.8243, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 2.9426398210290828e-05, |
|
"loss": 0.82, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 10.33, |
|
"learning_rate": 2.9351826994780017e-05, |
|
"loss": 0.8232, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 10.37, |
|
"learning_rate": 2.9277255779269202e-05, |
|
"loss": 0.8242, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 10.4, |
|
"learning_rate": 2.9202833706189415e-05, |
|
"loss": 0.8229, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 2.91282624906786e-05, |
|
"loss": 0.8172, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 2.905369127516779e-05, |
|
"loss": 0.823, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 10.51, |
|
"learning_rate": 2.897912005965697e-05, |
|
"loss": 0.8224, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 10.55, |
|
"learning_rate": 2.8904548844146163e-05, |
|
"loss": 0.8265, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 10.59, |
|
"learning_rate": 2.8829977628635345e-05, |
|
"loss": 0.8191, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 2.8755406413124537e-05, |
|
"loss": 0.8236, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 2.868083519761372e-05, |
|
"loss": 0.8158, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 10.7, |
|
"learning_rate": 2.860626398210291e-05, |
|
"loss": 0.8221, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 10.74, |
|
"learning_rate": 2.8531841909023117e-05, |
|
"loss": 0.8212, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 10.78, |
|
"learning_rate": 2.8457419835943326e-05, |
|
"loss": 0.825, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 10.81, |
|
"learning_rate": 2.838284862043251e-05, |
|
"loss": 0.8255, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 10.85, |
|
"learning_rate": 2.83082774049217e-05, |
|
"loss": 0.8139, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 10.89, |
|
"learning_rate": 2.8233706189410892e-05, |
|
"loss": 0.8157, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 10.92, |
|
"learning_rate": 2.8159134973900074e-05, |
|
"loss": 0.8177, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 10.96, |
|
"learning_rate": 2.8084563758389266e-05, |
|
"loss": 0.8097, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 2.8009992542878448e-05, |
|
"loss": 0.8206, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 2.793557046979866e-05, |
|
"loss": 0.8084, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 11.07, |
|
"learning_rate": 2.7860999254287846e-05, |
|
"loss": 0.8082, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"learning_rate": 2.7786428038777035e-05, |
|
"loss": 0.8034, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 2.771185682326622e-05, |
|
"loss": 0.8013, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 11.19, |
|
"learning_rate": 2.763728560775541e-05, |
|
"loss": 0.8098, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 11.22, |
|
"learning_rate": 2.7562714392244594e-05, |
|
"loss": 0.8062, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 11.26, |
|
"learning_rate": 2.7488143176733783e-05, |
|
"loss": 0.8112, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 11.3, |
|
"learning_rate": 2.741372110365399e-05, |
|
"loss": 0.8068, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 11.33, |
|
"learning_rate": 2.7339149888143178e-05, |
|
"loss": 0.8069, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 11.37, |
|
"learning_rate": 2.7264578672632363e-05, |
|
"loss": 0.8089, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 11.41, |
|
"learning_rate": 2.7190156599552575e-05, |
|
"loss": 0.8094, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 11.45, |
|
"learning_rate": 2.7115585384041757e-05, |
|
"loss": 0.8012, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 11.48, |
|
"learning_rate": 2.704101416853095e-05, |
|
"loss": 0.8105, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 11.52, |
|
"learning_rate": 2.696644295302014e-05, |
|
"loss": 0.8066, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 11.56, |
|
"learning_rate": 2.6891871737509324e-05, |
|
"loss": 0.8094, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 11.6, |
|
"learning_rate": 2.6817300521998512e-05, |
|
"loss": 0.809, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 11.63, |
|
"learning_rate": 2.6742729306487694e-05, |
|
"loss": 0.7998, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 11.67, |
|
"learning_rate": 2.6668158090976887e-05, |
|
"loss": 0.8112, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 11.71, |
|
"learning_rate": 2.6593736017897092e-05, |
|
"loss": 0.8068, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 11.74, |
|
"learning_rate": 2.651916480238628e-05, |
|
"loss": 0.8042, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 11.78, |
|
"learning_rate": 2.6444593586875466e-05, |
|
"loss": 0.7972, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 11.82, |
|
"learning_rate": 2.6370022371364655e-05, |
|
"loss": 0.8119, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 11.86, |
|
"learning_rate": 2.629545115585384e-05, |
|
"loss": 0.7982, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 11.89, |
|
"learning_rate": 2.622102908277405e-05, |
|
"loss": 0.7973, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 11.93, |
|
"learning_rate": 2.6146457867263235e-05, |
|
"loss": 0.7919, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 11.97, |
|
"learning_rate": 2.6071886651752424e-05, |
|
"loss": 0.8013, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 2.599731543624161e-05, |
|
"loss": 0.7986, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 2.592289336316182e-05, |
|
"loss": 0.7927, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 12.08, |
|
"learning_rate": 2.584832214765101e-05, |
|
"loss": 0.7983, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 12.12, |
|
"learning_rate": 2.5773750932140196e-05, |
|
"loss": 0.7913, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 12.15, |
|
"learning_rate": 2.5699179716629384e-05, |
|
"loss": 0.7979, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 12.19, |
|
"learning_rate": 2.562460850111857e-05, |
|
"loss": 0.7941, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 12.23, |
|
"learning_rate": 2.555003728560776e-05, |
|
"loss": 0.7871, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 12.27, |
|
"learning_rate": 2.5475466070096944e-05, |
|
"loss": 0.7927, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 12.3, |
|
"learning_rate": 2.5400894854586133e-05, |
|
"loss": 0.7908, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 12.34, |
|
"learning_rate": 2.532647278150634e-05, |
|
"loss": 0.7872, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 12.38, |
|
"learning_rate": 2.5251901565995527e-05, |
|
"loss": 0.7883, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 12.42, |
|
"learning_rate": 2.5177330350484713e-05, |
|
"loss": 0.785, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 12.45, |
|
"learning_rate": 2.51027591349739e-05, |
|
"loss": 0.7894, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 12.49, |
|
"learning_rate": 2.5028337061894107e-05, |
|
"loss": 0.7822, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 12.53, |
|
"learning_rate": 2.49537658463833e-05, |
|
"loss": 0.7952, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 2.4879194630872485e-05, |
|
"loss": 0.787, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 12.6, |
|
"learning_rate": 2.4804623415361673e-05, |
|
"loss": 0.7909, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 12.64, |
|
"learning_rate": 2.473005219985086e-05, |
|
"loss": 0.7913, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 12.68, |
|
"learning_rate": 2.4655480984340047e-05, |
|
"loss": 0.7831, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 12.71, |
|
"learning_rate": 2.4580909768829233e-05, |
|
"loss": 0.7899, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 12.75, |
|
"learning_rate": 2.4506338553318418e-05, |
|
"loss": 0.79, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 12.79, |
|
"learning_rate": 2.4431767337807607e-05, |
|
"loss": 0.7897, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 12.83, |
|
"learning_rate": 2.4357345264727816e-05, |
|
"loss": 0.7813, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 12.86, |
|
"learning_rate": 2.4282774049217e-05, |
|
"loss": 0.7852, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 12.9, |
|
"learning_rate": 2.420820283370619e-05, |
|
"loss": 0.7856, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 12.94, |
|
"learning_rate": 2.4133631618195376e-05, |
|
"loss": 0.7859, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 12.98, |
|
"learning_rate": 2.4059209545115585e-05, |
|
"loss": 0.7854, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 2.3984787472035794e-05, |
|
"loss": 0.7768, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 13.05, |
|
"learning_rate": 2.3910216256524982e-05, |
|
"loss": 0.7776, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 2.383564504101417e-05, |
|
"loss": 0.7801, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 13.12, |
|
"learning_rate": 2.3761073825503357e-05, |
|
"loss": 0.7801, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 13.16, |
|
"learning_rate": 2.3686502609992545e-05, |
|
"loss": 0.7767, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 13.2, |
|
"learning_rate": 2.3612080536912754e-05, |
|
"loss": 0.7765, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 13.24, |
|
"learning_rate": 2.353750932140194e-05, |
|
"loss": 0.7761, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 13.27, |
|
"learning_rate": 2.346293810589113e-05, |
|
"loss": 0.7767, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 13.31, |
|
"learning_rate": 2.3388366890380314e-05, |
|
"loss": 0.7699, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 13.35, |
|
"learning_rate": 2.3313795674869503e-05, |
|
"loss": 0.7793, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 13.39, |
|
"learning_rate": 2.3239373601789712e-05, |
|
"loss": 0.7755, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 13.42, |
|
"learning_rate": 2.3164802386278897e-05, |
|
"loss": 0.767, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 13.46, |
|
"learning_rate": 2.3090231170768082e-05, |
|
"loss": 0.7779, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 13.5, |
|
"learning_rate": 2.301565995525727e-05, |
|
"loss": 0.7851, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 13.53, |
|
"learning_rate": 2.2941088739746457e-05, |
|
"loss": 0.7793, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 13.57, |
|
"learning_rate": 2.2866517524235645e-05, |
|
"loss": 0.7716, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 13.61, |
|
"learning_rate": 2.279194630872483e-05, |
|
"loss": 0.7818, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 13.65, |
|
"learning_rate": 2.2717375093214023e-05, |
|
"loss": 0.7779, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 13.68, |
|
"learning_rate": 2.2642953020134232e-05, |
|
"loss": 0.7739, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 13.72, |
|
"learning_rate": 2.2568381804623417e-05, |
|
"loss": 0.766, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 13.76, |
|
"learning_rate": 2.2493810589112603e-05, |
|
"loss": 0.7742, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 13.8, |
|
"learning_rate": 2.2419388516032812e-05, |
|
"loss": 0.767, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 13.83, |
|
"learning_rate": 2.2344817300522e-05, |
|
"loss": 0.7672, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 13.87, |
|
"learning_rate": 2.2270246085011186e-05, |
|
"loss": 0.7774, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 13.91, |
|
"learning_rate": 2.2195674869500375e-05, |
|
"loss": 0.7696, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 13.94, |
|
"learning_rate": 2.2121252796420584e-05, |
|
"loss": 0.7736, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 13.98, |
|
"learning_rate": 2.204668158090977e-05, |
|
"loss": 0.7705, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 2.1972110365398958e-05, |
|
"loss": 0.7768, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 14.06, |
|
"learning_rate": 2.1897539149888143e-05, |
|
"loss": 0.7629, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 14.09, |
|
"learning_rate": 2.1822967934377332e-05, |
|
"loss": 0.758, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 14.13, |
|
"learning_rate": 2.174854586129754e-05, |
|
"loss": 0.7726, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 14.17, |
|
"learning_rate": 2.1673974645786726e-05, |
|
"loss": 0.765, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 14.21, |
|
"learning_rate": 2.1599403430275912e-05, |
|
"loss": 0.7639, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 14.24, |
|
"learning_rate": 2.15248322147651e-05, |
|
"loss": 0.7613, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 14.28, |
|
"learning_rate": 2.145026099925429e-05, |
|
"loss": 0.7624, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 14.32, |
|
"learning_rate": 2.13758389261745e-05, |
|
"loss": 0.7682, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 14.35, |
|
"learning_rate": 2.1301267710663687e-05, |
|
"loss": 0.7683, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 14.39, |
|
"learning_rate": 2.1226696495152873e-05, |
|
"loss": 0.7684, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 14.43, |
|
"learning_rate": 2.115212527964206e-05, |
|
"loss": 0.7561, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 14.47, |
|
"learning_rate": 2.1077554064131247e-05, |
|
"loss": 0.7672, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 14.5, |
|
"learning_rate": 2.1002982848620435e-05, |
|
"loss": 0.7658, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 14.54, |
|
"learning_rate": 2.092841163310962e-05, |
|
"loss": 0.7625, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 14.58, |
|
"learning_rate": 2.0853840417598806e-05, |
|
"loss": 0.757, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 14.62, |
|
"learning_rate": 2.0779269202087995e-05, |
|
"loss": 0.7596, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 14.65, |
|
"learning_rate": 2.0704847129008204e-05, |
|
"loss": 0.7603, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 14.69, |
|
"learning_rate": 2.0630425055928413e-05, |
|
"loss": 0.7666, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 14.73, |
|
"learning_rate": 2.05558538404176e-05, |
|
"loss": 0.7602, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 14.76, |
|
"learning_rate": 2.0481282624906787e-05, |
|
"loss": 0.7575, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 14.8, |
|
"learning_rate": 2.0406711409395973e-05, |
|
"loss": 0.7615, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 14.84, |
|
"learning_rate": 2.033214019388516e-05, |
|
"loss": 0.761, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 14.88, |
|
"learning_rate": 2.025771812080537e-05, |
|
"loss": 0.7628, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 14.91, |
|
"learning_rate": 2.0183296047725576e-05, |
|
"loss": 0.7629, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 14.95, |
|
"learning_rate": 2.0108724832214765e-05, |
|
"loss": 0.7623, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 14.99, |
|
"learning_rate": 2.0034153616703954e-05, |
|
"loss": 0.7528, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 1.9959582401193142e-05, |
|
"loss": 0.7458, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 15.06, |
|
"learning_rate": 1.9885011185682328e-05, |
|
"loss": 0.7596, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 15.1, |
|
"learning_rate": 1.9810439970171517e-05, |
|
"loss": 0.7547, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 1.9735868754660702e-05, |
|
"loss": 0.7581, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 15.18, |
|
"learning_rate": 1.966129753914989e-05, |
|
"loss": 0.7527, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 15.21, |
|
"learning_rate": 1.9586726323639076e-05, |
|
"loss": 0.7592, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 15.25, |
|
"learning_rate": 1.9512304250559285e-05, |
|
"loss": 0.7538, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 15.29, |
|
"learning_rate": 1.943773303504847e-05, |
|
"loss": 0.7466, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 15.32, |
|
"learning_rate": 1.936316181953766e-05, |
|
"loss": 0.7529, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 15.36, |
|
"learning_rate": 1.9288590604026845e-05, |
|
"loss": 0.7533, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 15.4, |
|
"learning_rate": 1.9214019388516033e-05, |
|
"loss": 0.7532, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 15.44, |
|
"learning_rate": 1.9139597315436243e-05, |
|
"loss": 0.7627, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 15.47, |
|
"learning_rate": 1.906517524235645e-05, |
|
"loss": 0.7514, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 15.51, |
|
"learning_rate": 1.8990604026845637e-05, |
|
"loss": 0.7564, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 15.55, |
|
"learning_rate": 1.8916032811334826e-05, |
|
"loss": 0.7587, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 15.59, |
|
"learning_rate": 1.884146159582401e-05, |
|
"loss": 0.7508, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 15.62, |
|
"learning_rate": 1.87668903803132e-05, |
|
"loss": 0.7443, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 15.66, |
|
"learning_rate": 1.869231916480239e-05, |
|
"loss": 0.7514, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 15.7, |
|
"learning_rate": 1.8617747949291574e-05, |
|
"loss": 0.7461, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 15.73, |
|
"learning_rate": 1.8543176733780763e-05, |
|
"loss": 0.7537, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 15.77, |
|
"learning_rate": 1.8468605518269948e-05, |
|
"loss": 0.7532, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 15.81, |
|
"learning_rate": 1.8394034302759137e-05, |
|
"loss": 0.7416, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 15.85, |
|
"learning_rate": 1.8319463087248322e-05, |
|
"loss": 0.7456, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 15.88, |
|
"learning_rate": 1.824489187173751e-05, |
|
"loss": 0.7477, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 15.92, |
|
"learning_rate": 1.8170320656226696e-05, |
|
"loss": 0.749, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 15.96, |
|
"learning_rate": 1.8095749440715885e-05, |
|
"loss": 0.7512, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 1.8021327367636094e-05, |
|
"loss": 0.7422, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 1.794675615212528e-05, |
|
"loss": 0.749, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 16.07, |
|
"learning_rate": 1.787218493661447e-05, |
|
"loss": 0.7409, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 16.11, |
|
"learning_rate": 1.7797613721103654e-05, |
|
"loss": 0.7411, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 16.14, |
|
"learning_rate": 1.7723191648023863e-05, |
|
"loss": 0.7397, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 16.18, |
|
"learning_rate": 1.764862043251305e-05, |
|
"loss": 0.7448, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 16.22, |
|
"learning_rate": 1.757404921700224e-05, |
|
"loss": 0.7482, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 16.26, |
|
"learning_rate": 1.7499478001491426e-05, |
|
"loss": 0.7399, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 16.29, |
|
"learning_rate": 1.7424906785980614e-05, |
|
"loss": 0.7347, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 16.33, |
|
"learning_rate": 1.7350633855331844e-05, |
|
"loss": 0.7431, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 16.37, |
|
"learning_rate": 1.727606263982103e-05, |
|
"loss": 0.7426, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 16.41, |
|
"learning_rate": 1.7201491424310218e-05, |
|
"loss": 0.7437, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 16.44, |
|
"learning_rate": 1.7126920208799403e-05, |
|
"loss": 0.7407, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 16.48, |
|
"learning_rate": 1.7052348993288592e-05, |
|
"loss": 0.7389, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 16.52, |
|
"learning_rate": 1.6977777777777777e-05, |
|
"loss": 0.7369, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 16.55, |
|
"learning_rate": 1.6903206562266966e-05, |
|
"loss": 0.7444, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 16.59, |
|
"learning_rate": 1.682863534675615e-05, |
|
"loss": 0.7385, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 16.63, |
|
"learning_rate": 1.675421327367636e-05, |
|
"loss": 0.7439, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 16.67, |
|
"learning_rate": 1.667964205816555e-05, |
|
"loss": 0.743, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 16.7, |
|
"learning_rate": 1.6605070842654735e-05, |
|
"loss": 0.7422, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 16.74, |
|
"learning_rate": 1.6530499627143924e-05, |
|
"loss": 0.7384, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 16.78, |
|
"learning_rate": 1.645592841163311e-05, |
|
"loss": 0.7388, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 16.82, |
|
"learning_rate": 1.6381357196122298e-05, |
|
"loss": 0.7384, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 16.85, |
|
"learning_rate": 1.6306785980611486e-05, |
|
"loss": 0.7387, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 16.89, |
|
"learning_rate": 1.6232363907531696e-05, |
|
"loss": 0.7306, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 16.93, |
|
"learning_rate": 1.615779269202088e-05, |
|
"loss": 0.7398, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 16.96, |
|
"learning_rate": 1.608322147651007e-05, |
|
"loss": 0.7371, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 1.6008650260999255e-05, |
|
"loss": 0.7357, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 1.5934377330350488e-05, |
|
"loss": 0.7357, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 17.08, |
|
"learning_rate": 1.5859806114839673e-05, |
|
"loss": 0.7397, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 17.11, |
|
"learning_rate": 1.578523489932886e-05, |
|
"loss": 0.735, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 17.15, |
|
"learning_rate": 1.5710663683818047e-05, |
|
"loss": 0.7345, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 17.19, |
|
"learning_rate": 1.5636092468307233e-05, |
|
"loss": 0.7339, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 17.23, |
|
"learning_rate": 1.556152125279642e-05, |
|
"loss": 0.7317, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 17.26, |
|
"learning_rate": 1.5486950037285607e-05, |
|
"loss": 0.7325, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 17.3, |
|
"learning_rate": 1.5412378821774796e-05, |
|
"loss": 0.7328, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 17.34, |
|
"learning_rate": 1.5337956748695005e-05, |
|
"loss": 0.7326, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 17.37, |
|
"learning_rate": 1.526338553318419e-05, |
|
"loss": 0.734, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 17.41, |
|
"learning_rate": 1.5188814317673377e-05, |
|
"loss": 0.7369, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 17.45, |
|
"learning_rate": 1.5114243102162564e-05, |
|
"loss": 0.7334, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 17.49, |
|
"learning_rate": 1.5039671886651755e-05, |
|
"loss": 0.73, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 17.52, |
|
"learning_rate": 1.4965100671140942e-05, |
|
"loss": 0.7294, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 17.56, |
|
"learning_rate": 1.4890529455630129e-05, |
|
"loss": 0.7238, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 17.6, |
|
"learning_rate": 1.4815958240119316e-05, |
|
"loss": 0.7297, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 17.64, |
|
"learning_rate": 1.4741536167039525e-05, |
|
"loss": 0.7347, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 17.67, |
|
"learning_rate": 1.466696495152871e-05, |
|
"loss": 0.7336, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 17.71, |
|
"learning_rate": 1.459254287844892e-05, |
|
"loss": 0.7238, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 17.75, |
|
"learning_rate": 1.4517971662938106e-05, |
|
"loss": 0.7348, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 17.79, |
|
"learning_rate": 1.4443549589858315e-05, |
|
"loss": 0.7348, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 17.82, |
|
"learning_rate": 1.4368978374347503e-05, |
|
"loss": 0.7221, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 17.86, |
|
"learning_rate": 1.429440715883669e-05, |
|
"loss": 0.7276, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 17.9, |
|
"learning_rate": 1.4219835943325877e-05, |
|
"loss": 0.725, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 17.93, |
|
"learning_rate": 1.4145264727815064e-05, |
|
"loss": 0.7271, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 17.97, |
|
"learning_rate": 1.407069351230425e-05, |
|
"loss": 0.7253, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 1.3996122296793438e-05, |
|
"loss": 0.7249, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 18.05, |
|
"learning_rate": 1.3921551081282625e-05, |
|
"loss": 0.7248, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 18.08, |
|
"learning_rate": 1.3846979865771812e-05, |
|
"loss": 0.7279, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 18.12, |
|
"learning_rate": 1.3772408650261e-05, |
|
"loss": 0.7313, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 18.16, |
|
"learning_rate": 1.3697837434750188e-05, |
|
"loss": 0.7249, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 18.2, |
|
"learning_rate": 1.3623266219239375e-05, |
|
"loss": 0.7154, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 18.23, |
|
"learning_rate": 1.3548695003728562e-05, |
|
"loss": 0.7206, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 18.27, |
|
"learning_rate": 1.3474272930648771e-05, |
|
"loss": 0.7222, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 18.31, |
|
"learning_rate": 1.3399701715137958e-05, |
|
"loss": 0.7212, |
|
"step": 245500 |
|
}, |
|
{ |
|
"epoch": 18.34, |
|
"learning_rate": 1.3325130499627145e-05, |
|
"loss": 0.7207, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 18.38, |
|
"learning_rate": 1.3250559284116332e-05, |
|
"loss": 0.7236, |
|
"step": 246500 |
|
}, |
|
{ |
|
"epoch": 18.42, |
|
"learning_rate": 1.3176137211036541e-05, |
|
"loss": 0.7235, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 18.46, |
|
"learning_rate": 1.3101715137956749e-05, |
|
"loss": 0.7207, |
|
"step": 247500 |
|
}, |
|
{ |
|
"epoch": 18.49, |
|
"learning_rate": 1.3027143922445936e-05, |
|
"loss": 0.7201, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 18.53, |
|
"learning_rate": 1.2952572706935123e-05, |
|
"loss": 0.7212, |
|
"step": 248500 |
|
}, |
|
{ |
|
"epoch": 18.57, |
|
"learning_rate": 1.287800149142431e-05, |
|
"loss": 0.7328, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 18.61, |
|
"learning_rate": 1.2803430275913497e-05, |
|
"loss": 0.7278, |
|
"step": 249500 |
|
}, |
|
{ |
|
"epoch": 18.64, |
|
"learning_rate": 1.2728859060402684e-05, |
|
"loss": 0.7188, |
|
"step": 250000 |
|
}, |
|
{ |
|
"epoch": 18.68, |
|
"learning_rate": 1.2654287844891871e-05, |
|
"loss": 0.7212, |
|
"step": 250500 |
|
}, |
|
{ |
|
"epoch": 18.72, |
|
"learning_rate": 1.2579716629381062e-05, |
|
"loss": 0.7273, |
|
"step": 251000 |
|
}, |
|
{ |
|
"epoch": 18.75, |
|
"learning_rate": 1.2505294556301269e-05, |
|
"loss": 0.718, |
|
"step": 251500 |
|
}, |
|
{ |
|
"epoch": 18.79, |
|
"learning_rate": 1.2430723340790456e-05, |
|
"loss": 0.7256, |
|
"step": 252000 |
|
}, |
|
{ |
|
"epoch": 18.83, |
|
"learning_rate": 1.2356152125279641e-05, |
|
"loss": 0.7269, |
|
"step": 252500 |
|
}, |
|
{ |
|
"epoch": 18.87, |
|
"learning_rate": 1.228173005219985e-05, |
|
"loss": 0.7144, |
|
"step": 253000 |
|
}, |
|
{ |
|
"epoch": 18.9, |
|
"learning_rate": 1.220730797912006e-05, |
|
"loss": 0.7169, |
|
"step": 253500 |
|
}, |
|
{ |
|
"epoch": 18.94, |
|
"learning_rate": 1.2132736763609247e-05, |
|
"loss": 0.7225, |
|
"step": 254000 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"learning_rate": 1.2058165548098435e-05, |
|
"loss": 0.7198, |
|
"step": 254500 |
|
}, |
|
{ |
|
"epoch": 19.02, |
|
"learning_rate": 1.1983594332587622e-05, |
|
"loss": 0.7165, |
|
"step": 255000 |
|
}, |
|
{ |
|
"epoch": 19.05, |
|
"learning_rate": 1.190902311707681e-05, |
|
"loss": 0.7114, |
|
"step": 255500 |
|
}, |
|
{ |
|
"epoch": 19.09, |
|
"learning_rate": 1.1834451901565997e-05, |
|
"loss": 0.7191, |
|
"step": 256000 |
|
}, |
|
{ |
|
"epoch": 19.13, |
|
"learning_rate": 1.1759880686055184e-05, |
|
"loss": 0.7164, |
|
"step": 256500 |
|
}, |
|
{ |
|
"epoch": 19.16, |
|
"learning_rate": 1.168530947054437e-05, |
|
"loss": 0.7177, |
|
"step": 257000 |
|
}, |
|
{ |
|
"epoch": 19.2, |
|
"learning_rate": 1.1610738255033558e-05, |
|
"loss": 0.7125, |
|
"step": 257500 |
|
}, |
|
{ |
|
"epoch": 19.24, |
|
"learning_rate": 1.1536167039522745e-05, |
|
"loss": 0.7177, |
|
"step": 258000 |
|
}, |
|
{ |
|
"epoch": 19.28, |
|
"learning_rate": 1.1461595824011932e-05, |
|
"loss": 0.7107, |
|
"step": 258500 |
|
}, |
|
{ |
|
"epoch": 19.31, |
|
"learning_rate": 1.1387024608501119e-05, |
|
"loss": 0.7199, |
|
"step": 259000 |
|
}, |
|
{ |
|
"epoch": 19.35, |
|
"learning_rate": 1.1312453392990306e-05, |
|
"loss": 0.7197, |
|
"step": 259500 |
|
}, |
|
{ |
|
"epoch": 19.39, |
|
"learning_rate": 1.1238031319910515e-05, |
|
"loss": 0.7166, |
|
"step": 260000 |
|
}, |
|
{ |
|
"epoch": 19.43, |
|
"learning_rate": 1.1163460104399702e-05, |
|
"loss": 0.7153, |
|
"step": 260500 |
|
}, |
|
{ |
|
"epoch": 19.46, |
|
"learning_rate": 1.108888888888889e-05, |
|
"loss": 0.7176, |
|
"step": 261000 |
|
}, |
|
{ |
|
"epoch": 19.5, |
|
"learning_rate": 1.1014466815809098e-05, |
|
"loss": 0.7143, |
|
"step": 261500 |
|
}, |
|
{ |
|
"epoch": 19.54, |
|
"learning_rate": 1.0940044742729306e-05, |
|
"loss": 0.7167, |
|
"step": 262000 |
|
}, |
|
{ |
|
"epoch": 19.57, |
|
"learning_rate": 1.0865473527218494e-05, |
|
"loss": 0.7157, |
|
"step": 262500 |
|
}, |
|
{ |
|
"epoch": 19.61, |
|
"learning_rate": 1.0790902311707682e-05, |
|
"loss": 0.7107, |
|
"step": 263000 |
|
}, |
|
{ |
|
"epoch": 19.65, |
|
"learning_rate": 1.0716331096196869e-05, |
|
"loss": 0.7156, |
|
"step": 263500 |
|
}, |
|
{ |
|
"epoch": 19.69, |
|
"learning_rate": 1.0641909023117078e-05, |
|
"loss": 0.7164, |
|
"step": 264000 |
|
}, |
|
{ |
|
"epoch": 19.72, |
|
"learning_rate": 1.0567337807606265e-05, |
|
"loss": 0.7131, |
|
"step": 264500 |
|
}, |
|
{ |
|
"epoch": 19.76, |
|
"learning_rate": 1.0492766592095452e-05, |
|
"loss": 0.7158, |
|
"step": 265000 |
|
}, |
|
{ |
|
"epoch": 19.8, |
|
"learning_rate": 1.0418195376584639e-05, |
|
"loss": 0.7129, |
|
"step": 265500 |
|
}, |
|
{ |
|
"epoch": 19.84, |
|
"learning_rate": 1.0343624161073826e-05, |
|
"loss": 0.7119, |
|
"step": 266000 |
|
}, |
|
{ |
|
"epoch": 19.87, |
|
"learning_rate": 1.0269052945563013e-05, |
|
"loss": 0.7208, |
|
"step": 266500 |
|
}, |
|
{ |
|
"epoch": 19.91, |
|
"learning_rate": 1.01944817300522e-05, |
|
"loss": 0.7132, |
|
"step": 267000 |
|
}, |
|
{ |
|
"epoch": 19.95, |
|
"learning_rate": 1.0119910514541387e-05, |
|
"loss": 0.715, |
|
"step": 267500 |
|
}, |
|
{ |
|
"epoch": 19.98, |
|
"learning_rate": 1.0045339299030574e-05, |
|
"loss": 0.7125, |
|
"step": 268000 |
|
}, |
|
{ |
|
"epoch": 20.02, |
|
"learning_rate": 9.970768083519761e-06, |
|
"loss": 0.7126, |
|
"step": 268500 |
|
}, |
|
{ |
|
"epoch": 20.06, |
|
"learning_rate": 9.89634601043997e-06, |
|
"loss": 0.7056, |
|
"step": 269000 |
|
}, |
|
{ |
|
"epoch": 20.1, |
|
"learning_rate": 9.821774794929157e-06, |
|
"loss": 0.7101, |
|
"step": 269500 |
|
}, |
|
{ |
|
"epoch": 20.13, |
|
"learning_rate": 9.747203579418344e-06, |
|
"loss": 0.7103, |
|
"step": 270000 |
|
}, |
|
{ |
|
"epoch": 20.17, |
|
"learning_rate": 9.672632363907533e-06, |
|
"loss": 0.7109, |
|
"step": 270500 |
|
}, |
|
{ |
|
"epoch": 20.21, |
|
"learning_rate": 9.59806114839672e-06, |
|
"loss": 0.707, |
|
"step": 271000 |
|
}, |
|
{ |
|
"epoch": 20.25, |
|
"learning_rate": 9.52363907531693e-06, |
|
"loss": 0.7121, |
|
"step": 271500 |
|
}, |
|
{ |
|
"epoch": 20.28, |
|
"learning_rate": 9.449067859806115e-06, |
|
"loss": 0.7088, |
|
"step": 272000 |
|
}, |
|
{ |
|
"epoch": 20.32, |
|
"learning_rate": 9.374496644295302e-06, |
|
"loss": 0.7078, |
|
"step": 272500 |
|
}, |
|
{ |
|
"epoch": 20.36, |
|
"learning_rate": 9.299925428784489e-06, |
|
"loss": 0.7094, |
|
"step": 273000 |
|
}, |
|
{ |
|
"epoch": 20.4, |
|
"learning_rate": 9.225354213273676e-06, |
|
"loss": 0.7103, |
|
"step": 273500 |
|
}, |
|
{ |
|
"epoch": 20.43, |
|
"learning_rate": 9.150932140193885e-06, |
|
"loss": 0.7097, |
|
"step": 274000 |
|
}, |
|
{ |
|
"epoch": 20.47, |
|
"learning_rate": 9.076360924683072e-06, |
|
"loss": 0.7092, |
|
"step": 274500 |
|
}, |
|
{ |
|
"epoch": 20.51, |
|
"learning_rate": 9.001938851603281e-06, |
|
"loss": 0.7092, |
|
"step": 275000 |
|
}, |
|
{ |
|
"epoch": 20.54, |
|
"learning_rate": 8.92736763609247e-06, |
|
"loss": 0.7061, |
|
"step": 275500 |
|
}, |
|
{ |
|
"epoch": 20.58, |
|
"learning_rate": 8.852796420581657e-06, |
|
"loss": 0.7076, |
|
"step": 276000 |
|
}, |
|
{ |
|
"epoch": 20.62, |
|
"learning_rate": 8.778225205070844e-06, |
|
"loss": 0.7124, |
|
"step": 276500 |
|
}, |
|
{ |
|
"epoch": 20.66, |
|
"learning_rate": 8.70365398956003e-06, |
|
"loss": 0.7051, |
|
"step": 277000 |
|
}, |
|
{ |
|
"epoch": 20.69, |
|
"learning_rate": 8.629082774049217e-06, |
|
"loss": 0.706, |
|
"step": 277500 |
|
}, |
|
{ |
|
"epoch": 20.73, |
|
"learning_rate": 8.554511558538404e-06, |
|
"loss": 0.7043, |
|
"step": 278000 |
|
}, |
|
{ |
|
"epoch": 20.77, |
|
"learning_rate": 8.47994034302759e-06, |
|
"loss": 0.7081, |
|
"step": 278500 |
|
}, |
|
{ |
|
"epoch": 20.81, |
|
"learning_rate": 8.40536912751678e-06, |
|
"loss": 0.7048, |
|
"step": 279000 |
|
}, |
|
{ |
|
"epoch": 20.84, |
|
"learning_rate": 8.330797912005966e-06, |
|
"loss": 0.715, |
|
"step": 279500 |
|
}, |
|
{ |
|
"epoch": 20.88, |
|
"learning_rate": 8.256226696495154e-06, |
|
"loss": 0.7035, |
|
"step": 280000 |
|
}, |
|
{ |
|
"epoch": 20.92, |
|
"learning_rate": 8.18165548098434e-06, |
|
"loss": 0.7097, |
|
"step": 280500 |
|
}, |
|
{ |
|
"epoch": 20.95, |
|
"learning_rate": 8.107084265473528e-06, |
|
"loss": 0.706, |
|
"step": 281000 |
|
}, |
|
{ |
|
"epoch": 20.99, |
|
"learning_rate": 8.032662192393737e-06, |
|
"loss": 0.7003, |
|
"step": 281500 |
|
}, |
|
{ |
|
"epoch": 21.03, |
|
"learning_rate": 7.958090976882924e-06, |
|
"loss": 0.7037, |
|
"step": 282000 |
|
}, |
|
{ |
|
"epoch": 21.07, |
|
"learning_rate": 7.883668903803131e-06, |
|
"loss": 0.7066, |
|
"step": 282500 |
|
}, |
|
{ |
|
"epoch": 21.1, |
|
"learning_rate": 7.809097688292318e-06, |
|
"loss": 0.7031, |
|
"step": 283000 |
|
}, |
|
{ |
|
"epoch": 21.14, |
|
"learning_rate": 7.734526472781507e-06, |
|
"loss": 0.6978, |
|
"step": 283500 |
|
}, |
|
{ |
|
"epoch": 21.18, |
|
"learning_rate": 7.659955257270694e-06, |
|
"loss": 0.7041, |
|
"step": 284000 |
|
}, |
|
{ |
|
"epoch": 21.22, |
|
"learning_rate": 7.585533184190903e-06, |
|
"loss": 0.7044, |
|
"step": 284500 |
|
}, |
|
{ |
|
"epoch": 21.25, |
|
"learning_rate": 7.51096196868009e-06, |
|
"loss": 0.6957, |
|
"step": 285000 |
|
}, |
|
{ |
|
"epoch": 21.29, |
|
"learning_rate": 7.436390753169277e-06, |
|
"loss": 0.702, |
|
"step": 285500 |
|
}, |
|
{ |
|
"epoch": 21.33, |
|
"learning_rate": 7.3619686800894855e-06, |
|
"loss": 0.7002, |
|
"step": 286000 |
|
}, |
|
{ |
|
"epoch": 21.36, |
|
"learning_rate": 7.287397464578673e-06, |
|
"loss": 0.7031, |
|
"step": 286500 |
|
}, |
|
{ |
|
"epoch": 21.4, |
|
"learning_rate": 7.21282624906786e-06, |
|
"loss": 0.7073, |
|
"step": 287000 |
|
}, |
|
{ |
|
"epoch": 21.44, |
|
"learning_rate": 7.138255033557047e-06, |
|
"loss": 0.706, |
|
"step": 287500 |
|
}, |
|
{ |
|
"epoch": 21.48, |
|
"learning_rate": 7.063683818046235e-06, |
|
"loss": 0.7052, |
|
"step": 288000 |
|
}, |
|
{ |
|
"epoch": 21.51, |
|
"learning_rate": 6.989112602535422e-06, |
|
"loss": 0.7019, |
|
"step": 288500 |
|
}, |
|
{ |
|
"epoch": 21.55, |
|
"learning_rate": 6.914690529455631e-06, |
|
"loss": 0.6996, |
|
"step": 289000 |
|
}, |
|
{ |
|
"epoch": 21.59, |
|
"learning_rate": 6.840119313944818e-06, |
|
"loss": 0.7, |
|
"step": 289500 |
|
}, |
|
{ |
|
"epoch": 21.63, |
|
"learning_rate": 6.765548098434005e-06, |
|
"loss": 0.7037, |
|
"step": 290000 |
|
}, |
|
{ |
|
"epoch": 21.66, |
|
"learning_rate": 6.690976882923192e-06, |
|
"loss": 0.7007, |
|
"step": 290500 |
|
}, |
|
{ |
|
"epoch": 21.7, |
|
"learning_rate": 6.616405667412379e-06, |
|
"loss": 0.7031, |
|
"step": 291000 |
|
}, |
|
{ |
|
"epoch": 21.74, |
|
"learning_rate": 6.541834451901565e-06, |
|
"loss": 0.7011, |
|
"step": 291500 |
|
}, |
|
{ |
|
"epoch": 21.77, |
|
"learning_rate": 6.467263236390754e-06, |
|
"loss": 0.7008, |
|
"step": 292000 |
|
}, |
|
{ |
|
"epoch": 21.81, |
|
"learning_rate": 6.392692020879941e-06, |
|
"loss": 0.696, |
|
"step": 292500 |
|
}, |
|
{ |
|
"epoch": 21.85, |
|
"learning_rate": 6.31826994780015e-06, |
|
"loss": 0.6973, |
|
"step": 293000 |
|
}, |
|
{ |
|
"epoch": 21.89, |
|
"learning_rate": 6.243698732289336e-06, |
|
"loss": 0.6995, |
|
"step": 293500 |
|
}, |
|
{ |
|
"epoch": 21.92, |
|
"learning_rate": 6.1691275167785235e-06, |
|
"loss": 0.6985, |
|
"step": 294000 |
|
}, |
|
{ |
|
"epoch": 21.96, |
|
"learning_rate": 6.0945563012677105e-06, |
|
"loss": 0.703, |
|
"step": 294500 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"learning_rate": 6.0199850857568984e-06, |
|
"loss": 0.6983, |
|
"step": 295000 |
|
}, |
|
{ |
|
"epoch": 22.04, |
|
"learning_rate": 5.9454138702460855e-06, |
|
"loss": 0.6993, |
|
"step": 295500 |
|
}, |
|
{ |
|
"epoch": 22.07, |
|
"learning_rate": 5.8708426547352726e-06, |
|
"loss": 0.6923, |
|
"step": 296000 |
|
}, |
|
{ |
|
"epoch": 22.11, |
|
"learning_rate": 5.79627143922446e-06, |
|
"loss": 0.6993, |
|
"step": 296500 |
|
}, |
|
{ |
|
"epoch": 22.15, |
|
"learning_rate": 5.721700223713647e-06, |
|
"loss": 0.6927, |
|
"step": 297000 |
|
}, |
|
{ |
|
"epoch": 22.18, |
|
"learning_rate": 5.647129008202834e-06, |
|
"loss": 0.6942, |
|
"step": 297500 |
|
}, |
|
{ |
|
"epoch": 22.22, |
|
"learning_rate": 5.572706935123043e-06, |
|
"loss": 0.6976, |
|
"step": 298000 |
|
}, |
|
{ |
|
"epoch": 22.26, |
|
"learning_rate": 5.49813571961223e-06, |
|
"loss": 0.6943, |
|
"step": 298500 |
|
}, |
|
{ |
|
"epoch": 22.3, |
|
"learning_rate": 5.423564504101417e-06, |
|
"loss": 0.7016, |
|
"step": 299000 |
|
}, |
|
{ |
|
"epoch": 22.33, |
|
"learning_rate": 5.348993288590604e-06, |
|
"loss": 0.6978, |
|
"step": 299500 |
|
}, |
|
{ |
|
"epoch": 22.37, |
|
"learning_rate": 5.274422073079791e-06, |
|
"loss": 0.6952, |
|
"step": 300000 |
|
}, |
|
{ |
|
"epoch": 22.41, |
|
"learning_rate": 5.199850857568979e-06, |
|
"loss": 0.6971, |
|
"step": 300500 |
|
}, |
|
{ |
|
"epoch": 22.45, |
|
"learning_rate": 5.125279642058166e-06, |
|
"loss": 0.6997, |
|
"step": 301000 |
|
}, |
|
{ |
|
"epoch": 22.48, |
|
"learning_rate": 5.050708426547353e-06, |
|
"loss": 0.6925, |
|
"step": 301500 |
|
}, |
|
{ |
|
"epoch": 22.52, |
|
"learning_rate": 4.976137211036539e-06, |
|
"loss": 0.692, |
|
"step": 302000 |
|
}, |
|
{ |
|
"epoch": 22.56, |
|
"learning_rate": 4.901715137956749e-06, |
|
"loss": 0.7017, |
|
"step": 302500 |
|
}, |
|
{ |
|
"epoch": 22.59, |
|
"learning_rate": 4.827143922445936e-06, |
|
"loss": 0.6978, |
|
"step": 303000 |
|
}, |
|
{ |
|
"epoch": 22.63, |
|
"learning_rate": 4.7525727069351234e-06, |
|
"loss": 0.698, |
|
"step": 303500 |
|
}, |
|
{ |
|
"epoch": 22.67, |
|
"learning_rate": 4.6780014914243105e-06, |
|
"loss": 0.6987, |
|
"step": 304000 |
|
}, |
|
{ |
|
"epoch": 22.71, |
|
"learning_rate": 4.603430275913498e-06, |
|
"loss": 0.6922, |
|
"step": 304500 |
|
}, |
|
{ |
|
"epoch": 22.74, |
|
"learning_rate": 4.529008202833707e-06, |
|
"loss": 0.693, |
|
"step": 305000 |
|
}, |
|
{ |
|
"epoch": 22.78, |
|
"learning_rate": 4.454436987322894e-06, |
|
"loss": 0.6959, |
|
"step": 305500 |
|
}, |
|
{ |
|
"epoch": 22.82, |
|
"learning_rate": 4.379865771812081e-06, |
|
"loss": 0.6993, |
|
"step": 306000 |
|
}, |
|
{ |
|
"epoch": 22.86, |
|
"learning_rate": 4.305294556301268e-06, |
|
"loss": 0.6949, |
|
"step": 306500 |
|
}, |
|
{ |
|
"epoch": 22.89, |
|
"learning_rate": 4.230872483221477e-06, |
|
"loss": 0.6936, |
|
"step": 307000 |
|
}, |
|
{ |
|
"epoch": 22.93, |
|
"learning_rate": 4.156301267710664e-06, |
|
"loss": 0.6956, |
|
"step": 307500 |
|
}, |
|
{ |
|
"epoch": 22.97, |
|
"learning_rate": 4.081730052199851e-06, |
|
"loss": 0.6929, |
|
"step": 308000 |
|
}, |
|
{ |
|
"epoch": 23.01, |
|
"learning_rate": 4.007158836689038e-06, |
|
"loss": 0.6897, |
|
"step": 308500 |
|
}, |
|
{ |
|
"epoch": 23.04, |
|
"learning_rate": 3.932736763609246e-06, |
|
"loss": 0.6975, |
|
"step": 309000 |
|
}, |
|
{ |
|
"epoch": 23.08, |
|
"learning_rate": 3.858165548098434e-06, |
|
"loss": 0.6971, |
|
"step": 309500 |
|
}, |
|
{ |
|
"epoch": 23.12, |
|
"learning_rate": 3.7835943325876213e-06, |
|
"loss": 0.6965, |
|
"step": 310000 |
|
}, |
|
{ |
|
"epoch": 23.15, |
|
"learning_rate": 3.70917225950783e-06, |
|
"loss": 0.697, |
|
"step": 310500 |
|
}, |
|
{ |
|
"epoch": 23.19, |
|
"learning_rate": 3.634601043997017e-06, |
|
"loss": 0.6885, |
|
"step": 311000 |
|
}, |
|
{ |
|
"epoch": 23.23, |
|
"learning_rate": 3.560029828486205e-06, |
|
"loss": 0.6854, |
|
"step": 311500 |
|
}, |
|
{ |
|
"epoch": 23.27, |
|
"learning_rate": 3.4854586129753916e-06, |
|
"loss": 0.6938, |
|
"step": 312000 |
|
}, |
|
{ |
|
"epoch": 23.3, |
|
"learning_rate": 3.4108873974645787e-06, |
|
"loss": 0.6905, |
|
"step": 312500 |
|
}, |
|
{ |
|
"epoch": 23.34, |
|
"learning_rate": 3.3364653243847873e-06, |
|
"loss": 0.6927, |
|
"step": 313000 |
|
}, |
|
{ |
|
"epoch": 23.38, |
|
"learning_rate": 3.2618941088739744e-06, |
|
"loss": 0.693, |
|
"step": 313500 |
|
}, |
|
{ |
|
"epoch": 23.42, |
|
"learning_rate": 3.1873228933631623e-06, |
|
"loss": 0.6894, |
|
"step": 314000 |
|
}, |
|
{ |
|
"epoch": 23.45, |
|
"learning_rate": 3.112751677852349e-06, |
|
"loss": 0.6899, |
|
"step": 314500 |
|
}, |
|
{ |
|
"epoch": 23.49, |
|
"learning_rate": 3.0381804623415364e-06, |
|
"loss": 0.6924, |
|
"step": 315000 |
|
}, |
|
{ |
|
"epoch": 23.53, |
|
"learning_rate": 2.9636092468307235e-06, |
|
"loss": 0.6936, |
|
"step": 315500 |
|
}, |
|
{ |
|
"epoch": 23.56, |
|
"learning_rate": 2.8890380313199106e-06, |
|
"loss": 0.6939, |
|
"step": 316000 |
|
}, |
|
{ |
|
"epoch": 23.6, |
|
"learning_rate": 2.8144668158090976e-06, |
|
"loss": 0.6975, |
|
"step": 316500 |
|
}, |
|
{ |
|
"epoch": 23.64, |
|
"learning_rate": 2.7400447427293063e-06, |
|
"loss": 0.6905, |
|
"step": 317000 |
|
}, |
|
{ |
|
"epoch": 23.68, |
|
"learning_rate": 2.6654735272184938e-06, |
|
"loss": 0.6908, |
|
"step": 317500 |
|
}, |
|
{ |
|
"epoch": 23.71, |
|
"learning_rate": 2.5909023117076813e-06, |
|
"loss": 0.6891, |
|
"step": 318000 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 2.516331096196868e-06, |
|
"loss": 0.6911, |
|
"step": 318500 |
|
}, |
|
{ |
|
"epoch": 23.79, |
|
"learning_rate": 2.4417598806860554e-06, |
|
"loss": 0.6956, |
|
"step": 319000 |
|
}, |
|
{ |
|
"epoch": 23.83, |
|
"learning_rate": 2.3671886651752425e-06, |
|
"loss": 0.6941, |
|
"step": 319500 |
|
}, |
|
{ |
|
"epoch": 23.86, |
|
"learning_rate": 2.292766592095451e-06, |
|
"loss": 0.6864, |
|
"step": 320000 |
|
}, |
|
{ |
|
"epoch": 23.9, |
|
"learning_rate": 2.2181953765846386e-06, |
|
"loss": 0.6959, |
|
"step": 320500 |
|
}, |
|
{ |
|
"epoch": 23.94, |
|
"learning_rate": 2.1436241610738253e-06, |
|
"loss": 0.689, |
|
"step": 321000 |
|
}, |
|
{ |
|
"epoch": 23.97, |
|
"learning_rate": 2.0690529455630127e-06, |
|
"loss": 0.6906, |
|
"step": 321500 |
|
}, |
|
{ |
|
"epoch": 24.01, |
|
"learning_rate": 1.9944817300522002e-06, |
|
"loss": 0.6921, |
|
"step": 322000 |
|
}, |
|
{ |
|
"epoch": 24.05, |
|
"learning_rate": 1.919910514541387e-06, |
|
"loss": 0.6915, |
|
"step": 322500 |
|
}, |
|
{ |
|
"epoch": 24.09, |
|
"learning_rate": 1.8454884414615957e-06, |
|
"loss": 0.6957, |
|
"step": 323000 |
|
}, |
|
{ |
|
"epoch": 24.12, |
|
"learning_rate": 1.7709172259507832e-06, |
|
"loss": 0.6892, |
|
"step": 323500 |
|
}, |
|
{ |
|
"epoch": 24.16, |
|
"learning_rate": 1.69634601043997e-06, |
|
"loss": 0.6875, |
|
"step": 324000 |
|
}, |
|
{ |
|
"epoch": 24.2, |
|
"learning_rate": 1.6217747949291574e-06, |
|
"loss": 0.6899, |
|
"step": 324500 |
|
}, |
|
{ |
|
"epoch": 24.24, |
|
"learning_rate": 1.5472035794183446e-06, |
|
"loss": 0.6888, |
|
"step": 325000 |
|
}, |
|
{ |
|
"epoch": 24.27, |
|
"learning_rate": 1.4726323639075317e-06, |
|
"loss": 0.6921, |
|
"step": 325500 |
|
}, |
|
{ |
|
"epoch": 24.31, |
|
"learning_rate": 1.3980611483967188e-06, |
|
"loss": 0.6866, |
|
"step": 326000 |
|
}, |
|
{ |
|
"epoch": 24.35, |
|
"learning_rate": 1.3234899328859063e-06, |
|
"loss": 0.6873, |
|
"step": 326500 |
|
}, |
|
{ |
|
"epoch": 24.38, |
|
"learning_rate": 1.249067859806115e-06, |
|
"loss": 0.6859, |
|
"step": 327000 |
|
}, |
|
{ |
|
"epoch": 24.42, |
|
"learning_rate": 1.1744966442953022e-06, |
|
"loss": 0.6825, |
|
"step": 327500 |
|
}, |
|
{ |
|
"epoch": 24.46, |
|
"learning_rate": 1.0999254287844893e-06, |
|
"loss": 0.6904, |
|
"step": 328000 |
|
}, |
|
{ |
|
"epoch": 24.5, |
|
"learning_rate": 1.0253542132736763e-06, |
|
"loss": 0.6941, |
|
"step": 328500 |
|
}, |
|
{ |
|
"epoch": 24.53, |
|
"learning_rate": 9.507829977628635e-07, |
|
"loss": 0.69, |
|
"step": 329000 |
|
}, |
|
{ |
|
"epoch": 24.57, |
|
"learning_rate": 8.762117822520507e-07, |
|
"loss": 0.6842, |
|
"step": 329500 |
|
}, |
|
{ |
|
"epoch": 24.61, |
|
"learning_rate": 8.01640566741238e-07, |
|
"loss": 0.6931, |
|
"step": 330000 |
|
}, |
|
{ |
|
"epoch": 24.65, |
|
"learning_rate": 7.27069351230425e-07, |
|
"loss": 0.6869, |
|
"step": 330500 |
|
}, |
|
{ |
|
"epoch": 24.68, |
|
"learning_rate": 6.527964205816555e-07, |
|
"loss": 0.6927, |
|
"step": 331000 |
|
}, |
|
{ |
|
"epoch": 24.72, |
|
"learning_rate": 5.782252050708427e-07, |
|
"loss": 0.6846, |
|
"step": 331500 |
|
}, |
|
{ |
|
"epoch": 24.76, |
|
"learning_rate": 5.036539895600298e-07, |
|
"loss": 0.6932, |
|
"step": 332000 |
|
}, |
|
{ |
|
"epoch": 24.79, |
|
"learning_rate": 4.29082774049217e-07, |
|
"loss": 0.6867, |
|
"step": 332500 |
|
}, |
|
{ |
|
"epoch": 24.83, |
|
"learning_rate": 3.545115585384042e-07, |
|
"loss": 0.6883, |
|
"step": 333000 |
|
}, |
|
{ |
|
"epoch": 24.87, |
|
"learning_rate": 2.799403430275914e-07, |
|
"loss": 0.6826, |
|
"step": 333500 |
|
}, |
|
{ |
|
"epoch": 24.91, |
|
"learning_rate": 2.053691275167785e-07, |
|
"loss": 0.6882, |
|
"step": 334000 |
|
}, |
|
{ |
|
"epoch": 24.94, |
|
"learning_rate": 1.3109619686800894e-07, |
|
"loss": 0.6854, |
|
"step": 334500 |
|
}, |
|
{ |
|
"epoch": 24.98, |
|
"learning_rate": 5.652498135719612e-08, |
|
"loss": 0.6878, |
|
"step": 335000 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"step": 335250, |
|
"total_flos": 2.835875553205027e+18, |
|
"train_loss": 0.891669009442653, |
|
"train_runtime": 467828.9629, |
|
"train_samples_per_second": 22.932, |
|
"train_steps_per_second": 0.717 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 335250, |
|
"num_train_epochs": 25, |
|
"save_steps": 10000, |
|
"total_flos": 2.835875553205027e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|