|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 7473, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0026763013515321826, |
|
"grad_norm": 21.75, |
|
"learning_rate": 9.973236986484679e-06, |
|
"loss": 1.9581, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.005352602703064365, |
|
"grad_norm": 11.0625, |
|
"learning_rate": 9.946473972969357e-06, |
|
"loss": 1.8767, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.008028904054596548, |
|
"grad_norm": 20.875, |
|
"learning_rate": 9.919710959454035e-06, |
|
"loss": 1.729, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01070520540612873, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 9.892947945938713e-06, |
|
"loss": 1.4764, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.013381506757660913, |
|
"grad_norm": 20.375, |
|
"learning_rate": 9.86618493242339e-06, |
|
"loss": 1.4558, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.016057808109193095, |
|
"grad_norm": 9.8125, |
|
"learning_rate": 9.83942191890807e-06, |
|
"loss": 1.1101, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.018734109460725276, |
|
"grad_norm": 14.8125, |
|
"learning_rate": 9.812658905392748e-06, |
|
"loss": 1.2423, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02141041081225746, |
|
"grad_norm": 15.3125, |
|
"learning_rate": 9.785895891877426e-06, |
|
"loss": 1.3871, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.02408671216378964, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 9.759132878362104e-06, |
|
"loss": 1.2922, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.026763013515321826, |
|
"grad_norm": 9.25, |
|
"learning_rate": 9.732369864846782e-06, |
|
"loss": 1.3144, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.029439314866854006, |
|
"grad_norm": 18.625, |
|
"learning_rate": 9.70560685133146e-06, |
|
"loss": 1.0756, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.03211561621838619, |
|
"grad_norm": 43.25, |
|
"learning_rate": 9.678843837816138e-06, |
|
"loss": 1.2612, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.034791917569918375, |
|
"grad_norm": 9.4375, |
|
"learning_rate": 9.652080824300816e-06, |
|
"loss": 1.0856, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.03746821892145055, |
|
"grad_norm": 17.5, |
|
"learning_rate": 9.625317810785494e-06, |
|
"loss": 1.0375, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.04014452027298274, |
|
"grad_norm": 7.75, |
|
"learning_rate": 9.598554797270172e-06, |
|
"loss": 0.9578, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.04282082162451492, |
|
"grad_norm": 12.875, |
|
"learning_rate": 9.571791783754852e-06, |
|
"loss": 0.9921, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.045497122976047105, |
|
"grad_norm": 14.4375, |
|
"learning_rate": 9.54502877023953e-06, |
|
"loss": 0.8984, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.04817342432757928, |
|
"grad_norm": 11.9375, |
|
"learning_rate": 9.518265756724208e-06, |
|
"loss": 1.0139, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.05084972567911147, |
|
"grad_norm": 8.1875, |
|
"learning_rate": 9.491502743208886e-06, |
|
"loss": 0.9266, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.05352602703064365, |
|
"grad_norm": 33.0, |
|
"learning_rate": 9.464739729693564e-06, |
|
"loss": 0.8922, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.056202328382175835, |
|
"grad_norm": 20.375, |
|
"learning_rate": 9.437976716178242e-06, |
|
"loss": 0.9903, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.05887862973370801, |
|
"grad_norm": 18.625, |
|
"learning_rate": 9.41121370266292e-06, |
|
"loss": 0.8287, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.0615549310852402, |
|
"grad_norm": 8.75, |
|
"learning_rate": 9.384450689147598e-06, |
|
"loss": 0.9283, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.06423123243677238, |
|
"grad_norm": 21.75, |
|
"learning_rate": 9.357687675632276e-06, |
|
"loss": 0.9585, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.06690753378830457, |
|
"grad_norm": 13.9375, |
|
"learning_rate": 9.330924662116954e-06, |
|
"loss": 0.8412, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.06958383513983675, |
|
"grad_norm": 25.0, |
|
"learning_rate": 9.304161648601634e-06, |
|
"loss": 0.9743, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.07226013649136893, |
|
"grad_norm": 18.625, |
|
"learning_rate": 9.277398635086312e-06, |
|
"loss": 0.869, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.0749364378429011, |
|
"grad_norm": 11.625, |
|
"learning_rate": 9.25063562157099e-06, |
|
"loss": 0.7945, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.07761273919443329, |
|
"grad_norm": 7.1875, |
|
"learning_rate": 9.223872608055667e-06, |
|
"loss": 0.8359, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.08028904054596547, |
|
"grad_norm": 14.3125, |
|
"learning_rate": 9.197109594540345e-06, |
|
"loss": 0.7731, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.08296534189749766, |
|
"grad_norm": 12.9375, |
|
"learning_rate": 9.170346581025023e-06, |
|
"loss": 0.8352, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.08564164324902984, |
|
"grad_norm": 19.75, |
|
"learning_rate": 9.143583567509703e-06, |
|
"loss": 0.8792, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.08831794460056203, |
|
"grad_norm": 8.25, |
|
"learning_rate": 9.116820553994381e-06, |
|
"loss": 0.8651, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.09099424595209421, |
|
"grad_norm": 12.125, |
|
"learning_rate": 9.090057540479059e-06, |
|
"loss": 0.812, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.0936705473036264, |
|
"grad_norm": 15.75, |
|
"learning_rate": 9.063294526963737e-06, |
|
"loss": 0.8949, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.09634684865515857, |
|
"grad_norm": 11.5625, |
|
"learning_rate": 9.036531513448415e-06, |
|
"loss": 0.8752, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.09902315000669075, |
|
"grad_norm": 14.6875, |
|
"learning_rate": 9.009768499933093e-06, |
|
"loss": 0.8544, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.10169945135822293, |
|
"grad_norm": 9.9375, |
|
"learning_rate": 8.983005486417771e-06, |
|
"loss": 0.8414, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.10437575270975512, |
|
"grad_norm": 11.5, |
|
"learning_rate": 8.95624247290245e-06, |
|
"loss": 0.8796, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.1070520540612873, |
|
"grad_norm": 13.0625, |
|
"learning_rate": 8.929479459387129e-06, |
|
"loss": 0.853, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.10972835541281949, |
|
"grad_norm": 12.1875, |
|
"learning_rate": 8.902716445871807e-06, |
|
"loss": 0.8864, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.11240465676435167, |
|
"grad_norm": 60.0, |
|
"learning_rate": 8.875953432356485e-06, |
|
"loss": 0.7659, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.11508095811588386, |
|
"grad_norm": 30.0, |
|
"learning_rate": 8.849190418841163e-06, |
|
"loss": 0.9269, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.11775725946741603, |
|
"grad_norm": 9.125, |
|
"learning_rate": 8.82242740532584e-06, |
|
"loss": 0.7939, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.12043356081894821, |
|
"grad_norm": 9.0, |
|
"learning_rate": 8.795664391810519e-06, |
|
"loss": 0.8343, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.1231098621704804, |
|
"grad_norm": 18.25, |
|
"learning_rate": 8.768901378295197e-06, |
|
"loss": 0.7588, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.12578616352201258, |
|
"grad_norm": 6.34375, |
|
"learning_rate": 8.742138364779875e-06, |
|
"loss": 0.8424, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.12846246487354476, |
|
"grad_norm": 11.875, |
|
"learning_rate": 8.715375351264553e-06, |
|
"loss": 0.7666, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.13113876622507695, |
|
"grad_norm": 10.0, |
|
"learning_rate": 8.688612337749232e-06, |
|
"loss": 0.7019, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.13381506757660913, |
|
"grad_norm": 8.25, |
|
"learning_rate": 8.66184932423391e-06, |
|
"loss": 0.7036, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.13649136892814132, |
|
"grad_norm": 10.75, |
|
"learning_rate": 8.635086310718588e-06, |
|
"loss": 0.7355, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.1391676702796735, |
|
"grad_norm": 14.3125, |
|
"learning_rate": 8.608323297203266e-06, |
|
"loss": 0.737, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.14184397163120568, |
|
"grad_norm": 9.5, |
|
"learning_rate": 8.581560283687944e-06, |
|
"loss": 0.7999, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.14452027298273787, |
|
"grad_norm": 10.125, |
|
"learning_rate": 8.554797270172622e-06, |
|
"loss": 0.7613, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.14719657433427003, |
|
"grad_norm": 22.625, |
|
"learning_rate": 8.5280342566573e-06, |
|
"loss": 0.8093, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.1498728756858022, |
|
"grad_norm": 45.0, |
|
"learning_rate": 8.501271243141978e-06, |
|
"loss": 0.8385, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.1525491770373344, |
|
"grad_norm": 12.5625, |
|
"learning_rate": 8.474508229626656e-06, |
|
"loss": 0.8014, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.15522547838886658, |
|
"grad_norm": 12.625, |
|
"learning_rate": 8.447745216111334e-06, |
|
"loss": 0.8474, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.15790177974039876, |
|
"grad_norm": 11.875, |
|
"learning_rate": 8.420982202596014e-06, |
|
"loss": 0.8009, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.16057808109193095, |
|
"grad_norm": 20.625, |
|
"learning_rate": 8.394219189080692e-06, |
|
"loss": 0.8726, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.16325438244346313, |
|
"grad_norm": 11.9375, |
|
"learning_rate": 8.36745617556537e-06, |
|
"loss": 0.751, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.16593068379499532, |
|
"grad_norm": 8.4375, |
|
"learning_rate": 8.340693162050048e-06, |
|
"loss": 0.729, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.1686069851465275, |
|
"grad_norm": 14.375, |
|
"learning_rate": 8.313930148534726e-06, |
|
"loss": 0.7642, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.17128328649805968, |
|
"grad_norm": 9.4375, |
|
"learning_rate": 8.287167135019404e-06, |
|
"loss": 0.799, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.17395958784959187, |
|
"grad_norm": 17.375, |
|
"learning_rate": 8.260404121504082e-06, |
|
"loss": 0.7899, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.17663588920112405, |
|
"grad_norm": 8.25, |
|
"learning_rate": 8.23364110798876e-06, |
|
"loss": 0.8224, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.17931219055265624, |
|
"grad_norm": 17.25, |
|
"learning_rate": 8.206878094473438e-06, |
|
"loss": 0.7836, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.18198849190418842, |
|
"grad_norm": 11.9375, |
|
"learning_rate": 8.180115080958116e-06, |
|
"loss": 0.8547, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.1846647932557206, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 8.153352067442795e-06, |
|
"loss": 0.8314, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.1873410946072528, |
|
"grad_norm": 12.25, |
|
"learning_rate": 8.126589053927473e-06, |
|
"loss": 0.7372, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.19001739595878495, |
|
"grad_norm": 17.625, |
|
"learning_rate": 8.099826040412151e-06, |
|
"loss": 0.7241, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.19269369731031713, |
|
"grad_norm": 13.25, |
|
"learning_rate": 8.07306302689683e-06, |
|
"loss": 0.75, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.19536999866184931, |
|
"grad_norm": 9.3125, |
|
"learning_rate": 8.046300013381507e-06, |
|
"loss": 0.7802, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.1980463000133815, |
|
"grad_norm": 9.1875, |
|
"learning_rate": 8.019536999866185e-06, |
|
"loss": 0.8704, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.20072260136491368, |
|
"grad_norm": 6.15625, |
|
"learning_rate": 7.992773986350863e-06, |
|
"loss": 0.7438, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.20339890271644587, |
|
"grad_norm": 9.125, |
|
"learning_rate": 7.966010972835541e-06, |
|
"loss": 0.7446, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.20607520406797805, |
|
"grad_norm": 21.25, |
|
"learning_rate": 7.93924795932022e-06, |
|
"loss": 0.6814, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.20875150541951024, |
|
"grad_norm": 6.65625, |
|
"learning_rate": 7.912484945804897e-06, |
|
"loss": 0.7361, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.21142780677104242, |
|
"grad_norm": 13.125, |
|
"learning_rate": 7.885721932289577e-06, |
|
"loss": 0.67, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.2141041081225746, |
|
"grad_norm": 15.6875, |
|
"learning_rate": 7.858958918774255e-06, |
|
"loss": 0.7145, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.2167804094741068, |
|
"grad_norm": 14.1875, |
|
"learning_rate": 7.832195905258933e-06, |
|
"loss": 0.818, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.21945671082563897, |
|
"grad_norm": 6.53125, |
|
"learning_rate": 7.805432891743611e-06, |
|
"loss": 0.7837, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.22213301217717116, |
|
"grad_norm": 13.5625, |
|
"learning_rate": 7.778669878228289e-06, |
|
"loss": 0.7751, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.22480931352870334, |
|
"grad_norm": 15.375, |
|
"learning_rate": 7.751906864712967e-06, |
|
"loss": 0.8476, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.22748561488023553, |
|
"grad_norm": 7.625, |
|
"learning_rate": 7.725143851197645e-06, |
|
"loss": 0.7734, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.2301619162317677, |
|
"grad_norm": 14.25, |
|
"learning_rate": 7.698380837682323e-06, |
|
"loss": 0.7714, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.23283821758329987, |
|
"grad_norm": 18.5, |
|
"learning_rate": 7.671617824167e-06, |
|
"loss": 0.7447, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.23551451893483205, |
|
"grad_norm": 15.75, |
|
"learning_rate": 7.644854810651679e-06, |
|
"loss": 0.7385, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.23819082028636424, |
|
"grad_norm": 10.5, |
|
"learning_rate": 7.618091797136358e-06, |
|
"loss": 0.7058, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.24086712163789642, |
|
"grad_norm": 38.75, |
|
"learning_rate": 7.591328783621036e-06, |
|
"loss": 0.737, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.2435434229894286, |
|
"grad_norm": 10.5, |
|
"learning_rate": 7.564565770105714e-06, |
|
"loss": 0.7831, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.2462197243409608, |
|
"grad_norm": 21.0, |
|
"learning_rate": 7.5378027565903925e-06, |
|
"loss": 0.8369, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.24889602569249297, |
|
"grad_norm": 20.5, |
|
"learning_rate": 7.5110397430750704e-06, |
|
"loss": 0.7846, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.25157232704402516, |
|
"grad_norm": 6.375, |
|
"learning_rate": 7.484276729559748e-06, |
|
"loss": 0.8011, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.25424862839555734, |
|
"grad_norm": 9.1875, |
|
"learning_rate": 7.457513716044428e-06, |
|
"loss": 0.7536, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.2569249297470895, |
|
"grad_norm": 9.0, |
|
"learning_rate": 7.430750702529106e-06, |
|
"loss": 0.7622, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.2596012310986217, |
|
"grad_norm": 7.875, |
|
"learning_rate": 7.403987689013784e-06, |
|
"loss": 0.7059, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.2622775324501539, |
|
"grad_norm": 7.46875, |
|
"learning_rate": 7.377224675498462e-06, |
|
"loss": 0.7635, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.2649538338016861, |
|
"grad_norm": 6.21875, |
|
"learning_rate": 7.35046166198314e-06, |
|
"loss": 0.7693, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.26763013515321826, |
|
"grad_norm": 9.375, |
|
"learning_rate": 7.323698648467819e-06, |
|
"loss": 0.7812, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.27030643650475045, |
|
"grad_norm": 13.1875, |
|
"learning_rate": 7.296935634952497e-06, |
|
"loss": 0.7372, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.27298273785628263, |
|
"grad_norm": 16.75, |
|
"learning_rate": 7.270172621437175e-06, |
|
"loss": 0.8489, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.2756590392078148, |
|
"grad_norm": 5.40625, |
|
"learning_rate": 7.243409607921853e-06, |
|
"loss": 0.6996, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.278335340559347, |
|
"grad_norm": 10.875, |
|
"learning_rate": 7.216646594406531e-06, |
|
"loss": 0.7596, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.2810116419108792, |
|
"grad_norm": 33.5, |
|
"learning_rate": 7.18988358089121e-06, |
|
"loss": 0.7819, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.28368794326241137, |
|
"grad_norm": 11.4375, |
|
"learning_rate": 7.163120567375888e-06, |
|
"loss": 0.8196, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.28636424461394355, |
|
"grad_norm": 7.78125, |
|
"learning_rate": 7.136357553860566e-06, |
|
"loss": 0.7454, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.28904054596547574, |
|
"grad_norm": 14.5625, |
|
"learning_rate": 7.109594540345244e-06, |
|
"loss": 0.7306, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.2917168473170079, |
|
"grad_norm": 9.5, |
|
"learning_rate": 7.082831526829922e-06, |
|
"loss": 0.7076, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.29439314866854005, |
|
"grad_norm": 12.875, |
|
"learning_rate": 7.0560685133146004e-06, |
|
"loss": 0.7679, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.29706945002007223, |
|
"grad_norm": 11.0, |
|
"learning_rate": 7.029305499799278e-06, |
|
"loss": 0.6691, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.2997457513716044, |
|
"grad_norm": 13.4375, |
|
"learning_rate": 7.002542486283956e-06, |
|
"loss": 0.7002, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.3024220527231366, |
|
"grad_norm": 13.25, |
|
"learning_rate": 6.975779472768634e-06, |
|
"loss": 0.8657, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.3050983540746688, |
|
"grad_norm": 16.25, |
|
"learning_rate": 6.949016459253312e-06, |
|
"loss": 0.6808, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.30777465542620097, |
|
"grad_norm": 11.75, |
|
"learning_rate": 6.922253445737991e-06, |
|
"loss": 0.7294, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.31045095677773316, |
|
"grad_norm": 12.5, |
|
"learning_rate": 6.895490432222669e-06, |
|
"loss": 0.7018, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.31312725812926534, |
|
"grad_norm": 11.6875, |
|
"learning_rate": 6.868727418707347e-06, |
|
"loss": 0.6812, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.3158035594807975, |
|
"grad_norm": 7.625, |
|
"learning_rate": 6.841964405192025e-06, |
|
"loss": 0.7062, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.3184798608323297, |
|
"grad_norm": 18.125, |
|
"learning_rate": 6.815201391676703e-06, |
|
"loss": 0.7192, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.3211561621838619, |
|
"grad_norm": 12.75, |
|
"learning_rate": 6.788438378161382e-06, |
|
"loss": 0.6406, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.3238324635353941, |
|
"grad_norm": 6.75, |
|
"learning_rate": 6.76167536464606e-06, |
|
"loss": 0.7128, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.32650876488692626, |
|
"grad_norm": 15.6875, |
|
"learning_rate": 6.734912351130738e-06, |
|
"loss": 0.6342, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.32918506623845845, |
|
"grad_norm": 14.0, |
|
"learning_rate": 6.708149337615416e-06, |
|
"loss": 0.7335, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.33186136758999063, |
|
"grad_norm": 13.0, |
|
"learning_rate": 6.681386324100094e-06, |
|
"loss": 0.629, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.3345376689415228, |
|
"grad_norm": 10.3125, |
|
"learning_rate": 6.654623310584773e-06, |
|
"loss": 0.7736, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.337213970293055, |
|
"grad_norm": 7.84375, |
|
"learning_rate": 6.627860297069451e-06, |
|
"loss": 0.826, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.3398902716445872, |
|
"grad_norm": 7.09375, |
|
"learning_rate": 6.601097283554129e-06, |
|
"loss": 0.6452, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.34256657299611937, |
|
"grad_norm": 6.125, |
|
"learning_rate": 6.574334270038807e-06, |
|
"loss": 0.705, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.34524287434765155, |
|
"grad_norm": 65.0, |
|
"learning_rate": 6.547571256523485e-06, |
|
"loss": 0.6926, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.34791917569918374, |
|
"grad_norm": 7.125, |
|
"learning_rate": 6.5208082430081635e-06, |
|
"loss": 0.6877, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.3505954770507159, |
|
"grad_norm": 12.4375, |
|
"learning_rate": 6.4940452294928415e-06, |
|
"loss": 0.8122, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.3532717784022481, |
|
"grad_norm": 5.75, |
|
"learning_rate": 6.4672822159775195e-06, |
|
"loss": 0.7015, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.3559480797537803, |
|
"grad_norm": 9.1875, |
|
"learning_rate": 6.4405192024621975e-06, |
|
"loss": 0.7502, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.3586243811053125, |
|
"grad_norm": 4.6875, |
|
"learning_rate": 6.4137561889468755e-06, |
|
"loss": 0.6114, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.36130068245684466, |
|
"grad_norm": 12.0, |
|
"learning_rate": 6.386993175431554e-06, |
|
"loss": 0.6149, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.36397698380837684, |
|
"grad_norm": 7.78125, |
|
"learning_rate": 6.360230161916232e-06, |
|
"loss": 0.6908, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.366653285159909, |
|
"grad_norm": 7.65625, |
|
"learning_rate": 6.33346714840091e-06, |
|
"loss": 0.7734, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.3693295865114412, |
|
"grad_norm": 12.0, |
|
"learning_rate": 6.306704134885588e-06, |
|
"loss": 0.7201, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.3720058878629734, |
|
"grad_norm": 17.875, |
|
"learning_rate": 6.279941121370266e-06, |
|
"loss": 0.7441, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.3746821892145056, |
|
"grad_norm": 18.375, |
|
"learning_rate": 6.253178107854945e-06, |
|
"loss": 0.6822, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.37735849056603776, |
|
"grad_norm": 10.125, |
|
"learning_rate": 6.226415094339623e-06, |
|
"loss": 0.758, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.3800347919175699, |
|
"grad_norm": 15.125, |
|
"learning_rate": 6.199652080824301e-06, |
|
"loss": 0.6623, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.3827110932691021, |
|
"grad_norm": 8.5, |
|
"learning_rate": 6.172889067308979e-06, |
|
"loss": 0.6744, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.38538739462063426, |
|
"grad_norm": 10.8125, |
|
"learning_rate": 6.146126053793657e-06, |
|
"loss": 0.6643, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.38806369597216644, |
|
"grad_norm": 11.5625, |
|
"learning_rate": 6.119363040278336e-06, |
|
"loss": 0.7278, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.39073999732369863, |
|
"grad_norm": 6.65625, |
|
"learning_rate": 6.092600026763014e-06, |
|
"loss": 0.6364, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.3934162986752308, |
|
"grad_norm": 8.75, |
|
"learning_rate": 6.065837013247692e-06, |
|
"loss": 0.7175, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.396092600026763, |
|
"grad_norm": 10.25, |
|
"learning_rate": 6.03907399973237e-06, |
|
"loss": 0.6905, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.3987689013782952, |
|
"grad_norm": 14.0625, |
|
"learning_rate": 6.012310986217048e-06, |
|
"loss": 0.599, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.40144520272982737, |
|
"grad_norm": 9.6875, |
|
"learning_rate": 5.985547972701727e-06, |
|
"loss": 0.7629, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.40412150408135955, |
|
"grad_norm": 19.125, |
|
"learning_rate": 5.958784959186405e-06, |
|
"loss": 0.7624, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.40679780543289173, |
|
"grad_norm": 9.0625, |
|
"learning_rate": 5.932021945671083e-06, |
|
"loss": 0.7122, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.4094741067844239, |
|
"grad_norm": 12.0625, |
|
"learning_rate": 5.905258932155761e-06, |
|
"loss": 0.711, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.4121504081359561, |
|
"grad_norm": 8.4375, |
|
"learning_rate": 5.878495918640439e-06, |
|
"loss": 0.685, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.4148267094874883, |
|
"grad_norm": 23.5, |
|
"learning_rate": 5.851732905125117e-06, |
|
"loss": 0.7422, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.41750301083902047, |
|
"grad_norm": 8.9375, |
|
"learning_rate": 5.824969891609795e-06, |
|
"loss": 0.7546, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.42017931219055266, |
|
"grad_norm": 10.375, |
|
"learning_rate": 5.798206878094474e-06, |
|
"loss": 0.7204, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.42285561354208484, |
|
"grad_norm": 7.3125, |
|
"learning_rate": 5.771443864579152e-06, |
|
"loss": 0.7457, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.425531914893617, |
|
"grad_norm": 10.9375, |
|
"learning_rate": 5.744680851063831e-06, |
|
"loss": 0.7411, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.4282082162451492, |
|
"grad_norm": 12.3125, |
|
"learning_rate": 5.717917837548509e-06, |
|
"loss": 0.7036, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.4308845175966814, |
|
"grad_norm": 10.0625, |
|
"learning_rate": 5.691154824033187e-06, |
|
"loss": 0.7014, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.4335608189482136, |
|
"grad_norm": 6.875, |
|
"learning_rate": 5.664391810517865e-06, |
|
"loss": 0.73, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.43623712029974576, |
|
"grad_norm": 8.375, |
|
"learning_rate": 5.637628797002543e-06, |
|
"loss": 0.6942, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.43891342165127795, |
|
"grad_norm": 13.125, |
|
"learning_rate": 5.610865783487222e-06, |
|
"loss": 0.6773, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.44158972300281013, |
|
"grad_norm": 10.5, |
|
"learning_rate": 5.5841027699719e-06, |
|
"loss": 0.6452, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.4442660243543423, |
|
"grad_norm": 7.59375, |
|
"learning_rate": 5.557339756456578e-06, |
|
"loss": 0.7561, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.4469423257058745, |
|
"grad_norm": 10.4375, |
|
"learning_rate": 5.530576742941256e-06, |
|
"loss": 0.7064, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.4496186270574067, |
|
"grad_norm": 6.90625, |
|
"learning_rate": 5.503813729425934e-06, |
|
"loss": 0.6614, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.45229492840893887, |
|
"grad_norm": 11.125, |
|
"learning_rate": 5.477050715910613e-06, |
|
"loss": 0.7513, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.45497122976047105, |
|
"grad_norm": 16.875, |
|
"learning_rate": 5.450287702395291e-06, |
|
"loss": 0.6785, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.45764753111200324, |
|
"grad_norm": 12.1875, |
|
"learning_rate": 5.4235246888799686e-06, |
|
"loss": 0.8056, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.4603238324635354, |
|
"grad_norm": 19.125, |
|
"learning_rate": 5.3967616753646466e-06, |
|
"loss": 0.7518, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.46300013381506755, |
|
"grad_norm": 9.8125, |
|
"learning_rate": 5.3699986618493245e-06, |
|
"loss": 0.6852, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.46567643516659973, |
|
"grad_norm": 7.59375, |
|
"learning_rate": 5.343235648334003e-06, |
|
"loss": 0.7327, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.4683527365181319, |
|
"grad_norm": 19.0, |
|
"learning_rate": 5.316472634818681e-06, |
|
"loss": 0.7463, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.4710290378696641, |
|
"grad_norm": 8.6875, |
|
"learning_rate": 5.289709621303359e-06, |
|
"loss": 0.6091, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.4737053392211963, |
|
"grad_norm": 14.0, |
|
"learning_rate": 5.262946607788037e-06, |
|
"loss": 0.6896, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.47638164057272847, |
|
"grad_norm": 8.5, |
|
"learning_rate": 5.236183594272715e-06, |
|
"loss": 0.7641, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.47905794192426066, |
|
"grad_norm": 12.3125, |
|
"learning_rate": 5.209420580757394e-06, |
|
"loss": 0.6582, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.48173424327579284, |
|
"grad_norm": 11.0625, |
|
"learning_rate": 5.182657567242072e-06, |
|
"loss": 0.7291, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.484410544627325, |
|
"grad_norm": 12.9375, |
|
"learning_rate": 5.15589455372675e-06, |
|
"loss": 0.7209, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.4870868459788572, |
|
"grad_norm": 14.625, |
|
"learning_rate": 5.129131540211428e-06, |
|
"loss": 0.7259, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.4897631473303894, |
|
"grad_norm": 9.4375, |
|
"learning_rate": 5.102368526696106e-06, |
|
"loss": 0.7259, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.4924394486819216, |
|
"grad_norm": 7.53125, |
|
"learning_rate": 5.075605513180785e-06, |
|
"loss": 0.6947, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.49511575003345376, |
|
"grad_norm": 6.59375, |
|
"learning_rate": 5.048842499665463e-06, |
|
"loss": 0.7026, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.49779205138498595, |
|
"grad_norm": 22.125, |
|
"learning_rate": 5.022079486150141e-06, |
|
"loss": 0.7332, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.5004683527365181, |
|
"grad_norm": 9.625, |
|
"learning_rate": 4.995316472634819e-06, |
|
"loss": 0.7487, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.5031446540880503, |
|
"grad_norm": 8.9375, |
|
"learning_rate": 4.968553459119497e-06, |
|
"loss": 0.6364, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.5058209554395825, |
|
"grad_norm": 10.125, |
|
"learning_rate": 4.941790445604176e-06, |
|
"loss": 0.7261, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.5084972567911147, |
|
"grad_norm": 19.0, |
|
"learning_rate": 4.915027432088854e-06, |
|
"loss": 0.725, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.5111735581426469, |
|
"grad_norm": 11.6875, |
|
"learning_rate": 4.888264418573532e-06, |
|
"loss": 0.6957, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.513849859494179, |
|
"grad_norm": 9.75, |
|
"learning_rate": 4.86150140505821e-06, |
|
"loss": 0.6555, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.5165261608457112, |
|
"grad_norm": 8.875, |
|
"learning_rate": 4.834738391542888e-06, |
|
"loss": 0.7902, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.5192024621972434, |
|
"grad_norm": 10.5625, |
|
"learning_rate": 4.8079753780275665e-06, |
|
"loss": 0.7535, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.5218787635487756, |
|
"grad_norm": 6.71875, |
|
"learning_rate": 4.7812123645122445e-06, |
|
"loss": 0.84, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.5245550649003078, |
|
"grad_norm": 8.3125, |
|
"learning_rate": 4.7544493509969225e-06, |
|
"loss": 0.6911, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.52723136625184, |
|
"grad_norm": 9.4375, |
|
"learning_rate": 4.7276863374816004e-06, |
|
"loss": 0.7435, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.5299076676033722, |
|
"grad_norm": 12.875, |
|
"learning_rate": 4.7009233239662784e-06, |
|
"loss": 0.7176, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.5325839689549043, |
|
"grad_norm": 6.6875, |
|
"learning_rate": 4.674160310450957e-06, |
|
"loss": 0.6389, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.5352602703064365, |
|
"grad_norm": 17.0, |
|
"learning_rate": 4.647397296935635e-06, |
|
"loss": 0.6902, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.5379365716579687, |
|
"grad_norm": 15.5, |
|
"learning_rate": 4.620634283420313e-06, |
|
"loss": 0.7835, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.5406128730095009, |
|
"grad_norm": 14.0625, |
|
"learning_rate": 4.593871269904991e-06, |
|
"loss": 0.7667, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.5432891743610331, |
|
"grad_norm": 15.5625, |
|
"learning_rate": 4.567108256389669e-06, |
|
"loss": 0.7173, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.5459654757125653, |
|
"grad_norm": 7.6875, |
|
"learning_rate": 4.540345242874348e-06, |
|
"loss": 0.6364, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.5486417770640974, |
|
"grad_norm": 8.0625, |
|
"learning_rate": 4.513582229359027e-06, |
|
"loss": 0.6644, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.5513180784156296, |
|
"grad_norm": 8.6875, |
|
"learning_rate": 4.486819215843705e-06, |
|
"loss": 0.6862, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.5539943797671618, |
|
"grad_norm": 16.875, |
|
"learning_rate": 4.460056202328383e-06, |
|
"loss": 0.6885, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.556670681118694, |
|
"grad_norm": 8.125, |
|
"learning_rate": 4.433293188813061e-06, |
|
"loss": 0.7585, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.5593469824702262, |
|
"grad_norm": 13.375, |
|
"learning_rate": 4.406530175297739e-06, |
|
"loss": 0.7361, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.5620232838217584, |
|
"grad_norm": 9.875, |
|
"learning_rate": 4.379767161782418e-06, |
|
"loss": 0.6836, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.5646995851732906, |
|
"grad_norm": 10.75, |
|
"learning_rate": 4.353004148267096e-06, |
|
"loss": 0.7238, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.5673758865248227, |
|
"grad_norm": 10.5625, |
|
"learning_rate": 4.326241134751774e-06, |
|
"loss": 0.7354, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.5700521878763549, |
|
"grad_norm": 8.125, |
|
"learning_rate": 4.299478121236452e-06, |
|
"loss": 0.7141, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.5727284892278871, |
|
"grad_norm": 9.9375, |
|
"learning_rate": 4.27271510772113e-06, |
|
"loss": 0.7113, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.5754047905794193, |
|
"grad_norm": 13.75, |
|
"learning_rate": 4.245952094205808e-06, |
|
"loss": 0.7338, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.5780810919309515, |
|
"grad_norm": 14.375, |
|
"learning_rate": 4.219189080690486e-06, |
|
"loss": 0.7292, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.5807573932824837, |
|
"grad_norm": 10.3125, |
|
"learning_rate": 4.192426067175164e-06, |
|
"loss": 0.6675, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.5834336946340158, |
|
"grad_norm": 27.375, |
|
"learning_rate": 4.165663053659842e-06, |
|
"loss": 0.6879, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.586109995985548, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 4.13890004014452e-06, |
|
"loss": 0.5935, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.5887862973370801, |
|
"grad_norm": 7.46875, |
|
"learning_rate": 4.112137026629199e-06, |
|
"loss": 0.7692, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.5914625986886123, |
|
"grad_norm": 13.875, |
|
"learning_rate": 4.085374013113877e-06, |
|
"loss": 0.7598, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.5941389000401445, |
|
"grad_norm": 14.6875, |
|
"learning_rate": 4.058610999598555e-06, |
|
"loss": 0.63, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.5968152013916767, |
|
"grad_norm": 17.75, |
|
"learning_rate": 4.031847986083233e-06, |
|
"loss": 0.7506, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.5994915027432088, |
|
"grad_norm": 9.625, |
|
"learning_rate": 4.005084972567911e-06, |
|
"loss": 0.6189, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.602167804094741, |
|
"grad_norm": 8.875, |
|
"learning_rate": 3.97832195905259e-06, |
|
"loss": 0.7193, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.6048441054462732, |
|
"grad_norm": 12.8125, |
|
"learning_rate": 3.951558945537268e-06, |
|
"loss": 0.7936, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.6075204067978054, |
|
"grad_norm": 8.3125, |
|
"learning_rate": 3.924795932021946e-06, |
|
"loss": 0.7412, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.6101967081493376, |
|
"grad_norm": 13.75, |
|
"learning_rate": 3.898032918506624e-06, |
|
"loss": 0.6437, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.6128730095008698, |
|
"grad_norm": 12.1875, |
|
"learning_rate": 3.871269904991302e-06, |
|
"loss": 0.6643, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.6155493108524019, |
|
"grad_norm": 14.5, |
|
"learning_rate": 3.844506891475981e-06, |
|
"loss": 0.6438, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.6182256122039341, |
|
"grad_norm": 8.6875, |
|
"learning_rate": 3.817743877960659e-06, |
|
"loss": 0.6731, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.6209019135554663, |
|
"grad_norm": 23.0, |
|
"learning_rate": 3.7909808644453367e-06, |
|
"loss": 0.6498, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.6235782149069985, |
|
"grad_norm": 7.3125, |
|
"learning_rate": 3.7642178509300147e-06, |
|
"loss": 0.7402, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.6262545162585307, |
|
"grad_norm": 16.625, |
|
"learning_rate": 3.737454837414693e-06, |
|
"loss": 0.7464, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.6289308176100629, |
|
"grad_norm": 8.5625, |
|
"learning_rate": 3.710691823899371e-06, |
|
"loss": 0.7294, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.631607118961595, |
|
"grad_norm": 7.1875, |
|
"learning_rate": 3.68392881038405e-06, |
|
"loss": 0.7184, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.6342834203131272, |
|
"grad_norm": 8.0625, |
|
"learning_rate": 3.657165796868728e-06, |
|
"loss": 0.6361, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.6369597216646594, |
|
"grad_norm": 10.6875, |
|
"learning_rate": 3.6304027833534063e-06, |
|
"loss": 0.6732, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.6396360230161916, |
|
"grad_norm": 10.1875, |
|
"learning_rate": 3.6036397698380843e-06, |
|
"loss": 0.6091, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.6423123243677238, |
|
"grad_norm": 10.75, |
|
"learning_rate": 3.5768767563227623e-06, |
|
"loss": 0.6382, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.644988625719256, |
|
"grad_norm": 10.6875, |
|
"learning_rate": 3.5501137428074407e-06, |
|
"loss": 0.7222, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.6476649270707882, |
|
"grad_norm": 16.5, |
|
"learning_rate": 3.5233507292921187e-06, |
|
"loss": 0.6297, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.6503412284223203, |
|
"grad_norm": 12.375, |
|
"learning_rate": 3.496587715776797e-06, |
|
"loss": 0.709, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.6530175297738525, |
|
"grad_norm": 15.8125, |
|
"learning_rate": 3.469824702261475e-06, |
|
"loss": 0.6754, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.6556938311253847, |
|
"grad_norm": 32.5, |
|
"learning_rate": 3.443061688746153e-06, |
|
"loss": 0.9194, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.6583701324769169, |
|
"grad_norm": 7.75, |
|
"learning_rate": 3.4162986752308315e-06, |
|
"loss": 0.6851, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.6610464338284491, |
|
"grad_norm": 24.5, |
|
"learning_rate": 3.3895356617155095e-06, |
|
"loss": 0.728, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.6637227351799813, |
|
"grad_norm": 13.3125, |
|
"learning_rate": 3.362772648200188e-06, |
|
"loss": 0.6899, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.6663990365315134, |
|
"grad_norm": 12.4375, |
|
"learning_rate": 3.336009634684866e-06, |
|
"loss": 0.714, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.6690753378830456, |
|
"grad_norm": 11.0625, |
|
"learning_rate": 3.309246621169544e-06, |
|
"loss": 0.721, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.6717516392345778, |
|
"grad_norm": 9.3125, |
|
"learning_rate": 3.2824836076542223e-06, |
|
"loss": 0.7363, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.67442794058611, |
|
"grad_norm": 3.921875, |
|
"learning_rate": 3.2557205941389002e-06, |
|
"loss": 0.6483, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.6771042419376422, |
|
"grad_norm": 6.8125, |
|
"learning_rate": 3.2289575806235787e-06, |
|
"loss": 0.7267, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.6797805432891744, |
|
"grad_norm": 8.4375, |
|
"learning_rate": 3.2021945671082566e-06, |
|
"loss": 0.73, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.6824568446407066, |
|
"grad_norm": 11.125, |
|
"learning_rate": 3.1754315535929346e-06, |
|
"loss": 0.7075, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.6851331459922387, |
|
"grad_norm": 15.0, |
|
"learning_rate": 3.148668540077613e-06, |
|
"loss": 0.7503, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.6878094473437709, |
|
"grad_norm": 5.46875, |
|
"learning_rate": 3.121905526562291e-06, |
|
"loss": 0.6256, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.6904857486953031, |
|
"grad_norm": 8.875, |
|
"learning_rate": 3.0951425130469694e-06, |
|
"loss": 0.7156, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.6931620500468353, |
|
"grad_norm": 9.9375, |
|
"learning_rate": 3.0683794995316474e-06, |
|
"loss": 0.6339, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.6958383513983675, |
|
"grad_norm": 6.09375, |
|
"learning_rate": 3.0416164860163254e-06, |
|
"loss": 0.7028, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.6985146527498997, |
|
"grad_norm": 7.65625, |
|
"learning_rate": 3.014853472501004e-06, |
|
"loss": 0.7132, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.7011909541014318, |
|
"grad_norm": 6.0625, |
|
"learning_rate": 2.988090458985682e-06, |
|
"loss": 0.6992, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.703867255452964, |
|
"grad_norm": 7.5625, |
|
"learning_rate": 2.96132744547036e-06, |
|
"loss": 0.7433, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.7065435568044962, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 2.934564431955038e-06, |
|
"loss": 0.6577, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.7092198581560284, |
|
"grad_norm": 37.0, |
|
"learning_rate": 2.907801418439716e-06, |
|
"loss": 0.7209, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.7118961595075606, |
|
"grad_norm": 17.75, |
|
"learning_rate": 2.8810384049243946e-06, |
|
"loss": 0.6553, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.7145724608590928, |
|
"grad_norm": 9.8125, |
|
"learning_rate": 2.854275391409073e-06, |
|
"loss": 0.8166, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.717248762210625, |
|
"grad_norm": 7.0625, |
|
"learning_rate": 2.8275123778937514e-06, |
|
"loss": 0.6525, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.7199250635621571, |
|
"grad_norm": 8.1875, |
|
"learning_rate": 2.8007493643784294e-06, |
|
"loss": 0.6576, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.7226013649136893, |
|
"grad_norm": 9.0, |
|
"learning_rate": 2.773986350863108e-06, |
|
"loss": 0.6565, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.7252776662652215, |
|
"grad_norm": 12.875, |
|
"learning_rate": 2.7472233373477858e-06, |
|
"loss": 0.6848, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.7279539676167537, |
|
"grad_norm": 9.9375, |
|
"learning_rate": 2.7204603238324638e-06, |
|
"loss": 0.6987, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.7306302689682859, |
|
"grad_norm": 16.375, |
|
"learning_rate": 2.693697310317142e-06, |
|
"loss": 0.8492, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.733306570319818, |
|
"grad_norm": 14.5625, |
|
"learning_rate": 2.66693429680182e-06, |
|
"loss": 0.6154, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.7359828716713502, |
|
"grad_norm": 10.3125, |
|
"learning_rate": 2.6401712832864986e-06, |
|
"loss": 0.6097, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.7386591730228824, |
|
"grad_norm": 17.5, |
|
"learning_rate": 2.6134082697711766e-06, |
|
"loss": 0.7904, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.7413354743744146, |
|
"grad_norm": 10.625, |
|
"learning_rate": 2.5866452562558545e-06, |
|
"loss": 0.6887, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.7440117757259468, |
|
"grad_norm": 9.8125, |
|
"learning_rate": 2.559882242740533e-06, |
|
"loss": 0.6463, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.746688077077479, |
|
"grad_norm": 12.0625, |
|
"learning_rate": 2.533119229225211e-06, |
|
"loss": 0.7327, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.7493643784290112, |
|
"grad_norm": 7.4375, |
|
"learning_rate": 2.5063562157098894e-06, |
|
"loss": 0.637, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.7520406797805433, |
|
"grad_norm": 7.78125, |
|
"learning_rate": 2.4795932021945673e-06, |
|
"loss": 0.7037, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.7547169811320755, |
|
"grad_norm": 12.625, |
|
"learning_rate": 2.4528301886792453e-06, |
|
"loss": 0.7834, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.7573932824836076, |
|
"grad_norm": 7.75, |
|
"learning_rate": 2.4260671751639237e-06, |
|
"loss": 0.5978, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.7600695838351398, |
|
"grad_norm": 11.25, |
|
"learning_rate": 2.3993041616486017e-06, |
|
"loss": 0.6954, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.762745885186672, |
|
"grad_norm": 22.75, |
|
"learning_rate": 2.37254114813328e-06, |
|
"loss": 0.6904, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.7654221865382042, |
|
"grad_norm": 9.8125, |
|
"learning_rate": 2.345778134617958e-06, |
|
"loss": 0.6372, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.7680984878897363, |
|
"grad_norm": 9.3125, |
|
"learning_rate": 2.319015121102636e-06, |
|
"loss": 0.5761, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.7707747892412685, |
|
"grad_norm": 7.1875, |
|
"learning_rate": 2.2922521075873145e-06, |
|
"loss": 0.7424, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.7734510905928007, |
|
"grad_norm": 7.3125, |
|
"learning_rate": 2.2654890940719925e-06, |
|
"loss": 0.6615, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.7761273919443329, |
|
"grad_norm": 6.5625, |
|
"learning_rate": 2.238726080556671e-06, |
|
"loss": 0.7074, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.7788036932958651, |
|
"grad_norm": 14.0, |
|
"learning_rate": 2.2119630670413493e-06, |
|
"loss": 0.7508, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.7814799946473973, |
|
"grad_norm": 24.0, |
|
"learning_rate": 2.1852000535260273e-06, |
|
"loss": 0.619, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.7841562959989294, |
|
"grad_norm": 6.21875, |
|
"learning_rate": 2.1584370400107053e-06, |
|
"loss": 0.7471, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.7868325973504616, |
|
"grad_norm": 8.9375, |
|
"learning_rate": 2.1316740264953837e-06, |
|
"loss": 0.6773, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.7895088987019938, |
|
"grad_norm": 13.375, |
|
"learning_rate": 2.1049110129800617e-06, |
|
"loss": 0.6566, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.792185200053526, |
|
"grad_norm": 9.9375, |
|
"learning_rate": 2.07814799946474e-06, |
|
"loss": 0.8063, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.7948615014050582, |
|
"grad_norm": 10.625, |
|
"learning_rate": 2.051384985949418e-06, |
|
"loss": 0.7928, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.7975378027565904, |
|
"grad_norm": 6.15625, |
|
"learning_rate": 2.024621972434096e-06, |
|
"loss": 0.714, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.8002141041081225, |
|
"grad_norm": 4.8125, |
|
"learning_rate": 1.9978589589187745e-06, |
|
"loss": 0.7086, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.8028904054596547, |
|
"grad_norm": 8.5, |
|
"learning_rate": 1.9710959454034525e-06, |
|
"loss": 0.665, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.8055667068111869, |
|
"grad_norm": 5.21875, |
|
"learning_rate": 1.944332931888131e-06, |
|
"loss": 0.6577, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.8082430081627191, |
|
"grad_norm": 10.0625, |
|
"learning_rate": 1.917569918372809e-06, |
|
"loss": 0.6418, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.8109193095142513, |
|
"grad_norm": 9.75, |
|
"learning_rate": 1.890806904857487e-06, |
|
"loss": 0.7353, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.8135956108657835, |
|
"grad_norm": 9.0625, |
|
"learning_rate": 1.8640438913421652e-06, |
|
"loss": 0.7601, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.8162719122173157, |
|
"grad_norm": 13.6875, |
|
"learning_rate": 1.8372808778268432e-06, |
|
"loss": 0.7325, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.8189482135688478, |
|
"grad_norm": 13.9375, |
|
"learning_rate": 1.8105178643115214e-06, |
|
"loss": 0.6905, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.82162451492038, |
|
"grad_norm": 16.0, |
|
"learning_rate": 1.7837548507961998e-06, |
|
"loss": 0.6566, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.8243008162719122, |
|
"grad_norm": 7.59375, |
|
"learning_rate": 1.756991837280878e-06, |
|
"loss": 0.6967, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.8269771176234444, |
|
"grad_norm": 18.75, |
|
"learning_rate": 1.7302288237655562e-06, |
|
"loss": 0.8126, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.8296534189749766, |
|
"grad_norm": 10.5625, |
|
"learning_rate": 1.7034658102502344e-06, |
|
"loss": 0.7035, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.8323297203265088, |
|
"grad_norm": 7.25, |
|
"learning_rate": 1.6767027967349126e-06, |
|
"loss": 0.7178, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.8350060216780409, |
|
"grad_norm": 5.0625, |
|
"learning_rate": 1.6499397832195906e-06, |
|
"loss": 0.7009, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.8376823230295731, |
|
"grad_norm": 7.28125, |
|
"learning_rate": 1.6231767697042688e-06, |
|
"loss": 0.65, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.8403586243811053, |
|
"grad_norm": 8.8125, |
|
"learning_rate": 1.596413756188947e-06, |
|
"loss": 0.6868, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.8430349257326375, |
|
"grad_norm": 12.4375, |
|
"learning_rate": 1.5696507426736252e-06, |
|
"loss": 0.7927, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.8457112270841697, |
|
"grad_norm": 9.3125, |
|
"learning_rate": 1.5428877291583034e-06, |
|
"loss": 0.7587, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.8483875284357019, |
|
"grad_norm": 7.3125, |
|
"learning_rate": 1.5161247156429814e-06, |
|
"loss": 0.7038, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.851063829787234, |
|
"grad_norm": 10.9375, |
|
"learning_rate": 1.4893617021276596e-06, |
|
"loss": 0.8245, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.8537401311387662, |
|
"grad_norm": 8.375, |
|
"learning_rate": 1.4625986886123378e-06, |
|
"loss": 0.6673, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.8564164324902984, |
|
"grad_norm": 7.09375, |
|
"learning_rate": 1.435835675097016e-06, |
|
"loss": 0.7309, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.8590927338418306, |
|
"grad_norm": 8.125, |
|
"learning_rate": 1.4090726615816942e-06, |
|
"loss": 0.725, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.8617690351933628, |
|
"grad_norm": 8.25, |
|
"learning_rate": 1.3823096480663726e-06, |
|
"loss": 0.6646, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.864445336544895, |
|
"grad_norm": 6.4375, |
|
"learning_rate": 1.3555466345510506e-06, |
|
"loss": 0.7158, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.8671216378964272, |
|
"grad_norm": 5.71875, |
|
"learning_rate": 1.3287836210357288e-06, |
|
"loss": 0.6649, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.8697979392479593, |
|
"grad_norm": 16.875, |
|
"learning_rate": 1.302020607520407e-06, |
|
"loss": 0.7767, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.8724742405994915, |
|
"grad_norm": 12.875, |
|
"learning_rate": 1.2752575940050852e-06, |
|
"loss": 0.6689, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.8751505419510237, |
|
"grad_norm": 9.5625, |
|
"learning_rate": 1.2484945804897634e-06, |
|
"loss": 0.6509, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.8778268433025559, |
|
"grad_norm": 8.625, |
|
"learning_rate": 1.2217315669744414e-06, |
|
"loss": 0.6574, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.8805031446540881, |
|
"grad_norm": 12.375, |
|
"learning_rate": 1.1949685534591195e-06, |
|
"loss": 0.7337, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.8831794460056203, |
|
"grad_norm": 9.375, |
|
"learning_rate": 1.1682055399437977e-06, |
|
"loss": 0.6465, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.8858557473571524, |
|
"grad_norm": 8.8125, |
|
"learning_rate": 1.141442526428476e-06, |
|
"loss": 0.7548, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.8885320487086846, |
|
"grad_norm": 11.125, |
|
"learning_rate": 1.1146795129131541e-06, |
|
"loss": 0.7219, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.8912083500602168, |
|
"grad_norm": 6.625, |
|
"learning_rate": 1.0879164993978321e-06, |
|
"loss": 0.8184, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.893884651411749, |
|
"grad_norm": 6.84375, |
|
"learning_rate": 1.0611534858825105e-06, |
|
"loss": 0.6956, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.8965609527632812, |
|
"grad_norm": 7.03125, |
|
"learning_rate": 1.0343904723671887e-06, |
|
"loss": 0.6756, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.8992372541148134, |
|
"grad_norm": 8.1875, |
|
"learning_rate": 1.007627458851867e-06, |
|
"loss": 0.6936, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.9019135554663456, |
|
"grad_norm": 10.0, |
|
"learning_rate": 9.80864445336545e-07, |
|
"loss": 0.6416, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.9045898568178777, |
|
"grad_norm": 8.1875, |
|
"learning_rate": 9.541014318212231e-07, |
|
"loss": 0.7569, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.9072661581694099, |
|
"grad_norm": 47.0, |
|
"learning_rate": 9.273384183059013e-07, |
|
"loss": 0.7972, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.9099424595209421, |
|
"grad_norm": 7.625, |
|
"learning_rate": 9.005754047905794e-07, |
|
"loss": 0.6925, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.9126187608724743, |
|
"grad_norm": 23.875, |
|
"learning_rate": 8.738123912752576e-07, |
|
"loss": 0.67, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.9152950622240065, |
|
"grad_norm": 7.6875, |
|
"learning_rate": 8.470493777599359e-07, |
|
"loss": 0.7114, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.9179713635755387, |
|
"grad_norm": 8.1875, |
|
"learning_rate": 8.202863642446141e-07, |
|
"loss": 0.6751, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.9206476649270708, |
|
"grad_norm": 11.0625, |
|
"learning_rate": 7.935233507292922e-07, |
|
"loss": 0.6867, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.923323966278603, |
|
"grad_norm": 15.375, |
|
"learning_rate": 7.667603372139704e-07, |
|
"loss": 0.7322, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.9260002676301351, |
|
"grad_norm": 6.15625, |
|
"learning_rate": 7.399973236986485e-07, |
|
"loss": 0.7396, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.9286765689816673, |
|
"grad_norm": 18.0, |
|
"learning_rate": 7.132343101833267e-07, |
|
"loss": 0.7134, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.9313528703331995, |
|
"grad_norm": 12.125, |
|
"learning_rate": 6.864712966680049e-07, |
|
"loss": 0.682, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.9340291716847317, |
|
"grad_norm": 8.5625, |
|
"learning_rate": 6.59708283152683e-07, |
|
"loss": 0.8207, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.9367054730362638, |
|
"grad_norm": 5.3125, |
|
"learning_rate": 6.329452696373613e-07, |
|
"loss": 0.6548, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.939381774387796, |
|
"grad_norm": 13.0, |
|
"learning_rate": 6.061822561220394e-07, |
|
"loss": 0.6879, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.9420580757393282, |
|
"grad_norm": 7.625, |
|
"learning_rate": 5.794192426067176e-07, |
|
"loss": 0.7555, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.9447343770908604, |
|
"grad_norm": 6.4375, |
|
"learning_rate": 5.526562290913958e-07, |
|
"loss": 0.6822, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.9474106784423926, |
|
"grad_norm": 6.375, |
|
"learning_rate": 5.258932155760739e-07, |
|
"loss": 0.7273, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.9500869797939248, |
|
"grad_norm": 9.1875, |
|
"learning_rate": 4.99130202060752e-07, |
|
"loss": 0.6715, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.9527632811454569, |
|
"grad_norm": 5.75, |
|
"learning_rate": 4.723671885454303e-07, |
|
"loss": 0.8109, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.9554395824969891, |
|
"grad_norm": 13.75, |
|
"learning_rate": 4.4560417503010845e-07, |
|
"loss": 0.6167, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.9581158838485213, |
|
"grad_norm": 7.59375, |
|
"learning_rate": 4.188411615147866e-07, |
|
"loss": 0.6606, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.9607921852000535, |
|
"grad_norm": 9.1875, |
|
"learning_rate": 3.9207814799946473e-07, |
|
"loss": 0.7073, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.9634684865515857, |
|
"grad_norm": 9.75, |
|
"learning_rate": 3.65315134484143e-07, |
|
"loss": 0.682, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.9661447879031179, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 3.3855212096882113e-07, |
|
"loss": 0.7581, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.96882108925465, |
|
"grad_norm": 8.875, |
|
"learning_rate": 3.117891074534993e-07, |
|
"loss": 0.6873, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.9714973906061822, |
|
"grad_norm": 27.375, |
|
"learning_rate": 2.8502609393817747e-07, |
|
"loss": 0.647, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.9741736919577144, |
|
"grad_norm": 12.1875, |
|
"learning_rate": 2.582630804228556e-07, |
|
"loss": 0.6391, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.9768499933092466, |
|
"grad_norm": 14.1875, |
|
"learning_rate": 2.3150006690753381e-07, |
|
"loss": 0.8126, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.9795262946607788, |
|
"grad_norm": 6.4375, |
|
"learning_rate": 2.0473705339221198e-07, |
|
"loss": 0.7084, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.982202596012311, |
|
"grad_norm": 25.0, |
|
"learning_rate": 1.7797403987689016e-07, |
|
"loss": 0.6649, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.9848788973638432, |
|
"grad_norm": 11.6875, |
|
"learning_rate": 1.5121102636156833e-07, |
|
"loss": 0.7436, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.9875551987153753, |
|
"grad_norm": 12.8125, |
|
"learning_rate": 1.244480128462465e-07, |
|
"loss": 0.6809, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.9902315000669075, |
|
"grad_norm": 7.125, |
|
"learning_rate": 9.768499933092467e-08, |
|
"loss": 0.6806, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.9929078014184397, |
|
"grad_norm": 15.625, |
|
"learning_rate": 7.092198581560284e-08, |
|
"loss": 0.7224, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.9955841027699719, |
|
"grad_norm": 7.46875, |
|
"learning_rate": 4.415897230028101e-08, |
|
"loss": 0.7614, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.9982604041215041, |
|
"grad_norm": 8.5, |
|
"learning_rate": 1.739595878495919e-08, |
|
"loss": 0.6988, |
|
"step": 7460 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 7473, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 5000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.1170284894683136e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|