{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 7473, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0026763013515321826, "grad_norm": 13.1875, "learning_rate": 9.973236986484679e-06, "loss": 1.7384, "step": 20 }, { "epoch": 0.005352602703064365, "grad_norm": 27.875, "learning_rate": 9.946473972969357e-06, "loss": 1.6485, "step": 40 }, { "epoch": 0.008028904054596548, "grad_norm": 9.8125, "learning_rate": 9.919710959454035e-06, "loss": 1.4827, "step": 60 }, { "epoch": 0.01070520540612873, "grad_norm": 6.4375, "learning_rate": 9.892947945938713e-06, "loss": 1.3008, "step": 80 }, { "epoch": 0.013381506757660913, "grad_norm": 15.9375, "learning_rate": 9.86618493242339e-06, "loss": 1.24, "step": 100 }, { "epoch": 0.016057808109193095, "grad_norm": 7.125, "learning_rate": 9.83942191890807e-06, "loss": 0.9532, "step": 120 }, { "epoch": 0.018734109460725276, "grad_norm": 6.375, "learning_rate": 9.812658905392748e-06, "loss": 1.1199, "step": 140 }, { "epoch": 0.02141041081225746, "grad_norm": 9.6875, "learning_rate": 9.785895891877426e-06, "loss": 1.2211, "step": 160 }, { "epoch": 0.02408671216378964, "grad_norm": 8.8125, "learning_rate": 9.759132878362104e-06, "loss": 1.1512, "step": 180 }, { "epoch": 0.026763013515321826, "grad_norm": 6.0625, "learning_rate": 9.732369864846782e-06, "loss": 1.2178, "step": 200 }, { "epoch": 0.029439314866854006, "grad_norm": 12.0, "learning_rate": 9.70560685133146e-06, "loss": 0.9836, "step": 220 }, { "epoch": 0.03211561621838619, "grad_norm": 19.5, "learning_rate": 9.678843837816138e-06, "loss": 1.1758, "step": 240 }, { "epoch": 0.034791917569918375, "grad_norm": 11.6875, "learning_rate": 9.652080824300816e-06, "loss": 1.0029, "step": 260 }, { "epoch": 0.03746821892145055, "grad_norm": 10.8125, "learning_rate": 9.625317810785494e-06, "loss": 0.917, "step": 280 }, { "epoch": 0.04014452027298274, "grad_norm": 5.59375, "learning_rate": 9.598554797270172e-06, "loss": 0.8177, "step": 300 }, { "epoch": 0.04282082162451492, "grad_norm": 7.75, "learning_rate": 9.571791783754852e-06, "loss": 0.8647, "step": 320 }, { "epoch": 0.045497122976047105, "grad_norm": 13.0625, "learning_rate": 9.54502877023953e-06, "loss": 0.7716, "step": 340 }, { "epoch": 0.04817342432757928, "grad_norm": 8.625, "learning_rate": 9.518265756724208e-06, "loss": 0.8501, "step": 360 }, { "epoch": 0.05084972567911147, "grad_norm": 7.53125, "learning_rate": 9.491502743208886e-06, "loss": 0.772, "step": 380 }, { "epoch": 0.05352602703064365, "grad_norm": 11.3125, "learning_rate": 9.464739729693564e-06, "loss": 0.7478, "step": 400 }, { "epoch": 0.056202328382175835, "grad_norm": 13.6875, "learning_rate": 9.437976716178242e-06, "loss": 0.8187, "step": 420 }, { "epoch": 0.05887862973370801, "grad_norm": 11.3125, "learning_rate": 9.41121370266292e-06, "loss": 0.729, "step": 440 }, { "epoch": 0.0615549310852402, "grad_norm": 6.34375, "learning_rate": 9.384450689147598e-06, "loss": 0.7931, "step": 460 }, { "epoch": 0.06423123243677238, "grad_norm": 23.0, "learning_rate": 9.357687675632276e-06, "loss": 0.7999, "step": 480 }, { "epoch": 0.06690753378830457, "grad_norm": 11.125, "learning_rate": 9.330924662116954e-06, "loss": 0.7258, "step": 500 }, { "epoch": 0.06958383513983675, "grad_norm": 33.75, "learning_rate": 9.304161648601634e-06, "loss": 0.8433, "step": 520 }, { "epoch": 0.07226013649136893, "grad_norm": 9.125, "learning_rate": 9.277398635086312e-06, "loss": 0.7657, "step": 540 }, { "epoch": 0.0749364378429011, "grad_norm": 12.875, "learning_rate": 9.25063562157099e-06, "loss": 0.668, "step": 560 }, { "epoch": 0.07761273919443329, "grad_norm": 4.6875, "learning_rate": 9.223872608055667e-06, "loss": 0.7134, "step": 580 }, { "epoch": 0.08028904054596547, "grad_norm": 13.125, "learning_rate": 9.197109594540345e-06, "loss": 0.6519, "step": 600 }, { "epoch": 0.08296534189749766, "grad_norm": 8.5625, "learning_rate": 9.170346581025023e-06, "loss": 0.7203, "step": 620 }, { "epoch": 0.08564164324902984, "grad_norm": 10.0, "learning_rate": 9.143583567509703e-06, "loss": 0.7479, "step": 640 }, { "epoch": 0.08831794460056203, "grad_norm": 8.6875, "learning_rate": 9.116820553994381e-06, "loss": 0.7291, "step": 660 }, { "epoch": 0.09099424595209421, "grad_norm": 9.125, "learning_rate": 9.090057540479059e-06, "loss": 0.7096, "step": 680 }, { "epoch": 0.0936705473036264, "grad_norm": 11.6875, "learning_rate": 9.063294526963737e-06, "loss": 0.7773, "step": 700 }, { "epoch": 0.09634684865515857, "grad_norm": 6.4375, "learning_rate": 9.036531513448415e-06, "loss": 0.7675, "step": 720 }, { "epoch": 0.09902315000669075, "grad_norm": 24.5, "learning_rate": 9.009768499933093e-06, "loss": 0.7259, "step": 740 }, { "epoch": 0.10169945135822293, "grad_norm": 7.84375, "learning_rate": 8.983005486417771e-06, "loss": 0.7389, "step": 760 }, { "epoch": 0.10437575270975512, "grad_norm": 14.0, "learning_rate": 8.95624247290245e-06, "loss": 0.7706, "step": 780 }, { "epoch": 0.1070520540612873, "grad_norm": 11.6875, "learning_rate": 8.929479459387129e-06, "loss": 0.7571, "step": 800 }, { "epoch": 0.10972835541281949, "grad_norm": 18.125, "learning_rate": 8.902716445871807e-06, "loss": 0.7931, "step": 820 }, { "epoch": 0.11240465676435167, "grad_norm": 19.0, "learning_rate": 8.875953432356485e-06, "loss": 0.6576, "step": 840 }, { "epoch": 0.11508095811588386, "grad_norm": 21.0, "learning_rate": 8.849190418841163e-06, "loss": 0.8381, "step": 860 }, { "epoch": 0.11775725946741603, "grad_norm": 16.375, "learning_rate": 8.82242740532584e-06, "loss": 0.6927, "step": 880 }, { "epoch": 0.12043356081894821, "grad_norm": 7.0625, "learning_rate": 8.795664391810519e-06, "loss": 0.725, "step": 900 }, { "epoch": 0.1231098621704804, "grad_norm": 16.375, "learning_rate": 8.768901378295197e-06, "loss": 0.6435, "step": 920 }, { "epoch": 0.12578616352201258, "grad_norm": 4.8125, "learning_rate": 8.742138364779875e-06, "loss": 0.7177, "step": 940 }, { "epoch": 0.12846246487354476, "grad_norm": 9.375, "learning_rate": 8.715375351264553e-06, "loss": 0.6612, "step": 960 }, { "epoch": 0.13113876622507695, "grad_norm": 7.1875, "learning_rate": 8.688612337749232e-06, "loss": 0.6302, "step": 980 }, { "epoch": 0.13381506757660913, "grad_norm": 5.5, "learning_rate": 8.66184932423391e-06, "loss": 0.6037, "step": 1000 }, { "epoch": 0.13649136892814132, "grad_norm": 7.34375, "learning_rate": 8.635086310718588e-06, "loss": 0.6317, "step": 1020 }, { "epoch": 0.1391676702796735, "grad_norm": 15.5625, "learning_rate": 8.608323297203266e-06, "loss": 0.6688, "step": 1040 }, { "epoch": 0.14184397163120568, "grad_norm": 4.53125, "learning_rate": 8.581560283687944e-06, "loss": 0.6914, "step": 1060 }, { "epoch": 0.14452027298273787, "grad_norm": 9.25, "learning_rate": 8.554797270172622e-06, "loss": 0.6758, "step": 1080 }, { "epoch": 0.14719657433427003, "grad_norm": 12.4375, "learning_rate": 8.5280342566573e-06, "loss": 0.7167, "step": 1100 }, { "epoch": 0.1498728756858022, "grad_norm": 11.6875, "learning_rate": 8.501271243141978e-06, "loss": 0.7415, "step": 1120 }, { "epoch": 0.1525491770373344, "grad_norm": 18.5, "learning_rate": 8.474508229626656e-06, "loss": 0.7248, "step": 1140 }, { "epoch": 0.15522547838886658, "grad_norm": 9.25, "learning_rate": 8.447745216111334e-06, "loss": 0.7272, "step": 1160 }, { "epoch": 0.15790177974039876, "grad_norm": 15.0625, "learning_rate": 8.420982202596014e-06, "loss": 0.713, "step": 1180 }, { "epoch": 0.16057808109193095, "grad_norm": 22.375, "learning_rate": 8.394219189080692e-06, "loss": 0.7674, "step": 1200 }, { "epoch": 0.16325438244346313, "grad_norm": 11.0625, "learning_rate": 8.36745617556537e-06, "loss": 0.6709, "step": 1220 }, { "epoch": 0.16593068379499532, "grad_norm": 6.75, "learning_rate": 8.340693162050048e-06, "loss": 0.6167, "step": 1240 }, { "epoch": 0.1686069851465275, "grad_norm": 15.0625, "learning_rate": 8.313930148534726e-06, "loss": 0.6662, "step": 1260 }, { "epoch": 0.17128328649805968, "grad_norm": 7.9375, "learning_rate": 8.287167135019404e-06, "loss": 0.6959, "step": 1280 }, { "epoch": 0.17395958784959187, "grad_norm": 7.625, "learning_rate": 8.260404121504082e-06, "loss": 0.7078, "step": 1300 }, { "epoch": 0.17663588920112405, "grad_norm": 4.78125, "learning_rate": 8.23364110798876e-06, "loss": 0.7281, "step": 1320 }, { "epoch": 0.17931219055265624, "grad_norm": 38.75, "learning_rate": 8.206878094473438e-06, "loss": 0.707, "step": 1340 }, { "epoch": 0.18198849190418842, "grad_norm": 10.6875, "learning_rate": 8.180115080958116e-06, "loss": 0.7312, "step": 1360 }, { "epoch": 0.1846647932557206, "grad_norm": 11.25, "learning_rate": 8.153352067442795e-06, "loss": 0.7054, "step": 1380 }, { "epoch": 0.1873410946072528, "grad_norm": 14.25, "learning_rate": 8.126589053927473e-06, "loss": 0.6535, "step": 1400 }, { "epoch": 0.19001739595878495, "grad_norm": 9.0, "learning_rate": 8.099826040412151e-06, "loss": 0.6309, "step": 1420 }, { "epoch": 0.19269369731031713, "grad_norm": 24.875, "learning_rate": 8.07306302689683e-06, "loss": 0.6625, "step": 1440 }, { "epoch": 0.19536999866184931, "grad_norm": 9.9375, "learning_rate": 8.046300013381507e-06, "loss": 0.702, "step": 1460 }, { "epoch": 0.1980463000133815, "grad_norm": 6.46875, "learning_rate": 8.019536999866185e-06, "loss": 0.7289, "step": 1480 }, { "epoch": 0.20072260136491368, "grad_norm": 7.3125, "learning_rate": 7.992773986350863e-06, "loss": 0.6527, "step": 1500 }, { "epoch": 0.20339890271644587, "grad_norm": 7.46875, "learning_rate": 7.966010972835541e-06, "loss": 0.6631, "step": 1520 }, { "epoch": 0.20607520406797805, "grad_norm": 12.5, "learning_rate": 7.93924795932022e-06, "loss": 0.5758, "step": 1540 }, { "epoch": 0.20875150541951024, "grad_norm": 4.875, "learning_rate": 7.912484945804897e-06, "loss": 0.6604, "step": 1560 }, { "epoch": 0.21142780677104242, "grad_norm": 9.125, "learning_rate": 7.885721932289577e-06, "loss": 0.6006, "step": 1580 }, { "epoch": 0.2141041081225746, "grad_norm": 11.0625, "learning_rate": 7.858958918774255e-06, "loss": 0.5989, "step": 1600 }, { "epoch": 0.2167804094741068, "grad_norm": 11.75, "learning_rate": 7.832195905258933e-06, "loss": 0.709, "step": 1620 }, { "epoch": 0.21945671082563897, "grad_norm": 16.375, "learning_rate": 7.805432891743611e-06, "loss": 0.6903, "step": 1640 }, { "epoch": 0.22213301217717116, "grad_norm": 9.4375, "learning_rate": 7.778669878228289e-06, "loss": 0.6708, "step": 1660 }, { "epoch": 0.22480931352870334, "grad_norm": 9.6875, "learning_rate": 7.751906864712967e-06, "loss": 0.7066, "step": 1680 }, { "epoch": 0.22748561488023553, "grad_norm": 5.71875, "learning_rate": 7.725143851197645e-06, "loss": 0.6876, "step": 1700 }, { "epoch": 0.2301619162317677, "grad_norm": 10.0625, "learning_rate": 7.698380837682323e-06, "loss": 0.6687, "step": 1720 }, { "epoch": 0.23283821758329987, "grad_norm": 22.875, "learning_rate": 7.671617824167e-06, "loss": 0.6319, "step": 1740 }, { "epoch": 0.23551451893483205, "grad_norm": 10.9375, "learning_rate": 7.644854810651679e-06, "loss": 0.6421, "step": 1760 }, { "epoch": 0.23819082028636424, "grad_norm": 42.5, "learning_rate": 7.618091797136358e-06, "loss": 0.607, "step": 1780 }, { "epoch": 0.24086712163789642, "grad_norm": 11.8125, "learning_rate": 7.591328783621036e-06, "loss": 0.6368, "step": 1800 }, { "epoch": 0.2435434229894286, "grad_norm": 8.0625, "learning_rate": 7.564565770105714e-06, "loss": 0.693, "step": 1820 }, { "epoch": 0.2462197243409608, "grad_norm": 21.75, "learning_rate": 7.5378027565903925e-06, "loss": 0.735, "step": 1840 }, { "epoch": 0.24889602569249297, "grad_norm": 25.5, "learning_rate": 7.5110397430750704e-06, "loss": 0.6872, "step": 1860 }, { "epoch": 0.25157232704402516, "grad_norm": 5.5, "learning_rate": 7.484276729559748e-06, "loss": 0.7235, "step": 1880 }, { "epoch": 0.25424862839555734, "grad_norm": 5.875, "learning_rate": 7.457513716044428e-06, "loss": 0.6518, "step": 1900 }, { "epoch": 0.2569249297470895, "grad_norm": 7.03125, "learning_rate": 7.430750702529106e-06, "loss": 0.6251, "step": 1920 }, { "epoch": 0.2596012310986217, "grad_norm": 5.59375, "learning_rate": 7.403987689013784e-06, "loss": 0.6323, "step": 1940 }, { "epoch": 0.2622775324501539, "grad_norm": 5.65625, "learning_rate": 7.377224675498462e-06, "loss": 0.7128, "step": 1960 }, { "epoch": 0.2649538338016861, "grad_norm": 5.53125, "learning_rate": 7.35046166198314e-06, "loss": 0.6354, "step": 1980 }, { "epoch": 0.26763013515321826, "grad_norm": 6.59375, "learning_rate": 7.323698648467819e-06, "loss": 0.6783, "step": 2000 }, { "epoch": 0.27030643650475045, "grad_norm": 9.5, "learning_rate": 7.296935634952497e-06, "loss": 0.6546, "step": 2020 }, { "epoch": 0.27298273785628263, "grad_norm": 14.6875, "learning_rate": 7.270172621437175e-06, "loss": 0.7473, "step": 2040 }, { "epoch": 0.2756590392078148, "grad_norm": 4.25, "learning_rate": 7.243409607921853e-06, "loss": 0.5906, "step": 2060 }, { "epoch": 0.278335340559347, "grad_norm": 7.53125, "learning_rate": 7.216646594406531e-06, "loss": 0.6394, "step": 2080 }, { "epoch": 0.2810116419108792, "grad_norm": 10.25, "learning_rate": 7.18988358089121e-06, "loss": 0.6928, "step": 2100 }, { "epoch": 0.28368794326241137, "grad_norm": 14.0625, "learning_rate": 7.163120567375888e-06, "loss": 0.7175, "step": 2120 }, { "epoch": 0.28636424461394355, "grad_norm": 5.9375, "learning_rate": 7.136357553860566e-06, "loss": 0.6362, "step": 2140 }, { "epoch": 0.28904054596547574, "grad_norm": 17.125, "learning_rate": 7.109594540345244e-06, "loss": 0.6462, "step": 2160 }, { "epoch": 0.2917168473170079, "grad_norm": 6.3125, "learning_rate": 7.082831526829922e-06, "loss": 0.6098, "step": 2180 }, { "epoch": 0.29439314866854005, "grad_norm": 13.8125, "learning_rate": 7.0560685133146004e-06, "loss": 0.6703, "step": 2200 }, { "epoch": 0.29706945002007223, "grad_norm": 10.1875, "learning_rate": 7.029305499799278e-06, "loss": 0.6079, "step": 2220 }, { "epoch": 0.2997457513716044, "grad_norm": 25.875, "learning_rate": 7.002542486283956e-06, "loss": 0.6221, "step": 2240 }, { "epoch": 0.3024220527231366, "grad_norm": 13.5625, "learning_rate": 6.975779472768634e-06, "loss": 0.7642, "step": 2260 }, { "epoch": 0.3050983540746688, "grad_norm": 21.0, "learning_rate": 6.949016459253312e-06, "loss": 0.5935, "step": 2280 }, { "epoch": 0.30777465542620097, "grad_norm": 9.625, "learning_rate": 6.922253445737991e-06, "loss": 0.658, "step": 2300 }, { "epoch": 0.31045095677773316, "grad_norm": 9.6875, "learning_rate": 6.895490432222669e-06, "loss": 0.653, "step": 2320 }, { "epoch": 0.31312725812926534, "grad_norm": 18.125, "learning_rate": 6.868727418707347e-06, "loss": 0.6094, "step": 2340 }, { "epoch": 0.3158035594807975, "grad_norm": 10.625, "learning_rate": 6.841964405192025e-06, "loss": 0.6108, "step": 2360 }, { "epoch": 0.3184798608323297, "grad_norm": 8.25, "learning_rate": 6.815201391676703e-06, "loss": 0.6254, "step": 2380 }, { "epoch": 0.3211561621838619, "grad_norm": 7.0625, "learning_rate": 6.788438378161382e-06, "loss": 0.5189, "step": 2400 }, { "epoch": 0.3238324635353941, "grad_norm": 12.3125, "learning_rate": 6.76167536464606e-06, "loss": 0.6326, "step": 2420 }, { "epoch": 0.32650876488692626, "grad_norm": 8.4375, "learning_rate": 6.734912351130738e-06, "loss": 0.541, "step": 2440 }, { "epoch": 0.32918506623845845, "grad_norm": 8.125, "learning_rate": 6.708149337615416e-06, "loss": 0.627, "step": 2460 }, { "epoch": 0.33186136758999063, "grad_norm": 9.6875, "learning_rate": 6.681386324100094e-06, "loss": 0.5483, "step": 2480 }, { "epoch": 0.3345376689415228, "grad_norm": 4.8125, "learning_rate": 6.654623310584773e-06, "loss": 0.6937, "step": 2500 }, { "epoch": 0.337213970293055, "grad_norm": 7.40625, "learning_rate": 6.627860297069451e-06, "loss": 0.7344, "step": 2520 }, { "epoch": 0.3398902716445872, "grad_norm": 6.3125, "learning_rate": 6.601097283554129e-06, "loss": 0.5555, "step": 2540 }, { "epoch": 0.34256657299611937, "grad_norm": 8.4375, "learning_rate": 6.574334270038807e-06, "loss": 0.6078, "step": 2560 }, { "epoch": 0.34524287434765155, "grad_norm": 35.25, "learning_rate": 6.547571256523485e-06, "loss": 0.5849, "step": 2580 }, { "epoch": 0.34791917569918374, "grad_norm": 5.3125, "learning_rate": 6.5208082430081635e-06, "loss": 0.5821, "step": 2600 }, { "epoch": 0.3505954770507159, "grad_norm": 11.25, "learning_rate": 6.4940452294928415e-06, "loss": 0.6821, "step": 2620 }, { "epoch": 0.3532717784022481, "grad_norm": 4.40625, "learning_rate": 6.4672822159775195e-06, "loss": 0.608, "step": 2640 }, { "epoch": 0.3559480797537803, "grad_norm": 9.75, "learning_rate": 6.4405192024621975e-06, "loss": 0.6809, "step": 2660 }, { "epoch": 0.3586243811053125, "grad_norm": 4.9375, "learning_rate": 6.4137561889468755e-06, "loss": 0.5272, "step": 2680 }, { "epoch": 0.36130068245684466, "grad_norm": 14.25, "learning_rate": 6.386993175431554e-06, "loss": 0.5292, "step": 2700 }, { "epoch": 0.36397698380837684, "grad_norm": 12.125, "learning_rate": 6.360230161916232e-06, "loss": 0.6039, "step": 2720 }, { "epoch": 0.366653285159909, "grad_norm": 5.71875, "learning_rate": 6.33346714840091e-06, "loss": 0.681, "step": 2740 }, { "epoch": 0.3693295865114412, "grad_norm": 24.5, "learning_rate": 6.306704134885588e-06, "loss": 0.6367, "step": 2760 }, { "epoch": 0.3720058878629734, "grad_norm": 5.84375, "learning_rate": 6.279941121370266e-06, "loss": 0.6731, "step": 2780 }, { "epoch": 0.3746821892145056, "grad_norm": 16.125, "learning_rate": 6.253178107854945e-06, "loss": 0.5908, "step": 2800 }, { "epoch": 0.37735849056603776, "grad_norm": 11.0, "learning_rate": 6.226415094339623e-06, "loss": 0.6916, "step": 2820 }, { "epoch": 0.3800347919175699, "grad_norm": 8.75, "learning_rate": 6.199652080824301e-06, "loss": 0.5824, "step": 2840 }, { "epoch": 0.3827110932691021, "grad_norm": 6.59375, "learning_rate": 6.172889067308979e-06, "loss": 0.5781, "step": 2860 }, { "epoch": 0.38538739462063426, "grad_norm": 12.75, "learning_rate": 6.146126053793657e-06, "loss": 0.5776, "step": 2880 }, { "epoch": 0.38806369597216644, "grad_norm": 16.5, "learning_rate": 6.119363040278336e-06, "loss": 0.6522, "step": 2900 }, { "epoch": 0.39073999732369863, "grad_norm": 6.90625, "learning_rate": 6.092600026763014e-06, "loss": 0.5461, "step": 2920 }, { "epoch": 0.3934162986752308, "grad_norm": 4.4375, "learning_rate": 6.065837013247692e-06, "loss": 0.6251, "step": 2940 }, { "epoch": 0.396092600026763, "grad_norm": 27.0, "learning_rate": 6.03907399973237e-06, "loss": 0.5881, "step": 2960 }, { "epoch": 0.3987689013782952, "grad_norm": 23.625, "learning_rate": 6.012310986217048e-06, "loss": 0.5181, "step": 2980 }, { "epoch": 0.40144520272982737, "grad_norm": 7.375, "learning_rate": 5.985547972701727e-06, "loss": 0.666, "step": 3000 }, { "epoch": 0.40412150408135955, "grad_norm": 16.125, "learning_rate": 5.958784959186405e-06, "loss": 0.6627, "step": 3020 }, { "epoch": 0.40679780543289173, "grad_norm": 7.09375, "learning_rate": 5.932021945671083e-06, "loss": 0.6242, "step": 3040 }, { "epoch": 0.4094741067844239, "grad_norm": 11.5, "learning_rate": 5.905258932155761e-06, "loss": 0.6298, "step": 3060 }, { "epoch": 0.4121504081359561, "grad_norm": 8.0625, "learning_rate": 5.878495918640439e-06, "loss": 0.5965, "step": 3080 }, { "epoch": 0.4148267094874883, "grad_norm": 11.1875, "learning_rate": 5.851732905125117e-06, "loss": 0.6369, "step": 3100 }, { "epoch": 0.41750301083902047, "grad_norm": 8.5, "learning_rate": 5.824969891609795e-06, "loss": 0.6442, "step": 3120 }, { "epoch": 0.42017931219055266, "grad_norm": 33.0, "learning_rate": 5.798206878094474e-06, "loss": 0.6203, "step": 3140 }, { "epoch": 0.42285561354208484, "grad_norm": 5.0625, "learning_rate": 5.771443864579152e-06, "loss": 0.6406, "step": 3160 }, { "epoch": 0.425531914893617, "grad_norm": 10.4375, "learning_rate": 5.744680851063831e-06, "loss": 0.6375, "step": 3180 }, { "epoch": 0.4282082162451492, "grad_norm": 5.625, "learning_rate": 5.717917837548509e-06, "loss": 0.6048, "step": 3200 }, { "epoch": 0.4308845175966814, "grad_norm": 7.75, "learning_rate": 5.691154824033187e-06, "loss": 0.6119, "step": 3220 }, { "epoch": 0.4335608189482136, "grad_norm": 6.75, "learning_rate": 5.664391810517865e-06, "loss": 0.6523, "step": 3240 }, { "epoch": 0.43623712029974576, "grad_norm": 7.09375, "learning_rate": 5.637628797002543e-06, "loss": 0.6033, "step": 3260 }, { "epoch": 0.43891342165127795, "grad_norm": 9.1875, "learning_rate": 5.610865783487222e-06, "loss": 0.5765, "step": 3280 }, { "epoch": 0.44158972300281013, "grad_norm": 4.90625, "learning_rate": 5.5841027699719e-06, "loss": 0.5753, "step": 3300 }, { "epoch": 0.4442660243543423, "grad_norm": 5.5625, "learning_rate": 5.557339756456578e-06, "loss": 0.6599, "step": 3320 }, { "epoch": 0.4469423257058745, "grad_norm": 9.9375, "learning_rate": 5.530576742941256e-06, "loss": 0.5884, "step": 3340 }, { "epoch": 0.4496186270574067, "grad_norm": 5.65625, "learning_rate": 5.503813729425934e-06, "loss": 0.5796, "step": 3360 }, { "epoch": 0.45229492840893887, "grad_norm": 11.0625, "learning_rate": 5.477050715910613e-06, "loss": 0.664, "step": 3380 }, { "epoch": 0.45497122976047105, "grad_norm": 7.3125, "learning_rate": 5.450287702395291e-06, "loss": 0.5953, "step": 3400 }, { "epoch": 0.45764753111200324, "grad_norm": 8.6875, "learning_rate": 5.4235246888799686e-06, "loss": 0.6893, "step": 3420 }, { "epoch": 0.4603238324635354, "grad_norm": 13.375, "learning_rate": 5.3967616753646466e-06, "loss": 0.6659, "step": 3440 }, { "epoch": 0.46300013381506755, "grad_norm": 8.625, "learning_rate": 5.3699986618493245e-06, "loss": 0.578, "step": 3460 }, { "epoch": 0.46567643516659973, "grad_norm": 7.40625, "learning_rate": 5.343235648334003e-06, "loss": 0.651, "step": 3480 }, { "epoch": 0.4683527365181319, "grad_norm": 30.875, "learning_rate": 5.316472634818681e-06, "loss": 0.6255, "step": 3500 }, { "epoch": 0.4710290378696641, "grad_norm": 4.9375, "learning_rate": 5.289709621303359e-06, "loss": 0.514, "step": 3520 }, { "epoch": 0.4737053392211963, "grad_norm": 7.84375, "learning_rate": 5.262946607788037e-06, "loss": 0.6143, "step": 3540 }, { "epoch": 0.47638164057272847, "grad_norm": 7.875, "learning_rate": 5.236183594272715e-06, "loss": 0.7061, "step": 3560 }, { "epoch": 0.47905794192426066, "grad_norm": 14.9375, "learning_rate": 5.209420580757394e-06, "loss": 0.5776, "step": 3580 }, { "epoch": 0.48173424327579284, "grad_norm": 6.71875, "learning_rate": 5.182657567242072e-06, "loss": 0.6318, "step": 3600 }, { "epoch": 0.484410544627325, "grad_norm": 18.5, "learning_rate": 5.15589455372675e-06, "loss": 0.6403, "step": 3620 }, { "epoch": 0.4870868459788572, "grad_norm": 9.625, "learning_rate": 5.129131540211428e-06, "loss": 0.6161, "step": 3640 }, { "epoch": 0.4897631473303894, "grad_norm": 6.09375, "learning_rate": 5.102368526696106e-06, "loss": 0.6192, "step": 3660 }, { "epoch": 0.4924394486819216, "grad_norm": 5.0, "learning_rate": 5.075605513180785e-06, "loss": 0.61, "step": 3680 }, { "epoch": 0.49511575003345376, "grad_norm": 5.9375, "learning_rate": 5.048842499665463e-06, "loss": 0.6077, "step": 3700 }, { "epoch": 0.49779205138498595, "grad_norm": 48.25, "learning_rate": 5.022079486150141e-06, "loss": 0.6666, "step": 3720 }, { "epoch": 0.5004683527365181, "grad_norm": 5.71875, "learning_rate": 4.995316472634819e-06, "loss": 0.6599, "step": 3740 }, { "epoch": 0.5031446540880503, "grad_norm": 5.75, "learning_rate": 4.968553459119497e-06, "loss": 0.5544, "step": 3760 }, { "epoch": 0.5058209554395825, "grad_norm": 6.21875, "learning_rate": 4.941790445604176e-06, "loss": 0.6117, "step": 3780 }, { "epoch": 0.5084972567911147, "grad_norm": 10.5625, "learning_rate": 4.915027432088854e-06, "loss": 0.6413, "step": 3800 }, { "epoch": 0.5111735581426469, "grad_norm": 6.65625, "learning_rate": 4.888264418573532e-06, "loss": 0.5872, "step": 3820 }, { "epoch": 0.513849859494179, "grad_norm": 10.375, "learning_rate": 4.86150140505821e-06, "loss": 0.5602, "step": 3840 }, { "epoch": 0.5165261608457112, "grad_norm": 6.03125, "learning_rate": 4.834738391542888e-06, "loss": 0.6806, "step": 3860 }, { "epoch": 0.5192024621972434, "grad_norm": 7.0, "learning_rate": 4.8079753780275665e-06, "loss": 0.632, "step": 3880 }, { "epoch": 0.5218787635487756, "grad_norm": 6.03125, "learning_rate": 4.7812123645122445e-06, "loss": 0.7411, "step": 3900 }, { "epoch": 0.5245550649003078, "grad_norm": 4.375, "learning_rate": 4.7544493509969225e-06, "loss": 0.6028, "step": 3920 }, { "epoch": 0.52723136625184, "grad_norm": 9.6875, "learning_rate": 4.7276863374816004e-06, "loss": 0.654, "step": 3940 }, { "epoch": 0.5299076676033722, "grad_norm": 9.25, "learning_rate": 4.7009233239662784e-06, "loss": 0.6127, "step": 3960 }, { "epoch": 0.5325839689549043, "grad_norm": 5.875, "learning_rate": 4.674160310450957e-06, "loss": 0.5728, "step": 3980 }, { "epoch": 0.5352602703064365, "grad_norm": 19.375, "learning_rate": 4.647397296935635e-06, "loss": 0.5786, "step": 4000 }, { "epoch": 0.5379365716579687, "grad_norm": 16.75, "learning_rate": 4.620634283420313e-06, "loss": 0.6569, "step": 4020 }, { "epoch": 0.5406128730095009, "grad_norm": 4.46875, "learning_rate": 4.593871269904991e-06, "loss": 0.6774, "step": 4040 }, { "epoch": 0.5432891743610331, "grad_norm": 41.75, "learning_rate": 4.567108256389669e-06, "loss": 0.603, "step": 4060 }, { "epoch": 0.5459654757125653, "grad_norm": 5.15625, "learning_rate": 4.540345242874348e-06, "loss": 0.5679, "step": 4080 }, { "epoch": 0.5486417770640974, "grad_norm": 18.125, "learning_rate": 4.513582229359027e-06, "loss": 0.5686, "step": 4100 }, { "epoch": 0.5513180784156296, "grad_norm": 6.90625, "learning_rate": 4.486819215843705e-06, "loss": 0.5776, "step": 4120 }, { "epoch": 0.5539943797671618, "grad_norm": 6.15625, "learning_rate": 4.460056202328383e-06, "loss": 0.6051, "step": 4140 }, { "epoch": 0.556670681118694, "grad_norm": 8.25, "learning_rate": 4.433293188813061e-06, "loss": 0.7062, "step": 4160 }, { "epoch": 0.5593469824702262, "grad_norm": 8.6875, "learning_rate": 4.406530175297739e-06, "loss": 0.6409, "step": 4180 }, { "epoch": 0.5620232838217584, "grad_norm": 7.59375, "learning_rate": 4.379767161782418e-06, "loss": 0.5696, "step": 4200 }, { "epoch": 0.5646995851732906, "grad_norm": 8.0625, "learning_rate": 4.353004148267096e-06, "loss": 0.6228, "step": 4220 }, { "epoch": 0.5673758865248227, "grad_norm": 6.03125, "learning_rate": 4.326241134751774e-06, "loss": 0.6532, "step": 4240 }, { "epoch": 0.5700521878763549, "grad_norm": 6.375, "learning_rate": 4.299478121236452e-06, "loss": 0.6136, "step": 4260 }, { "epoch": 0.5727284892278871, "grad_norm": 11.9375, "learning_rate": 4.27271510772113e-06, "loss": 0.6101, "step": 4280 }, { "epoch": 0.5754047905794193, "grad_norm": 57.25, "learning_rate": 4.245952094205808e-06, "loss": 0.6618, "step": 4300 }, { "epoch": 0.5780810919309515, "grad_norm": 40.0, "learning_rate": 4.219189080690486e-06, "loss": 0.6603, "step": 4320 }, { "epoch": 0.5807573932824837, "grad_norm": 8.0, "learning_rate": 4.192426067175164e-06, "loss": 0.5636, "step": 4340 }, { "epoch": 0.5834336946340158, "grad_norm": 10.6875, "learning_rate": 4.165663053659842e-06, "loss": 0.5962, "step": 4360 }, { "epoch": 0.586109995985548, "grad_norm": 17.0, "learning_rate": 4.13890004014452e-06, "loss": 0.5037, "step": 4380 }, { "epoch": 0.5887862973370801, "grad_norm": 4.40625, "learning_rate": 4.112137026629199e-06, "loss": 0.6774, "step": 4400 }, { "epoch": 0.5914625986886123, "grad_norm": 18.75, "learning_rate": 4.085374013113877e-06, "loss": 0.681, "step": 4420 }, { "epoch": 0.5941389000401445, "grad_norm": 15.875, "learning_rate": 4.058610999598555e-06, "loss": 0.5419, "step": 4440 }, { "epoch": 0.5968152013916767, "grad_norm": 11.75, "learning_rate": 4.031847986083233e-06, "loss": 0.6745, "step": 4460 }, { "epoch": 0.5994915027432088, "grad_norm": 9.375, "learning_rate": 4.005084972567911e-06, "loss": 0.5266, "step": 4480 }, { "epoch": 0.602167804094741, "grad_norm": 8.0, "learning_rate": 3.97832195905259e-06, "loss": 0.6198, "step": 4500 }, { "epoch": 0.6048441054462732, "grad_norm": 7.90625, "learning_rate": 3.951558945537268e-06, "loss": 0.6739, "step": 4520 }, { "epoch": 0.6075204067978054, "grad_norm": 5.625, "learning_rate": 3.924795932021946e-06, "loss": 0.6192, "step": 4540 }, { "epoch": 0.6101967081493376, "grad_norm": 7.9375, "learning_rate": 3.898032918506624e-06, "loss": 0.548, "step": 4560 }, { "epoch": 0.6128730095008698, "grad_norm": 20.875, "learning_rate": 3.871269904991302e-06, "loss": 0.5932, "step": 4580 }, { "epoch": 0.6155493108524019, "grad_norm": 13.8125, "learning_rate": 3.844506891475981e-06, "loss": 0.5685, "step": 4600 }, { "epoch": 0.6182256122039341, "grad_norm": 6.71875, "learning_rate": 3.817743877960659e-06, "loss": 0.5804, "step": 4620 }, { "epoch": 0.6209019135554663, "grad_norm": 8.375, "learning_rate": 3.7909808644453367e-06, "loss": 0.5648, "step": 4640 }, { "epoch": 0.6235782149069985, "grad_norm": 5.875, "learning_rate": 3.7642178509300147e-06, "loss": 0.6447, "step": 4660 }, { "epoch": 0.6262545162585307, "grad_norm": 29.625, "learning_rate": 3.737454837414693e-06, "loss": 0.6378, "step": 4680 }, { "epoch": 0.6289308176100629, "grad_norm": 5.40625, "learning_rate": 3.710691823899371e-06, "loss": 0.6327, "step": 4700 }, { "epoch": 0.631607118961595, "grad_norm": 5.3125, "learning_rate": 3.68392881038405e-06, "loss": 0.6192, "step": 4720 }, { "epoch": 0.6342834203131272, "grad_norm": 9.8125, "learning_rate": 3.657165796868728e-06, "loss": 0.5407, "step": 4740 }, { "epoch": 0.6369597216646594, "grad_norm": 10.5, "learning_rate": 3.6304027833534063e-06, "loss": 0.6072, "step": 4760 }, { "epoch": 0.6396360230161916, "grad_norm": 6.40625, "learning_rate": 3.6036397698380843e-06, "loss": 0.5386, "step": 4780 }, { "epoch": 0.6423123243677238, "grad_norm": 13.9375, "learning_rate": 3.5768767563227623e-06, "loss": 0.5386, "step": 4800 }, { "epoch": 0.644988625719256, "grad_norm": 7.03125, "learning_rate": 3.5501137428074407e-06, "loss": 0.6285, "step": 4820 }, { "epoch": 0.6476649270707882, "grad_norm": 42.5, "learning_rate": 3.5233507292921187e-06, "loss": 0.5539, "step": 4840 }, { "epoch": 0.6503412284223203, "grad_norm": 10.875, "learning_rate": 3.496587715776797e-06, "loss": 0.6209, "step": 4860 }, { "epoch": 0.6530175297738525, "grad_norm": 9.1875, "learning_rate": 3.469824702261475e-06, "loss": 0.6097, "step": 4880 }, { "epoch": 0.6556938311253847, "grad_norm": 42.0, "learning_rate": 3.443061688746153e-06, "loss": 0.7998, "step": 4900 }, { "epoch": 0.6583701324769169, "grad_norm": 8.375, "learning_rate": 3.4162986752308315e-06, "loss": 0.5952, "step": 4920 }, { "epoch": 0.6610464338284491, "grad_norm": 8.25, "learning_rate": 3.3895356617155095e-06, "loss": 0.6066, "step": 4940 }, { "epoch": 0.6637227351799813, "grad_norm": 9.6875, "learning_rate": 3.362772648200188e-06, "loss": 0.5866, "step": 4960 }, { "epoch": 0.6663990365315134, "grad_norm": 15.5, "learning_rate": 3.336009634684866e-06, "loss": 0.6064, "step": 4980 }, { "epoch": 0.6690753378830456, "grad_norm": 6.0, "learning_rate": 3.309246621169544e-06, "loss": 0.6234, "step": 5000 }, { "epoch": 0.6717516392345778, "grad_norm": 8.6875, "learning_rate": 3.2824836076542223e-06, "loss": 0.65, "step": 5020 }, { "epoch": 0.67442794058611, "grad_norm": 7.59375, "learning_rate": 3.2557205941389002e-06, "loss": 0.5471, "step": 5040 }, { "epoch": 0.6771042419376422, "grad_norm": 5.71875, "learning_rate": 3.2289575806235787e-06, "loss": 0.6251, "step": 5060 }, { "epoch": 0.6797805432891744, "grad_norm": 5.34375, "learning_rate": 3.2021945671082566e-06, "loss": 0.6402, "step": 5080 }, { "epoch": 0.6824568446407066, "grad_norm": 10.125, "learning_rate": 3.1754315535929346e-06, "loss": 0.6108, "step": 5100 }, { "epoch": 0.6851331459922387, "grad_norm": 11.0, "learning_rate": 3.148668540077613e-06, "loss": 0.6438, "step": 5120 }, { "epoch": 0.6878094473437709, "grad_norm": 4.28125, "learning_rate": 3.121905526562291e-06, "loss": 0.5578, "step": 5140 }, { "epoch": 0.6904857486953031, "grad_norm": 6.0625, "learning_rate": 3.0951425130469694e-06, "loss": 0.6331, "step": 5160 }, { "epoch": 0.6931620500468353, "grad_norm": 10.25, "learning_rate": 3.0683794995316474e-06, "loss": 0.5664, "step": 5180 }, { "epoch": 0.6958383513983675, "grad_norm": 4.78125, "learning_rate": 3.0416164860163254e-06, "loss": 0.6339, "step": 5200 }, { "epoch": 0.6985146527498997, "grad_norm": 4.875, "learning_rate": 3.014853472501004e-06, "loss": 0.6027, "step": 5220 }, { "epoch": 0.7011909541014318, "grad_norm": 5.59375, "learning_rate": 2.988090458985682e-06, "loss": 0.6055, "step": 5240 }, { "epoch": 0.703867255452964, "grad_norm": 5.03125, "learning_rate": 2.96132744547036e-06, "loss": 0.6562, "step": 5260 }, { "epoch": 0.7065435568044962, "grad_norm": 5.8125, "learning_rate": 2.934564431955038e-06, "loss": 0.5782, "step": 5280 }, { "epoch": 0.7092198581560284, "grad_norm": 28.125, "learning_rate": 2.907801418439716e-06, "loss": 0.6352, "step": 5300 }, { "epoch": 0.7118961595075606, "grad_norm": 8.75, "learning_rate": 2.8810384049243946e-06, "loss": 0.5951, "step": 5320 }, { "epoch": 0.7145724608590928, "grad_norm": 5.71875, "learning_rate": 2.854275391409073e-06, "loss": 0.6997, "step": 5340 }, { "epoch": 0.717248762210625, "grad_norm": 6.0625, "learning_rate": 2.8275123778937514e-06, "loss": 0.5509, "step": 5360 }, { "epoch": 0.7199250635621571, "grad_norm": 5.875, "learning_rate": 2.8007493643784294e-06, "loss": 0.5825, "step": 5380 }, { "epoch": 0.7226013649136893, "grad_norm": 5.84375, "learning_rate": 2.773986350863108e-06, "loss": 0.5669, "step": 5400 }, { "epoch": 0.7252776662652215, "grad_norm": 16.875, "learning_rate": 2.7472233373477858e-06, "loss": 0.6002, "step": 5420 }, { "epoch": 0.7279539676167537, "grad_norm": 6.1875, "learning_rate": 2.7204603238324638e-06, "loss": 0.5909, "step": 5440 }, { "epoch": 0.7306302689682859, "grad_norm": 18.5, "learning_rate": 2.693697310317142e-06, "loss": 0.7438, "step": 5460 }, { "epoch": 0.733306570319818, "grad_norm": 7.71875, "learning_rate": 2.66693429680182e-06, "loss": 0.5287, "step": 5480 }, { "epoch": 0.7359828716713502, "grad_norm": 8.625, "learning_rate": 2.6401712832864986e-06, "loss": 0.5491, "step": 5500 }, { "epoch": 0.7386591730228824, "grad_norm": 16.25, "learning_rate": 2.6134082697711766e-06, "loss": 0.6913, "step": 5520 }, { "epoch": 0.7413354743744146, "grad_norm": 8.8125, "learning_rate": 2.5866452562558545e-06, "loss": 0.5872, "step": 5540 }, { "epoch": 0.7440117757259468, "grad_norm": 14.4375, "learning_rate": 2.559882242740533e-06, "loss": 0.5462, "step": 5560 }, { "epoch": 0.746688077077479, "grad_norm": 5.21875, "learning_rate": 2.533119229225211e-06, "loss": 0.6182, "step": 5580 }, { "epoch": 0.7493643784290112, "grad_norm": 5.21875, "learning_rate": 2.5063562157098894e-06, "loss": 0.5553, "step": 5600 }, { "epoch": 0.7520406797805433, "grad_norm": 5.6875, "learning_rate": 2.4795932021945673e-06, "loss": 0.6205, "step": 5620 }, { "epoch": 0.7547169811320755, "grad_norm": 9.25, "learning_rate": 2.4528301886792453e-06, "loss": 0.6668, "step": 5640 }, { "epoch": 0.7573932824836076, "grad_norm": 5.25, "learning_rate": 2.4260671751639237e-06, "loss": 0.525, "step": 5660 }, { "epoch": 0.7600695838351398, "grad_norm": 11.9375, "learning_rate": 2.3993041616486017e-06, "loss": 0.6166, "step": 5680 }, { "epoch": 0.762745885186672, "grad_norm": 10.9375, "learning_rate": 2.37254114813328e-06, "loss": 0.6011, "step": 5700 }, { "epoch": 0.7654221865382042, "grad_norm": 14.5625, "learning_rate": 2.345778134617958e-06, "loss": 0.5479, "step": 5720 }, { "epoch": 0.7680984878897363, "grad_norm": 6.90625, "learning_rate": 2.319015121102636e-06, "loss": 0.4926, "step": 5740 }, { "epoch": 0.7707747892412685, "grad_norm": 7.84375, "learning_rate": 2.2922521075873145e-06, "loss": 0.6657, "step": 5760 }, { "epoch": 0.7734510905928007, "grad_norm": 7.34375, "learning_rate": 2.2654890940719925e-06, "loss": 0.569, "step": 5780 }, { "epoch": 0.7761273919443329, "grad_norm": 8.75, "learning_rate": 2.238726080556671e-06, "loss": 0.6252, "step": 5800 }, { "epoch": 0.7788036932958651, "grad_norm": 12.6875, "learning_rate": 2.2119630670413493e-06, "loss": 0.6618, "step": 5820 }, { "epoch": 0.7814799946473973, "grad_norm": 7.125, "learning_rate": 2.1852000535260273e-06, "loss": 0.5305, "step": 5840 }, { "epoch": 0.7841562959989294, "grad_norm": 7.1875, "learning_rate": 2.1584370400107053e-06, "loss": 0.6592, "step": 5860 }, { "epoch": 0.7868325973504616, "grad_norm": 7.71875, "learning_rate": 2.1316740264953837e-06, "loss": 0.57, "step": 5880 }, { "epoch": 0.7895088987019938, "grad_norm": 12.1875, "learning_rate": 2.1049110129800617e-06, "loss": 0.5813, "step": 5900 }, { "epoch": 0.792185200053526, "grad_norm": 9.25, "learning_rate": 2.07814799946474e-06, "loss": 0.7133, "step": 5920 }, { "epoch": 0.7948615014050582, "grad_norm": 9.6875, "learning_rate": 2.051384985949418e-06, "loss": 0.6929, "step": 5940 }, { "epoch": 0.7975378027565904, "grad_norm": 7.96875, "learning_rate": 2.024621972434096e-06, "loss": 0.612, "step": 5960 }, { "epoch": 0.8002141041081225, "grad_norm": 5.09375, "learning_rate": 1.9978589589187745e-06, "loss": 0.6441, "step": 5980 }, { "epoch": 0.8028904054596547, "grad_norm": 8.875, "learning_rate": 1.9710959454034525e-06, "loss": 0.5688, "step": 6000 }, { "epoch": 0.8055667068111869, "grad_norm": 4.375, "learning_rate": 1.944332931888131e-06, "loss": 0.5676, "step": 6020 }, { "epoch": 0.8082430081627191, "grad_norm": 19.875, "learning_rate": 1.917569918372809e-06, "loss": 0.5413, "step": 6040 }, { "epoch": 0.8109193095142513, "grad_norm": 6.84375, "learning_rate": 1.890806904857487e-06, "loss": 0.6589, "step": 6060 }, { "epoch": 0.8135956108657835, "grad_norm": 7.90625, "learning_rate": 1.8640438913421652e-06, "loss": 0.6596, "step": 6080 }, { "epoch": 0.8162719122173157, "grad_norm": 6.625, "learning_rate": 1.8372808778268432e-06, "loss": 0.6569, "step": 6100 }, { "epoch": 0.8189482135688478, "grad_norm": 13.375, "learning_rate": 1.8105178643115214e-06, "loss": 0.6127, "step": 6120 }, { "epoch": 0.82162451492038, "grad_norm": 10.1875, "learning_rate": 1.7837548507961998e-06, "loss": 0.5638, "step": 6140 }, { "epoch": 0.8243008162719122, "grad_norm": 5.34375, "learning_rate": 1.756991837280878e-06, "loss": 0.591, "step": 6160 }, { "epoch": 0.8269771176234444, "grad_norm": 10.4375, "learning_rate": 1.7302288237655562e-06, "loss": 0.7082, "step": 6180 }, { "epoch": 0.8296534189749766, "grad_norm": 8.625, "learning_rate": 1.7034658102502344e-06, "loss": 0.6457, "step": 6200 }, { "epoch": 0.8323297203265088, "grad_norm": 5.875, "learning_rate": 1.6767027967349126e-06, "loss": 0.6338, "step": 6220 }, { "epoch": 0.8350060216780409, "grad_norm": 4.8125, "learning_rate": 1.6499397832195906e-06, "loss": 0.6131, "step": 6240 }, { "epoch": 0.8376823230295731, "grad_norm": 4.6875, "learning_rate": 1.6231767697042688e-06, "loss": 0.5457, "step": 6260 }, { "epoch": 0.8403586243811053, "grad_norm": 9.375, "learning_rate": 1.596413756188947e-06, "loss": 0.5772, "step": 6280 }, { "epoch": 0.8430349257326375, "grad_norm": 10.875, "learning_rate": 1.5696507426736252e-06, "loss": 0.6949, "step": 6300 }, { "epoch": 0.8457112270841697, "grad_norm": 8.625, "learning_rate": 1.5428877291583034e-06, "loss": 0.6427, "step": 6320 }, { "epoch": 0.8483875284357019, "grad_norm": 8.6875, "learning_rate": 1.5161247156429814e-06, "loss": 0.6007, "step": 6340 }, { "epoch": 0.851063829787234, "grad_norm": 7.5, "learning_rate": 1.4893617021276596e-06, "loss": 0.7373, "step": 6360 }, { "epoch": 0.8537401311387662, "grad_norm": 4.59375, "learning_rate": 1.4625986886123378e-06, "loss": 0.577, "step": 6380 }, { "epoch": 0.8564164324902984, "grad_norm": 4.84375, "learning_rate": 1.435835675097016e-06, "loss": 0.6276, "step": 6400 }, { "epoch": 0.8590927338418306, "grad_norm": 5.75, "learning_rate": 1.4090726615816942e-06, "loss": 0.628, "step": 6420 }, { "epoch": 0.8617690351933628, "grad_norm": 5.90625, "learning_rate": 1.3823096480663726e-06, "loss": 0.5873, "step": 6440 }, { "epoch": 0.864445336544895, "grad_norm": 4.59375, "learning_rate": 1.3555466345510506e-06, "loss": 0.6292, "step": 6460 }, { "epoch": 0.8671216378964272, "grad_norm": 3.828125, "learning_rate": 1.3287836210357288e-06, "loss": 0.5848, "step": 6480 }, { "epoch": 0.8697979392479593, "grad_norm": 19.5, "learning_rate": 1.302020607520407e-06, "loss": 0.6519, "step": 6500 }, { "epoch": 0.8724742405994915, "grad_norm": 8.5, "learning_rate": 1.2752575940050852e-06, "loss": 0.5921, "step": 6520 }, { "epoch": 0.8751505419510237, "grad_norm": 7.21875, "learning_rate": 1.2484945804897634e-06, "loss": 0.5985, "step": 6540 }, { "epoch": 0.8778268433025559, "grad_norm": 5.28125, "learning_rate": 1.2217315669744414e-06, "loss": 0.5775, "step": 6560 }, { "epoch": 0.8805031446540881, "grad_norm": 10.625, "learning_rate": 1.1949685534591195e-06, "loss": 0.6396, "step": 6580 }, { "epoch": 0.8831794460056203, "grad_norm": 9.6875, "learning_rate": 1.1682055399437977e-06, "loss": 0.5473, "step": 6600 }, { "epoch": 0.8858557473571524, "grad_norm": 7.34375, "learning_rate": 1.141442526428476e-06, "loss": 0.6597, "step": 6620 }, { "epoch": 0.8885320487086846, "grad_norm": 11.125, "learning_rate": 1.1146795129131541e-06, "loss": 0.6388, "step": 6640 }, { "epoch": 0.8912083500602168, "grad_norm": 8.125, "learning_rate": 1.0879164993978321e-06, "loss": 0.7043, "step": 6660 }, { "epoch": 0.893884651411749, "grad_norm": 4.53125, "learning_rate": 1.0611534858825105e-06, "loss": 0.6001, "step": 6680 }, { "epoch": 0.8965609527632812, "grad_norm": 5.34375, "learning_rate": 1.0343904723671887e-06, "loss": 0.592, "step": 6700 }, { "epoch": 0.8992372541148134, "grad_norm": 8.4375, "learning_rate": 1.007627458851867e-06, "loss": 0.6011, "step": 6720 }, { "epoch": 0.9019135554663456, "grad_norm": 5.90625, "learning_rate": 9.80864445336545e-07, "loss": 0.5611, "step": 6740 }, { "epoch": 0.9045898568178777, "grad_norm": 6.375, "learning_rate": 9.541014318212231e-07, "loss": 0.658, "step": 6760 }, { "epoch": 0.9072661581694099, "grad_norm": 52.5, "learning_rate": 9.273384183059013e-07, "loss": 0.6819, "step": 6780 }, { "epoch": 0.9099424595209421, "grad_norm": 16.5, "learning_rate": 9.005754047905794e-07, "loss": 0.6003, "step": 6800 }, { "epoch": 0.9126187608724743, "grad_norm": 10.125, "learning_rate": 8.738123912752576e-07, "loss": 0.5508, "step": 6820 }, { "epoch": 0.9152950622240065, "grad_norm": 7.75, "learning_rate": 8.470493777599359e-07, "loss": 0.641, "step": 6840 }, { "epoch": 0.9179713635755387, "grad_norm": 6.84375, "learning_rate": 8.202863642446141e-07, "loss": 0.5932, "step": 6860 }, { "epoch": 0.9206476649270708, "grad_norm": 13.0, "learning_rate": 7.935233507292922e-07, "loss": 0.6069, "step": 6880 }, { "epoch": 0.923323966278603, "grad_norm": 6.28125, "learning_rate": 7.667603372139704e-07, "loss": 0.6463, "step": 6900 }, { "epoch": 0.9260002676301351, "grad_norm": 3.828125, "learning_rate": 7.399973236986485e-07, "loss": 0.6454, "step": 6920 }, { "epoch": 0.9286765689816673, "grad_norm": 15.1875, "learning_rate": 7.132343101833267e-07, "loss": 0.5931, "step": 6940 }, { "epoch": 0.9313528703331995, "grad_norm": 24.375, "learning_rate": 6.864712966680049e-07, "loss": 0.5925, "step": 6960 }, { "epoch": 0.9340291716847317, "grad_norm": 6.03125, "learning_rate": 6.59708283152683e-07, "loss": 0.6978, "step": 6980 }, { "epoch": 0.9367054730362638, "grad_norm": 7.71875, "learning_rate": 6.329452696373613e-07, "loss": 0.5375, "step": 7000 }, { "epoch": 0.939381774387796, "grad_norm": 9.375, "learning_rate": 6.061822561220394e-07, "loss": 0.6064, "step": 7020 }, { "epoch": 0.9420580757393282, "grad_norm": 4.84375, "learning_rate": 5.794192426067176e-07, "loss": 0.6413, "step": 7040 }, { "epoch": 0.9447343770908604, "grad_norm": 9.3125, "learning_rate": 5.526562290913958e-07, "loss": 0.5989, "step": 7060 }, { "epoch": 0.9474106784423926, "grad_norm": 12.8125, "learning_rate": 5.258932155760739e-07, "loss": 0.6521, "step": 7080 }, { "epoch": 0.9500869797939248, "grad_norm": 4.59375, "learning_rate": 4.99130202060752e-07, "loss": 0.5703, "step": 7100 }, { "epoch": 0.9527632811454569, "grad_norm": 4.25, "learning_rate": 4.723671885454303e-07, "loss": 0.7012, "step": 7120 }, { "epoch": 0.9554395824969891, "grad_norm": 23.0, "learning_rate": 4.4560417503010845e-07, "loss": 0.535, "step": 7140 }, { "epoch": 0.9581158838485213, "grad_norm": 7.375, "learning_rate": 4.188411615147866e-07, "loss": 0.5745, "step": 7160 }, { "epoch": 0.9607921852000535, "grad_norm": 11.0625, "learning_rate": 3.9207814799946473e-07, "loss": 0.6288, "step": 7180 }, { "epoch": 0.9634684865515857, "grad_norm": 11.9375, "learning_rate": 3.65315134484143e-07, "loss": 0.5873, "step": 7200 }, { "epoch": 0.9661447879031179, "grad_norm": 4.46875, "learning_rate": 3.3855212096882113e-07, "loss": 0.6598, "step": 7220 }, { "epoch": 0.96882108925465, "grad_norm": 4.71875, "learning_rate": 3.117891074534993e-07, "loss": 0.5922, "step": 7240 }, { "epoch": 0.9714973906061822, "grad_norm": 9.5625, "learning_rate": 2.8502609393817747e-07, "loss": 0.5569, "step": 7260 }, { "epoch": 0.9741736919577144, "grad_norm": 11.3125, "learning_rate": 2.582630804228556e-07, "loss": 0.5598, "step": 7280 }, { "epoch": 0.9768499933092466, "grad_norm": 12.0, "learning_rate": 2.3150006690753381e-07, "loss": 0.6874, "step": 7300 }, { "epoch": 0.9795262946607788, "grad_norm": 5.4375, "learning_rate": 2.0473705339221198e-07, "loss": 0.616, "step": 7320 }, { "epoch": 0.982202596012311, "grad_norm": 17.0, "learning_rate": 1.7797403987689016e-07, "loss": 0.5929, "step": 7340 }, { "epoch": 0.9848788973638432, "grad_norm": 10.5625, "learning_rate": 1.5121102636156833e-07, "loss": 0.641, "step": 7360 }, { "epoch": 0.9875551987153753, "grad_norm": 11.5625, "learning_rate": 1.244480128462465e-07, "loss": 0.6001, "step": 7380 }, { "epoch": 0.9902315000669075, "grad_norm": 6.8125, "learning_rate": 9.768499933092467e-08, "loss": 0.5997, "step": 7400 }, { "epoch": 0.9929078014184397, "grad_norm": 11.3125, "learning_rate": 7.092198581560284e-08, "loss": 0.6143, "step": 7420 }, { "epoch": 0.9955841027699719, "grad_norm": 3.953125, "learning_rate": 4.415897230028101e-08, "loss": 0.6439, "step": 7440 }, { "epoch": 0.9982604041215041, "grad_norm": 6.625, "learning_rate": 1.739595878495919e-08, "loss": 0.5995, "step": 7460 } ], "logging_steps": 20, "max_steps": 7473, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.2355070863998976e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }