{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2422, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00041288191577208916, "grad_norm": 24.94086805103786, "learning_rate": 4.1152263374485605e-08, "loss": 1.4345, "step": 1 }, { "epoch": 0.002064409578860446, "grad_norm": 24.898038168161193, "learning_rate": 2.05761316872428e-07, "loss": 1.4101, "step": 5 }, { "epoch": 0.004128819157720892, "grad_norm": 15.938300194174698, "learning_rate": 4.11522633744856e-07, "loss": 1.3943, "step": 10 }, { "epoch": 0.006193228736581338, "grad_norm": 8.846886408597587, "learning_rate": 6.17283950617284e-07, "loss": 1.2771, "step": 15 }, { "epoch": 0.008257638315441783, "grad_norm": 9.948610282503696, "learning_rate": 8.23045267489712e-07, "loss": 1.1572, "step": 20 }, { "epoch": 0.01032204789430223, "grad_norm": 4.744946112205854, "learning_rate": 1.02880658436214e-06, "loss": 1.0575, "step": 25 }, { "epoch": 0.012386457473162676, "grad_norm": 3.3793530185545926, "learning_rate": 1.234567901234568e-06, "loss": 0.9708, "step": 30 }, { "epoch": 0.014450867052023121, "grad_norm": 3.4379702794812124, "learning_rate": 1.440329218106996e-06, "loss": 0.9668, "step": 35 }, { "epoch": 0.016515276630883566, "grad_norm": 3.110466178670075, "learning_rate": 1.646090534979424e-06, "loss": 0.9508, "step": 40 }, { "epoch": 0.018579686209744012, "grad_norm": 3.0934122499344614, "learning_rate": 1.8518518518518519e-06, "loss": 0.9496, "step": 45 }, { "epoch": 0.02064409578860446, "grad_norm": 3.2046299727793572, "learning_rate": 2.05761316872428e-06, "loss": 0.9256, "step": 50 }, { "epoch": 0.022708505367464906, "grad_norm": 2.9857050496874358, "learning_rate": 2.263374485596708e-06, "loss": 0.9146, "step": 55 }, { "epoch": 0.02477291494632535, "grad_norm": 3.010697976532333, "learning_rate": 2.469135802469136e-06, "loss": 0.9185, "step": 60 }, { "epoch": 0.026837324525185797, "grad_norm": 2.9986835294430274, "learning_rate": 2.674897119341564e-06, "loss": 0.9008, "step": 65 }, { "epoch": 0.028901734104046242, "grad_norm": 3.2527007862069186, "learning_rate": 2.880658436213992e-06, "loss": 0.8855, "step": 70 }, { "epoch": 0.030966143682906688, "grad_norm": 3.2007644630686833, "learning_rate": 3.08641975308642e-06, "loss": 0.8783, "step": 75 }, { "epoch": 0.03303055326176713, "grad_norm": 3.196196143650837, "learning_rate": 3.292181069958848e-06, "loss": 0.8775, "step": 80 }, { "epoch": 0.03509496284062758, "grad_norm": 3.004102067065696, "learning_rate": 3.4979423868312762e-06, "loss": 0.8808, "step": 85 }, { "epoch": 0.037159372419488024, "grad_norm": 3.118135767316795, "learning_rate": 3.7037037037037037e-06, "loss": 0.8775, "step": 90 }, { "epoch": 0.03922378199834847, "grad_norm": 3.0856528450327354, "learning_rate": 3.909465020576132e-06, "loss": 0.8634, "step": 95 }, { "epoch": 0.04128819157720892, "grad_norm": 2.9700335979493167, "learning_rate": 4.11522633744856e-06, "loss": 0.872, "step": 100 }, { "epoch": 0.04335260115606936, "grad_norm": 3.1350056589977533, "learning_rate": 4.3209876543209875e-06, "loss": 0.8706, "step": 105 }, { "epoch": 0.04541701073492981, "grad_norm": 3.2497160658033852, "learning_rate": 4.526748971193416e-06, "loss": 0.8539, "step": 110 }, { "epoch": 0.047481420313790254, "grad_norm": 3.382546616433861, "learning_rate": 4.732510288065844e-06, "loss": 0.8573, "step": 115 }, { "epoch": 0.0495458298926507, "grad_norm": 3.1741955682207124, "learning_rate": 4.938271604938272e-06, "loss": 0.8433, "step": 120 }, { "epoch": 0.051610239471511145, "grad_norm": 3.13507489217465, "learning_rate": 5.1440329218107e-06, "loss": 0.8546, "step": 125 }, { "epoch": 0.053674649050371594, "grad_norm": 3.182938519722128, "learning_rate": 5.349794238683128e-06, "loss": 0.8505, "step": 130 }, { "epoch": 0.05573905862923204, "grad_norm": 3.3061978280140916, "learning_rate": 5.555555555555557e-06, "loss": 0.8596, "step": 135 }, { "epoch": 0.057803468208092484, "grad_norm": 3.172304485577032, "learning_rate": 5.761316872427984e-06, "loss": 0.8454, "step": 140 }, { "epoch": 0.05986787778695293, "grad_norm": 3.222279073856588, "learning_rate": 5.967078189300412e-06, "loss": 0.8552, "step": 145 }, { "epoch": 0.061932287365813375, "grad_norm": 3.026540545603275, "learning_rate": 6.17283950617284e-06, "loss": 0.8365, "step": 150 }, { "epoch": 0.06399669694467382, "grad_norm": 3.1724013332685304, "learning_rate": 6.3786008230452675e-06, "loss": 0.8397, "step": 155 }, { "epoch": 0.06606110652353427, "grad_norm": 3.4516704362207222, "learning_rate": 6.584362139917696e-06, "loss": 0.8346, "step": 160 }, { "epoch": 0.06812551610239471, "grad_norm": 3.1199583778413285, "learning_rate": 6.790123456790124e-06, "loss": 0.8296, "step": 165 }, { "epoch": 0.07018992568125516, "grad_norm": 3.448087429701013, "learning_rate": 6.9958847736625525e-06, "loss": 0.8384, "step": 170 }, { "epoch": 0.07225433526011561, "grad_norm": 3.072968814322973, "learning_rate": 7.201646090534981e-06, "loss": 0.846, "step": 175 }, { "epoch": 0.07431874483897605, "grad_norm": 3.1647169459938076, "learning_rate": 7.4074074074074075e-06, "loss": 0.8256, "step": 180 }, { "epoch": 0.0763831544178365, "grad_norm": 3.105263189462399, "learning_rate": 7.613168724279836e-06, "loss": 0.8011, "step": 185 }, { "epoch": 0.07844756399669695, "grad_norm": 2.9946543968073294, "learning_rate": 7.818930041152263e-06, "loss": 0.8214, "step": 190 }, { "epoch": 0.0805119735755574, "grad_norm": 3.25501723509638, "learning_rate": 8.024691358024692e-06, "loss": 0.8201, "step": 195 }, { "epoch": 0.08257638315441784, "grad_norm": 3.0087568189695237, "learning_rate": 8.23045267489712e-06, "loss": 0.8139, "step": 200 }, { "epoch": 0.08464079273327828, "grad_norm": 3.0911976211351817, "learning_rate": 8.43621399176955e-06, "loss": 0.8242, "step": 205 }, { "epoch": 0.08670520231213873, "grad_norm": 3.310286578361364, "learning_rate": 8.641975308641975e-06, "loss": 0.8364, "step": 210 }, { "epoch": 0.08876961189099918, "grad_norm": 2.986473336622259, "learning_rate": 8.847736625514404e-06, "loss": 0.8316, "step": 215 }, { "epoch": 0.09083402146985962, "grad_norm": 3.102015428130825, "learning_rate": 9.053497942386832e-06, "loss": 0.8128, "step": 220 }, { "epoch": 0.09289843104872006, "grad_norm": 2.906374211555821, "learning_rate": 9.25925925925926e-06, "loss": 0.7945, "step": 225 }, { "epoch": 0.09496284062758051, "grad_norm": 3.0506200713500826, "learning_rate": 9.465020576131688e-06, "loss": 0.8158, "step": 230 }, { "epoch": 0.09702725020644096, "grad_norm": 3.0384361806454017, "learning_rate": 9.670781893004116e-06, "loss": 0.8228, "step": 235 }, { "epoch": 0.0990916597853014, "grad_norm": 3.100843008960617, "learning_rate": 9.876543209876543e-06, "loss": 0.8099, "step": 240 }, { "epoch": 0.10115606936416185, "grad_norm": 2.935521796714598, "learning_rate": 9.999979213327006e-06, "loss": 0.8108, "step": 245 }, { "epoch": 0.10322047894302229, "grad_norm": 3.361172390200279, "learning_rate": 9.999745365240722e-06, "loss": 0.7923, "step": 250 }, { "epoch": 0.10528488852188274, "grad_norm": 2.915556545903911, "learning_rate": 9.999251697919687e-06, "loss": 0.8045, "step": 255 }, { "epoch": 0.10734929810074319, "grad_norm": 3.212286307862521, "learning_rate": 9.998498237018056e-06, "loss": 0.8112, "step": 260 }, { "epoch": 0.10941370767960364, "grad_norm": 3.191428090938715, "learning_rate": 9.997485021690553e-06, "loss": 0.8001, "step": 265 }, { "epoch": 0.11147811725846409, "grad_norm": 2.838764527991799, "learning_rate": 9.996212104590426e-06, "loss": 0.8074, "step": 270 }, { "epoch": 0.11354252683732452, "grad_norm": 2.9644134550569943, "learning_rate": 9.994679551866712e-06, "loss": 0.8033, "step": 275 }, { "epoch": 0.11560693641618497, "grad_norm": 3.118728479888768, "learning_rate": 9.992887443160804e-06, "loss": 0.7905, "step": 280 }, { "epoch": 0.11767134599504542, "grad_norm": 2.8425828854985737, "learning_rate": 9.990835871602305e-06, "loss": 0.7987, "step": 285 }, { "epoch": 0.11973575557390587, "grad_norm": 3.091935940986862, "learning_rate": 9.988524943804196e-06, "loss": 0.7919, "step": 290 }, { "epoch": 0.12180016515276632, "grad_norm": 2.9471270363090976, "learning_rate": 9.985954779857292e-06, "loss": 0.7885, "step": 295 }, { "epoch": 0.12386457473162675, "grad_norm": 2.761224162387467, "learning_rate": 9.983125513324e-06, "loss": 0.784, "step": 300 }, { "epoch": 0.1259289843104872, "grad_norm": 2.779726805954528, "learning_rate": 9.98003729123138e-06, "loss": 0.7762, "step": 305 }, { "epoch": 0.12799339388934763, "grad_norm": 2.8806127663541625, "learning_rate": 9.976690274063507e-06, "loss": 0.7995, "step": 310 }, { "epoch": 0.13005780346820808, "grad_norm": 3.0861309652893163, "learning_rate": 9.973084635753125e-06, "loss": 0.8002, "step": 315 }, { "epoch": 0.13212221304706853, "grad_norm": 2.886159344081087, "learning_rate": 9.969220563672614e-06, "loss": 0.7742, "step": 320 }, { "epoch": 0.13418662262592898, "grad_norm": 2.9262415184999835, "learning_rate": 9.965098258624253e-06, "loss": 0.7807, "step": 325 }, { "epoch": 0.13625103220478943, "grad_norm": 2.9922747460068324, "learning_rate": 9.960717934829777e-06, "loss": 0.7582, "step": 330 }, { "epoch": 0.13831544178364988, "grad_norm": 2.9066701618612596, "learning_rate": 9.956079819919254e-06, "loss": 0.7849, "step": 335 }, { "epoch": 0.14037985136251033, "grad_norm": 2.9577891790066766, "learning_rate": 9.951184154919253e-06, "loss": 0.7676, "step": 340 }, { "epoch": 0.14244426094137078, "grad_norm": 2.7242674149455923, "learning_rate": 9.946031194240317e-06, "loss": 0.7613, "step": 345 }, { "epoch": 0.14450867052023122, "grad_norm": 2.738187421683147, "learning_rate": 9.940621205663745e-06, "loss": 0.7476, "step": 350 }, { "epoch": 0.14657308009909165, "grad_norm": 2.959447667783807, "learning_rate": 9.934954470327671e-06, "loss": 0.757, "step": 355 }, { "epoch": 0.1486374896779521, "grad_norm": 2.9344152274329693, "learning_rate": 9.92903128271246e-06, "loss": 0.7559, "step": 360 }, { "epoch": 0.15070189925681254, "grad_norm": 3.0358566235662554, "learning_rate": 9.9228519506254e-06, "loss": 0.7671, "step": 365 }, { "epoch": 0.152766308835673, "grad_norm": 2.917481842444454, "learning_rate": 9.916416795184716e-06, "loss": 0.7423, "step": 370 }, { "epoch": 0.15483071841453344, "grad_norm": 2.858632389269523, "learning_rate": 9.90972615080287e-06, "loss": 0.7625, "step": 375 }, { "epoch": 0.1568951279933939, "grad_norm": 2.7940389389869997, "learning_rate": 9.902780365169186e-06, "loss": 0.7344, "step": 380 }, { "epoch": 0.15895953757225434, "grad_norm": 2.966296190863466, "learning_rate": 9.895579799231791e-06, "loss": 0.7431, "step": 385 }, { "epoch": 0.1610239471511148, "grad_norm": 2.734933032890328, "learning_rate": 9.888124827178847e-06, "loss": 0.7543, "step": 390 }, { "epoch": 0.16308835672997524, "grad_norm": 3.0106930594709183, "learning_rate": 9.880415836419108e-06, "loss": 0.7322, "step": 395 }, { "epoch": 0.16515276630883569, "grad_norm": 3.1160711879023277, "learning_rate": 9.872453227561795e-06, "loss": 0.7217, "step": 400 }, { "epoch": 0.1672171758876961, "grad_norm": 3.025574483100628, "learning_rate": 9.864237414395768e-06, "loss": 0.7234, "step": 405 }, { "epoch": 0.16928158546655656, "grad_norm": 3.0800133917199024, "learning_rate": 9.855768823868026e-06, "loss": 0.7451, "step": 410 }, { "epoch": 0.171345995045417, "grad_norm": 2.878566581474389, "learning_rate": 9.847047896061526e-06, "loss": 0.7479, "step": 415 }, { "epoch": 0.17341040462427745, "grad_norm": 3.168491396422218, "learning_rate": 9.838075084172303e-06, "loss": 0.7258, "step": 420 }, { "epoch": 0.1754748142031379, "grad_norm": 2.815061772988165, "learning_rate": 9.828850854485929e-06, "loss": 0.7055, "step": 425 }, { "epoch": 0.17753922378199835, "grad_norm": 2.6873037207867356, "learning_rate": 9.819375686353277e-06, "loss": 0.7253, "step": 430 }, { "epoch": 0.1796036333608588, "grad_norm": 2.8648079803750077, "learning_rate": 9.809650072165606e-06, "loss": 0.7218, "step": 435 }, { "epoch": 0.18166804293971925, "grad_norm": 2.6485478231791446, "learning_rate": 9.799674517328985e-06, "loss": 0.7211, "step": 440 }, { "epoch": 0.1837324525185797, "grad_norm": 2.9086971891718365, "learning_rate": 9.78944954023802e-06, "loss": 0.7113, "step": 445 }, { "epoch": 0.18579686209744012, "grad_norm": 2.777331603392391, "learning_rate": 9.778975672248909e-06, "loss": 0.7236, "step": 450 }, { "epoch": 0.18786127167630057, "grad_norm": 2.896862579927988, "learning_rate": 9.768253457651853e-06, "loss": 0.7239, "step": 455 }, { "epoch": 0.18992568125516102, "grad_norm": 2.788383349290103, "learning_rate": 9.757283453642741e-06, "loss": 0.7083, "step": 460 }, { "epoch": 0.19199009083402147, "grad_norm": 2.7799217861819816, "learning_rate": 9.746066230294216e-06, "loss": 0.7008, "step": 465 }, { "epoch": 0.19405450041288191, "grad_norm": 3.0828748770941603, "learning_rate": 9.734602370526042e-06, "loss": 0.6949, "step": 470 }, { "epoch": 0.19611890999174236, "grad_norm": 2.8583233928158145, "learning_rate": 9.722892470074811e-06, "loss": 0.6923, "step": 475 }, { "epoch": 0.1981833195706028, "grad_norm": 2.850888921108626, "learning_rate": 9.710937137462992e-06, "loss": 0.7056, "step": 480 }, { "epoch": 0.20024772914946326, "grad_norm": 2.8804381008997697, "learning_rate": 9.698736993967295e-06, "loss": 0.7134, "step": 485 }, { "epoch": 0.2023121387283237, "grad_norm": 2.5675849510127473, "learning_rate": 9.686292673586397e-06, "loss": 0.6854, "step": 490 }, { "epoch": 0.20437654830718416, "grad_norm": 2.752726728070582, "learning_rate": 9.673604823007994e-06, "loss": 0.6944, "step": 495 }, { "epoch": 0.20644095788604458, "grad_norm": 2.8340105200706507, "learning_rate": 9.660674101575188e-06, "loss": 0.6959, "step": 500 }, { "epoch": 0.20850536746490503, "grad_norm": 2.6749836979567325, "learning_rate": 9.647501181252228e-06, "loss": 0.7181, "step": 505 }, { "epoch": 0.21056977704376548, "grad_norm": 2.871887495235821, "learning_rate": 9.634086746589597e-06, "loss": 0.6855, "step": 510 }, { "epoch": 0.21263418662262593, "grad_norm": 3.022510499989922, "learning_rate": 9.62043149468842e-06, "loss": 0.6786, "step": 515 }, { "epoch": 0.21469859620148637, "grad_norm": 2.798631786402729, "learning_rate": 9.60653613516426e-06, "loss": 0.6878, "step": 520 }, { "epoch": 0.21676300578034682, "grad_norm": 2.7665584034808557, "learning_rate": 9.592401390110224e-06, "loss": 0.6689, "step": 525 }, { "epoch": 0.21882741535920727, "grad_norm": 2.6972597849127995, "learning_rate": 9.578027994059448e-06, "loss": 0.6938, "step": 530 }, { "epoch": 0.22089182493806772, "grad_norm": 2.876624388604322, "learning_rate": 9.563416693946927e-06, "loss": 0.6831, "step": 535 }, { "epoch": 0.22295623451692817, "grad_norm": 2.760614779991045, "learning_rate": 9.548568249070688e-06, "loss": 0.6721, "step": 540 }, { "epoch": 0.2250206440957886, "grad_norm": 2.711332679097759, "learning_rate": 9.533483431052347e-06, "loss": 0.6734, "step": 545 }, { "epoch": 0.22708505367464904, "grad_norm": 2.9419038213760573, "learning_rate": 9.518163023797001e-06, "loss": 0.6826, "step": 550 }, { "epoch": 0.2291494632535095, "grad_norm": 2.708449916455307, "learning_rate": 9.502607823452492e-06, "loss": 0.6909, "step": 555 }, { "epoch": 0.23121387283236994, "grad_norm": 2.6744723555279712, "learning_rate": 9.486818638368037e-06, "loss": 0.6727, "step": 560 }, { "epoch": 0.2332782824112304, "grad_norm": 2.740519766192663, "learning_rate": 9.47079628905222e-06, "loss": 0.6643, "step": 565 }, { "epoch": 0.23534269199009084, "grad_norm": 2.6127044855477544, "learning_rate": 9.45454160813035e-06, "loss": 0.641, "step": 570 }, { "epoch": 0.23740710156895128, "grad_norm": 2.5898360039465547, "learning_rate": 9.438055440301198e-06, "loss": 0.654, "step": 575 }, { "epoch": 0.23947151114781173, "grad_norm": 2.594315428105211, "learning_rate": 9.421338642293096e-06, "loss": 0.6612, "step": 580 }, { "epoch": 0.24153592072667218, "grad_norm": 2.675279309210325, "learning_rate": 9.404392082819418e-06, "loss": 0.6835, "step": 585 }, { "epoch": 0.24360033030553263, "grad_norm": 2.720481221371884, "learning_rate": 9.387216642533436e-06, "loss": 0.638, "step": 590 }, { "epoch": 0.24566473988439305, "grad_norm": 2.7274539164197567, "learning_rate": 9.369813213982557e-06, "loss": 0.6742, "step": 595 }, { "epoch": 0.2477291494632535, "grad_norm": 2.881159963630694, "learning_rate": 9.352182701561932e-06, "loss": 0.643, "step": 600 }, { "epoch": 0.24979355904211395, "grad_norm": 2.672785271513205, "learning_rate": 9.334326021467473e-06, "loss": 0.6518, "step": 605 }, { "epoch": 0.2518579686209744, "grad_norm": 2.6806455710314365, "learning_rate": 9.316244101648229e-06, "loss": 0.644, "step": 610 }, { "epoch": 0.2539223781998348, "grad_norm": 2.5531072290720296, "learning_rate": 9.297937881758165e-06, "loss": 0.6738, "step": 615 }, { "epoch": 0.25598678777869527, "grad_norm": 2.8562519058356335, "learning_rate": 9.279408313107342e-06, "loss": 0.6497, "step": 620 }, { "epoch": 0.2580511973575557, "grad_norm": 2.9118129566241273, "learning_rate": 9.260656358612461e-06, "loss": 0.6544, "step": 625 }, { "epoch": 0.26011560693641617, "grad_norm": 2.6582669873812117, "learning_rate": 9.241682992746847e-06, "loss": 0.6492, "step": 630 }, { "epoch": 0.2621800165152766, "grad_norm": 3.135088950824475, "learning_rate": 9.222489201489792e-06, "loss": 0.6324, "step": 635 }, { "epoch": 0.26424442609413706, "grad_norm": 2.5674033607777282, "learning_rate": 9.203075982275323e-06, "loss": 0.6461, "step": 640 }, { "epoch": 0.2663088356729975, "grad_norm": 2.615329482313224, "learning_rate": 9.18344434394037e-06, "loss": 0.6322, "step": 645 }, { "epoch": 0.26837324525185796, "grad_norm": 2.6662637432037144, "learning_rate": 9.163595306672337e-06, "loss": 0.6245, "step": 650 }, { "epoch": 0.2704376548307184, "grad_norm": 2.788831071229617, "learning_rate": 9.143529901956089e-06, "loss": 0.6395, "step": 655 }, { "epoch": 0.27250206440957886, "grad_norm": 2.7354638041171344, "learning_rate": 9.12324917252035e-06, "loss": 0.6239, "step": 660 }, { "epoch": 0.2745664739884393, "grad_norm": 2.6074213157412682, "learning_rate": 9.102754172283513e-06, "loss": 0.6146, "step": 665 }, { "epoch": 0.27663088356729976, "grad_norm": 2.6642140176865485, "learning_rate": 9.082045966298871e-06, "loss": 0.6207, "step": 670 }, { "epoch": 0.2786952931461602, "grad_norm": 2.7423993871208547, "learning_rate": 9.061125630699276e-06, "loss": 0.6244, "step": 675 }, { "epoch": 0.28075970272502065, "grad_norm": 2.639661943202539, "learning_rate": 9.039994252641214e-06, "loss": 0.6212, "step": 680 }, { "epoch": 0.2828241123038811, "grad_norm": 2.7917609886441443, "learning_rate": 9.018652930248299e-06, "loss": 0.6279, "step": 685 }, { "epoch": 0.28488852188274155, "grad_norm": 2.629871903545716, "learning_rate": 8.997102772554226e-06, "loss": 0.6229, "step": 690 }, { "epoch": 0.286952931461602, "grad_norm": 2.625664182045412, "learning_rate": 8.975344899445119e-06, "loss": 0.6177, "step": 695 }, { "epoch": 0.28901734104046245, "grad_norm": 2.80144553747866, "learning_rate": 8.953380441601352e-06, "loss": 0.6225, "step": 700 }, { "epoch": 0.2910817506193229, "grad_norm": 2.672640679384628, "learning_rate": 8.931210540438778e-06, "loss": 0.6169, "step": 705 }, { "epoch": 0.2931461601981833, "grad_norm": 2.709692648262891, "learning_rate": 8.908836348049421e-06, "loss": 0.6198, "step": 710 }, { "epoch": 0.29521056977704374, "grad_norm": 2.681528317767069, "learning_rate": 8.886259027141603e-06, "loss": 0.6149, "step": 715 }, { "epoch": 0.2972749793559042, "grad_norm": 3.03293644975451, "learning_rate": 8.863479750979523e-06, "loss": 0.6048, "step": 720 }, { "epoch": 0.29933938893476464, "grad_norm": 2.670868153191048, "learning_rate": 8.840499703322286e-06, "loss": 0.6244, "step": 725 }, { "epoch": 0.3014037985136251, "grad_norm": 2.7608395704880646, "learning_rate": 8.817320078362388e-06, "loss": 0.6035, "step": 730 }, { "epoch": 0.30346820809248554, "grad_norm": 2.62588384982977, "learning_rate": 8.793942080663658e-06, "loss": 0.6172, "step": 735 }, { "epoch": 0.305532617671346, "grad_norm": 2.5889068256927006, "learning_rate": 8.77036692509866e-06, "loss": 0.6128, "step": 740 }, { "epoch": 0.30759702725020643, "grad_norm": 2.8092080143766216, "learning_rate": 8.746595836785561e-06, "loss": 0.6218, "step": 745 }, { "epoch": 0.3096614368290669, "grad_norm": 2.6419072436310205, "learning_rate": 8.722630051024467e-06, "loss": 0.5877, "step": 750 }, { "epoch": 0.31172584640792733, "grad_norm": 2.4997349328580363, "learning_rate": 8.698470813233223e-06, "loss": 0.6091, "step": 755 }, { "epoch": 0.3137902559867878, "grad_norm": 2.6466710278909074, "learning_rate": 8.674119378882707e-06, "loss": 0.6087, "step": 760 }, { "epoch": 0.31585466556564823, "grad_norm": 2.9297154437714363, "learning_rate": 8.649577013431569e-06, "loss": 0.592, "step": 765 }, { "epoch": 0.3179190751445087, "grad_norm": 2.549997409403058, "learning_rate": 8.624844992260483e-06, "loss": 0.6023, "step": 770 }, { "epoch": 0.3199834847233691, "grad_norm": 2.7080008761741827, "learning_rate": 8.599924600605865e-06, "loss": 0.6087, "step": 775 }, { "epoch": 0.3220478943022296, "grad_norm": 2.567143331934158, "learning_rate": 8.574817133493085e-06, "loss": 0.5847, "step": 780 }, { "epoch": 0.32411230388109, "grad_norm": 3.0113192932757418, "learning_rate": 8.549523895669167e-06, "loss": 0.5908, "step": 785 }, { "epoch": 0.3261767134599505, "grad_norm": 2.7819521209578206, "learning_rate": 8.52404620153499e-06, "loss": 0.5953, "step": 790 }, { "epoch": 0.3282411230388109, "grad_norm": 2.747216748021648, "learning_rate": 8.498385375076979e-06, "loss": 0.6114, "step": 795 }, { "epoch": 0.33030553261767137, "grad_norm": 2.8476886670194523, "learning_rate": 8.472542749798302e-06, "loss": 0.5923, "step": 800 }, { "epoch": 0.33236994219653176, "grad_norm": 3.078148326221217, "learning_rate": 8.44651966864958e-06, "loss": 0.5854, "step": 805 }, { "epoch": 0.3344343517753922, "grad_norm": 2.7069355571781553, "learning_rate": 8.420317483959086e-06, "loss": 0.5963, "step": 810 }, { "epoch": 0.33649876135425266, "grad_norm": 2.651998023029158, "learning_rate": 8.39393755736248e-06, "loss": 0.578, "step": 815 }, { "epoch": 0.3385631709331131, "grad_norm": 2.6251615353567117, "learning_rate": 8.367381259732042e-06, "loss": 0.587, "step": 820 }, { "epoch": 0.34062758051197356, "grad_norm": 2.5308341360997595, "learning_rate": 8.340649971105444e-06, "loss": 0.5732, "step": 825 }, { "epoch": 0.342691990090834, "grad_norm": 2.823077213728183, "learning_rate": 8.313745080614017e-06, "loss": 0.5672, "step": 830 }, { "epoch": 0.34475639966969446, "grad_norm": 2.659469137057744, "learning_rate": 8.286667986410578e-06, "loss": 0.5858, "step": 835 }, { "epoch": 0.3468208092485549, "grad_norm": 2.845018948853141, "learning_rate": 8.259420095596766e-06, "loss": 0.5662, "step": 840 }, { "epoch": 0.34888521882741536, "grad_norm": 2.6028478895400116, "learning_rate": 8.232002824149917e-06, "loss": 0.5606, "step": 845 }, { "epoch": 0.3509496284062758, "grad_norm": 2.664709732227379, "learning_rate": 8.204417596849492e-06, "loss": 0.5807, "step": 850 }, { "epoch": 0.35301403798513625, "grad_norm": 2.6109659834823478, "learning_rate": 8.176665847203023e-06, "loss": 0.5678, "step": 855 }, { "epoch": 0.3550784475639967, "grad_norm": 2.576771155175357, "learning_rate": 8.148749017371627e-06, "loss": 0.5656, "step": 860 }, { "epoch": 0.35714285714285715, "grad_norm": 2.6688801593883817, "learning_rate": 8.120668558095056e-06, "loss": 0.5749, "step": 865 }, { "epoch": 0.3592072667217176, "grad_norm": 2.801881409304358, "learning_rate": 8.092425928616317e-06, "loss": 0.5603, "step": 870 }, { "epoch": 0.36127167630057805, "grad_norm": 2.7158675577145512, "learning_rate": 8.064022596605825e-06, "loss": 0.5878, "step": 875 }, { "epoch": 0.3633360858794385, "grad_norm": 2.7112475864205208, "learning_rate": 8.03546003808515e-06, "loss": 0.5697, "step": 880 }, { "epoch": 0.36540049545829895, "grad_norm": 2.757081302105068, "learning_rate": 8.006739737350303e-06, "loss": 0.5549, "step": 885 }, { "epoch": 0.3674649050371594, "grad_norm": 2.6798695533176, "learning_rate": 7.977863186894598e-06, "loss": 0.5621, "step": 890 }, { "epoch": 0.36952931461601984, "grad_norm": 2.901794116790705, "learning_rate": 7.948831887331113e-06, "loss": 0.5531, "step": 895 }, { "epoch": 0.37159372419488024, "grad_norm": 2.6209628611019333, "learning_rate": 7.91964734731468e-06, "loss": 0.5529, "step": 900 }, { "epoch": 0.3736581337737407, "grad_norm": 2.631776999408232, "learning_rate": 7.890311083463511e-06, "loss": 0.5341, "step": 905 }, { "epoch": 0.37572254335260113, "grad_norm": 2.6460839052940424, "learning_rate": 7.860824620280367e-06, "loss": 0.5589, "step": 910 }, { "epoch": 0.3777869529314616, "grad_norm": 2.7027858769480604, "learning_rate": 7.831189490073343e-06, "loss": 0.5662, "step": 915 }, { "epoch": 0.37985136251032203, "grad_norm": 2.686204866691535, "learning_rate": 7.80140723287624e-06, "loss": 0.5557, "step": 920 }, { "epoch": 0.3819157720891825, "grad_norm": 2.537669294604416, "learning_rate": 7.771479396368533e-06, "loss": 0.5392, "step": 925 }, { "epoch": 0.38398018166804293, "grad_norm": 2.634245727933428, "learning_rate": 7.741407535794939e-06, "loss": 0.5476, "step": 930 }, { "epoch": 0.3860445912469034, "grad_norm": 2.63998365984595, "learning_rate": 7.711193213884602e-06, "loss": 0.5411, "step": 935 }, { "epoch": 0.38810900082576383, "grad_norm": 2.5066641406139163, "learning_rate": 7.680838000769889e-06, "loss": 0.5436, "step": 940 }, { "epoch": 0.3901734104046243, "grad_norm": 2.4698510563626384, "learning_rate": 7.650343473904776e-06, "loss": 0.536, "step": 945 }, { "epoch": 0.3922378199834847, "grad_norm": 2.5102808206533753, "learning_rate": 7.619711217982899e-06, "loss": 0.5391, "step": 950 }, { "epoch": 0.3943022295623452, "grad_norm": 2.648139467398325, "learning_rate": 7.5889428248551854e-06, "loss": 0.5368, "step": 955 }, { "epoch": 0.3963666391412056, "grad_norm": 2.6648738415553765, "learning_rate": 7.558039893447131e-06, "loss": 0.5496, "step": 960 }, { "epoch": 0.39843104872006607, "grad_norm": 2.6157509498551526, "learning_rate": 7.527004029675722e-06, "loss": 0.5318, "step": 965 }, { "epoch": 0.4004954582989265, "grad_norm": 2.5975829607560574, "learning_rate": 7.495836846365965e-06, "loss": 0.5352, "step": 970 }, { "epoch": 0.40255986787778697, "grad_norm": 2.92450937390704, "learning_rate": 7.464539963167091e-06, "loss": 0.5177, "step": 975 }, { "epoch": 0.4046242774566474, "grad_norm": 2.7355952122295264, "learning_rate": 7.433115006468373e-06, "loss": 0.5322, "step": 980 }, { "epoch": 0.40668868703550787, "grad_norm": 2.5829736114469184, "learning_rate": 7.40156360931462e-06, "loss": 0.5298, "step": 985 }, { "epoch": 0.4087530966143683, "grad_norm": 2.705958846647064, "learning_rate": 7.369887411321305e-06, "loss": 0.5477, "step": 990 }, { "epoch": 0.4108175061932287, "grad_norm": 2.4846959730474145, "learning_rate": 7.338088058589362e-06, "loss": 0.5324, "step": 995 }, { "epoch": 0.41288191577208916, "grad_norm": 2.5252160694956762, "learning_rate": 7.306167203619647e-06, "loss": 0.5335, "step": 1000 }, { "epoch": 0.4149463253509496, "grad_norm": 2.509240890701763, "learning_rate": 7.27412650522706e-06, "loss": 0.5219, "step": 1005 }, { "epoch": 0.41701073492981006, "grad_norm": 2.567378202198639, "learning_rate": 7.241967628454345e-06, "loss": 0.5142, "step": 1010 }, { "epoch": 0.4190751445086705, "grad_norm": 2.479116521394592, "learning_rate": 7.20969224448556e-06, "loss": 0.5198, "step": 1015 }, { "epoch": 0.42113955408753095, "grad_norm": 2.456783739959803, "learning_rate": 7.1773020305592355e-06, "loss": 0.5141, "step": 1020 }, { "epoch": 0.4232039636663914, "grad_norm": 2.6855797799019356, "learning_rate": 7.1447986698812115e-06, "loss": 0.5307, "step": 1025 }, { "epoch": 0.42526837324525185, "grad_norm": 2.5323199215752203, "learning_rate": 7.112183851537166e-06, "loss": 0.5256, "step": 1030 }, { "epoch": 0.4273327828241123, "grad_norm": 2.735418738176186, "learning_rate": 7.079459270404842e-06, "loss": 0.5112, "step": 1035 }, { "epoch": 0.42939719240297275, "grad_norm": 2.6703449875287735, "learning_rate": 7.046626627065974e-06, "loss": 0.5099, "step": 1040 }, { "epoch": 0.4314616019818332, "grad_norm": 2.4631394786044747, "learning_rate": 7.013687627717905e-06, "loss": 0.4978, "step": 1045 }, { "epoch": 0.43352601156069365, "grad_norm": 2.8328555379260196, "learning_rate": 6.980643984084927e-06, "loss": 0.5179, "step": 1050 }, { "epoch": 0.4355904211395541, "grad_norm": 2.504430665948721, "learning_rate": 6.947497413329333e-06, "loss": 0.5199, "step": 1055 }, { "epoch": 0.43765483071841454, "grad_norm": 2.5711442943268663, "learning_rate": 6.914249637962174e-06, "loss": 0.5133, "step": 1060 }, { "epoch": 0.439719240297275, "grad_norm": 2.5576807334846747, "learning_rate": 6.880902385753749e-06, "loss": 0.51, "step": 1065 }, { "epoch": 0.44178364987613544, "grad_norm": 2.4378114750971833, "learning_rate": 6.847457389643823e-06, "loss": 0.4966, "step": 1070 }, { "epoch": 0.4438480594549959, "grad_norm": 2.57922332477375, "learning_rate": 6.8139163876515694e-06, "loss": 0.5018, "step": 1075 }, { "epoch": 0.44591246903385634, "grad_norm": 2.4923563722799145, "learning_rate": 6.780281122785243e-06, "loss": 0.5089, "step": 1080 }, { "epoch": 0.4479768786127168, "grad_norm": 2.380488675820629, "learning_rate": 6.74655334295162e-06, "loss": 0.4969, "step": 1085 }, { "epoch": 0.4500412881915772, "grad_norm": 2.622744707083321, "learning_rate": 6.712734800865146e-06, "loss": 0.5006, "step": 1090 }, { "epoch": 0.45210569777043763, "grad_norm": 2.5279298037448514, "learning_rate": 6.67882725395687e-06, "loss": 0.4919, "step": 1095 }, { "epoch": 0.4541701073492981, "grad_norm": 2.5696353167432076, "learning_rate": 6.644832464283105e-06, "loss": 0.5006, "step": 1100 }, { "epoch": 0.45623451692815853, "grad_norm": 2.6472120693562675, "learning_rate": 6.6107521984338654e-06, "loss": 0.5222, "step": 1105 }, { "epoch": 0.458298926507019, "grad_norm": 2.5641259833492915, "learning_rate": 6.576588227441063e-06, "loss": 0.5021, "step": 1110 }, { "epoch": 0.4603633360858794, "grad_norm": 2.508674684822535, "learning_rate": 6.5423423266864715e-06, "loss": 0.5019, "step": 1115 }, { "epoch": 0.4624277456647399, "grad_norm": 2.5803067400324546, "learning_rate": 6.508016275809472e-06, "loss": 0.5116, "step": 1120 }, { "epoch": 0.4644921552436003, "grad_norm": 2.4783246238450296, "learning_rate": 6.473611858614557e-06, "loss": 0.4939, "step": 1125 }, { "epoch": 0.4665565648224608, "grad_norm": 2.3961470840043986, "learning_rate": 6.439130862978653e-06, "loss": 0.4964, "step": 1130 }, { "epoch": 0.4686209744013212, "grad_norm": 2.523268688947089, "learning_rate": 6.4045750807581875e-06, "loss": 0.4878, "step": 1135 }, { "epoch": 0.47068538398018167, "grad_norm": 2.464908065215536, "learning_rate": 6.369946307695994e-06, "loss": 0.5004, "step": 1140 }, { "epoch": 0.4727497935590421, "grad_norm": 2.4978854529114556, "learning_rate": 6.335246343327983e-06, "loss": 0.4928, "step": 1145 }, { "epoch": 0.47481420313790257, "grad_norm": 2.5910220265362405, "learning_rate": 6.300476990889622e-06, "loss": 0.4859, "step": 1150 }, { "epoch": 0.476878612716763, "grad_norm": 2.6474515998209314, "learning_rate": 6.2656400572222374e-06, "loss": 0.4968, "step": 1155 }, { "epoch": 0.47894302229562347, "grad_norm": 2.5111508875593627, "learning_rate": 6.230737352679115e-06, "loss": 0.4825, "step": 1160 }, { "epoch": 0.4810074318744839, "grad_norm": 2.6001140755401067, "learning_rate": 6.1957706910314195e-06, "loss": 0.4903, "step": 1165 }, { "epoch": 0.48307184145334436, "grad_norm": 2.4405093084108853, "learning_rate": 6.160741889373948e-06, "loss": 0.4738, "step": 1170 }, { "epoch": 0.4851362510322048, "grad_norm": 2.6122999506914817, "learning_rate": 6.12565276803069e-06, "loss": 0.4851, "step": 1175 }, { "epoch": 0.48720066061106526, "grad_norm": 2.5491172182391284, "learning_rate": 6.090505150460239e-06, "loss": 0.4969, "step": 1180 }, { "epoch": 0.48926507018992565, "grad_norm": 2.5615374385855256, "learning_rate": 6.055300863161035e-06, "loss": 0.4778, "step": 1185 }, { "epoch": 0.4913294797687861, "grad_norm": 2.4469903993231514, "learning_rate": 6.020041735576444e-06, "loss": 0.4787, "step": 1190 }, { "epoch": 0.49339388934764655, "grad_norm": 2.4860974782023346, "learning_rate": 5.9847295999996835e-06, "loss": 0.4914, "step": 1195 }, { "epoch": 0.495458298926507, "grad_norm": 2.4900103589842395, "learning_rate": 5.949366291478621e-06, "loss": 0.4906, "step": 1200 }, { "epoch": 0.49752270850536745, "grad_norm": 2.5283147963068964, "learning_rate": 5.9139536477203945e-06, "loss": 0.4759, "step": 1205 }, { "epoch": 0.4995871180842279, "grad_norm": 2.573137486004109, "learning_rate": 5.878493508995924e-06, "loss": 0.4768, "step": 1210 }, { "epoch": 0.5016515276630884, "grad_norm": 2.579476813837634, "learning_rate": 5.842987718044275e-06, "loss": 0.4679, "step": 1215 }, { "epoch": 0.5037159372419489, "grad_norm": 2.497492462831134, "learning_rate": 5.807438119976904e-06, "loss": 0.4698, "step": 1220 }, { "epoch": 0.5057803468208093, "grad_norm": 2.590655156263669, "learning_rate": 5.771846562181761e-06, "loss": 0.4728, "step": 1225 }, { "epoch": 0.5078447563996696, "grad_norm": 2.5177297962486493, "learning_rate": 5.736214894227304e-06, "loss": 0.473, "step": 1230 }, { "epoch": 0.5099091659785301, "grad_norm": 2.6021365054723775, "learning_rate": 5.700544967766373e-06, "loss": 0.4753, "step": 1235 }, { "epoch": 0.5119735755573905, "grad_norm": 2.604951347166713, "learning_rate": 5.6648386364399645e-06, "loss": 0.47, "step": 1240 }, { "epoch": 0.514037985136251, "grad_norm": 2.4157020679627794, "learning_rate": 5.629097755780913e-06, "loss": 0.4711, "step": 1245 }, { "epoch": 0.5161023947151114, "grad_norm": 2.3659395254457167, "learning_rate": 5.59332418311746e-06, "loss": 0.4773, "step": 1250 }, { "epoch": 0.5181668042939719, "grad_norm": 2.499361991412955, "learning_rate": 5.557519777476733e-06, "loss": 0.4527, "step": 1255 }, { "epoch": 0.5202312138728323, "grad_norm": 2.5585105365426513, "learning_rate": 5.521686399488145e-06, "loss": 0.4573, "step": 1260 }, { "epoch": 0.5222956234516928, "grad_norm": 2.4814508055210567, "learning_rate": 5.485825911286697e-06, "loss": 0.4691, "step": 1265 }, { "epoch": 0.5243600330305532, "grad_norm": 2.6081291748684716, "learning_rate": 5.449940176416213e-06, "loss": 0.4526, "step": 1270 }, { "epoch": 0.5264244426094137, "grad_norm": 2.587126119981003, "learning_rate": 5.414031059732502e-06, "loss": 0.4605, "step": 1275 }, { "epoch": 0.5284888521882741, "grad_norm": 2.5458768564010765, "learning_rate": 5.378100427306439e-06, "loss": 0.4467, "step": 1280 }, { "epoch": 0.5305532617671346, "grad_norm": 2.422344978356131, "learning_rate": 5.3421501463269976e-06, "loss": 0.4563, "step": 1285 }, { "epoch": 0.532617671345995, "grad_norm": 2.481246363328885, "learning_rate": 5.30618208500422e-06, "loss": 0.4496, "step": 1290 }, { "epoch": 0.5346820809248555, "grad_norm": 2.5964265890691274, "learning_rate": 5.270198112472128e-06, "loss": 0.4487, "step": 1295 }, { "epoch": 0.5367464905037159, "grad_norm": 2.375375823310664, "learning_rate": 5.234200098691597e-06, "loss": 0.4492, "step": 1300 }, { "epoch": 0.5388109000825764, "grad_norm": 2.482312531712164, "learning_rate": 5.19818991435317e-06, "loss": 0.4524, "step": 1305 }, { "epoch": 0.5408753096614368, "grad_norm": 2.6268917122173154, "learning_rate": 5.162169430779861e-06, "loss": 0.4343, "step": 1310 }, { "epoch": 0.5429397192402973, "grad_norm": 2.543083790336233, "learning_rate": 5.126140519829888e-06, "loss": 0.4738, "step": 1315 }, { "epoch": 0.5450041288191577, "grad_norm": 2.4547121862030865, "learning_rate": 5.090105053799418e-06, "loss": 0.4604, "step": 1320 }, { "epoch": 0.5470685383980182, "grad_norm": 2.5806078709462295, "learning_rate": 5.054064905325262e-06, "loss": 0.4571, "step": 1325 }, { "epoch": 0.5491329479768786, "grad_norm": 2.538503365518751, "learning_rate": 5.018021947287556e-06, "loss": 0.4432, "step": 1330 }, { "epoch": 0.5511973575557391, "grad_norm": 2.4205037183193157, "learning_rate": 4.981978052712446e-06, "loss": 0.4545, "step": 1335 }, { "epoch": 0.5532617671345995, "grad_norm": 2.5910957054221817, "learning_rate": 4.9459350946747405e-06, "loss": 0.4584, "step": 1340 }, { "epoch": 0.55532617671346, "grad_norm": 2.423593643872453, "learning_rate": 4.909894946200583e-06, "loss": 0.454, "step": 1345 }, { "epoch": 0.5573905862923204, "grad_norm": 2.491439793223575, "learning_rate": 4.873859480170113e-06, "loss": 0.4535, "step": 1350 }, { "epoch": 0.5594549958711809, "grad_norm": 2.6251023268801106, "learning_rate": 4.83783056922014e-06, "loss": 0.4475, "step": 1355 }, { "epoch": 0.5615194054500413, "grad_norm": 2.379012672081149, "learning_rate": 4.801810085646831e-06, "loss": 0.4427, "step": 1360 }, { "epoch": 0.5635838150289018, "grad_norm": 2.438307211574731, "learning_rate": 4.765799901308405e-06, "loss": 0.4392, "step": 1365 }, { "epoch": 0.5656482246077622, "grad_norm": 2.4743966890129534, "learning_rate": 4.729801887527872e-06, "loss": 0.4376, "step": 1370 }, { "epoch": 0.5677126341866227, "grad_norm": 2.4438402043185636, "learning_rate": 4.6938179149957815e-06, "loss": 0.4448, "step": 1375 }, { "epoch": 0.5697770437654831, "grad_norm": 2.3233363056641587, "learning_rate": 4.657849853673004e-06, "loss": 0.4332, "step": 1380 }, { "epoch": 0.5718414533443436, "grad_norm": 2.3513159156694536, "learning_rate": 4.621899572693564e-06, "loss": 0.4286, "step": 1385 }, { "epoch": 0.573905862923204, "grad_norm": 2.452790907750448, "learning_rate": 4.585968940267499e-06, "loss": 0.4329, "step": 1390 }, { "epoch": 0.5759702725020645, "grad_norm": 2.4876165011978544, "learning_rate": 4.550059823583788e-06, "loss": 0.4368, "step": 1395 }, { "epoch": 0.5780346820809249, "grad_norm": 2.4110190508295926, "learning_rate": 4.514174088713305e-06, "loss": 0.4294, "step": 1400 }, { "epoch": 0.5800990916597853, "grad_norm": 2.4790636104654724, "learning_rate": 4.478313600511856e-06, "loss": 0.438, "step": 1405 }, { "epoch": 0.5821635012386458, "grad_norm": 2.5304633815274062, "learning_rate": 4.442480222523269e-06, "loss": 0.4383, "step": 1410 }, { "epoch": 0.5842279108175062, "grad_norm": 2.365309103378621, "learning_rate": 4.4066758168825414e-06, "loss": 0.4383, "step": 1415 }, { "epoch": 0.5862923203963666, "grad_norm": 2.4670833952729687, "learning_rate": 4.370902244219088e-06, "loss": 0.4389, "step": 1420 }, { "epoch": 0.588356729975227, "grad_norm": 2.3390938676909734, "learning_rate": 4.335161363560037e-06, "loss": 0.4291, "step": 1425 }, { "epoch": 0.5904211395540875, "grad_norm": 2.362644766829775, "learning_rate": 4.299455032233631e-06, "loss": 0.4288, "step": 1430 }, { "epoch": 0.5924855491329479, "grad_norm": 2.445730007800526, "learning_rate": 4.263785105772696e-06, "loss": 0.4335, "step": 1435 }, { "epoch": 0.5945499587118084, "grad_norm": 2.303973634970439, "learning_rate": 4.22815343781824e-06, "loss": 0.4366, "step": 1440 }, { "epoch": 0.5966143682906688, "grad_norm": 2.391887494831502, "learning_rate": 4.192561880023099e-06, "loss": 0.4261, "step": 1445 }, { "epoch": 0.5986787778695293, "grad_norm": 2.293302291662098, "learning_rate": 4.157012281955726e-06, "loss": 0.424, "step": 1450 }, { "epoch": 0.6007431874483897, "grad_norm": 2.3712447663578655, "learning_rate": 4.1215064910040795e-06, "loss": 0.4314, "step": 1455 }, { "epoch": 0.6028075970272502, "grad_norm": 2.381812200033987, "learning_rate": 4.086046352279606e-06, "loss": 0.4227, "step": 1460 }, { "epoch": 0.6048720066061106, "grad_norm": 2.4743735981326114, "learning_rate": 4.050633708521381e-06, "loss": 0.4158, "step": 1465 }, { "epoch": 0.6069364161849711, "grad_norm": 2.396322395044398, "learning_rate": 4.015270400000317e-06, "loss": 0.42, "step": 1470 }, { "epoch": 0.6090008257638315, "grad_norm": 2.565114051952279, "learning_rate": 3.979958264423558e-06, "loss": 0.4155, "step": 1475 }, { "epoch": 0.611065235342692, "grad_norm": 2.4934554445972954, "learning_rate": 3.944699136838965e-06, "loss": 0.4126, "step": 1480 }, { "epoch": 0.6131296449215524, "grad_norm": 2.4162398419293187, "learning_rate": 3.909494849539761e-06, "loss": 0.4213, "step": 1485 }, { "epoch": 0.6151940545004129, "grad_norm": 2.464180248251729, "learning_rate": 3.874347231969312e-06, "loss": 0.4235, "step": 1490 }, { "epoch": 0.6172584640792733, "grad_norm": 2.4941582294474434, "learning_rate": 3.839258110626053e-06, "loss": 0.4307, "step": 1495 }, { "epoch": 0.6193228736581338, "grad_norm": 2.276754213405921, "learning_rate": 3.8042293089685813e-06, "loss": 0.4065, "step": 1500 }, { "epoch": 0.6213872832369942, "grad_norm": 2.324659425092549, "learning_rate": 3.769262647320886e-06, "loss": 0.4179, "step": 1505 }, { "epoch": 0.6234516928158547, "grad_norm": 2.3013999943076815, "learning_rate": 3.7343599427777634e-06, "loss": 0.4157, "step": 1510 }, { "epoch": 0.6255161023947151, "grad_norm": 2.3853366654087633, "learning_rate": 3.69952300911038e-06, "loss": 0.4087, "step": 1515 }, { "epoch": 0.6275805119735756, "grad_norm": 2.3274220274703206, "learning_rate": 3.6647536566720186e-06, "loss": 0.4203, "step": 1520 }, { "epoch": 0.629644921552436, "grad_norm": 2.4392433402643947, "learning_rate": 3.630053692304005e-06, "loss": 0.4091, "step": 1525 }, { "epoch": 0.6317093311312965, "grad_norm": 2.4428540121878894, "learning_rate": 3.5954249192418133e-06, "loss": 0.4255, "step": 1530 }, { "epoch": 0.6337737407101569, "grad_norm": 2.3810770043048692, "learning_rate": 3.560869137021349e-06, "loss": 0.4103, "step": 1535 }, { "epoch": 0.6358381502890174, "grad_norm": 2.301863999851205, "learning_rate": 3.526388141385444e-06, "loss": 0.4155, "step": 1540 }, { "epoch": 0.6379025598678778, "grad_norm": 2.400363519907276, "learning_rate": 3.4919837241905308e-06, "loss": 0.4224, "step": 1545 }, { "epoch": 0.6399669694467383, "grad_norm": 2.477498090820749, "learning_rate": 3.4576576733135284e-06, "loss": 0.3938, "step": 1550 }, { "epoch": 0.6420313790255987, "grad_norm": 2.707387721012179, "learning_rate": 3.4234117725589385e-06, "loss": 0.4111, "step": 1555 }, { "epoch": 0.6440957886044592, "grad_norm": 2.341459106750255, "learning_rate": 3.3892478015661362e-06, "loss": 0.4139, "step": 1560 }, { "epoch": 0.6461601981833196, "grad_norm": 2.4202740830209137, "learning_rate": 3.355167535716897e-06, "loss": 0.4127, "step": 1565 }, { "epoch": 0.64822460776218, "grad_norm": 2.3816575196973577, "learning_rate": 3.321172746043131e-06, "loss": 0.4002, "step": 1570 }, { "epoch": 0.6502890173410405, "grad_norm": 2.454435864594131, "learning_rate": 3.2872651991348548e-06, "loss": 0.403, "step": 1575 }, { "epoch": 0.652353426919901, "grad_norm": 2.591163202406022, "learning_rate": 3.253446657048382e-06, "loss": 0.4072, "step": 1580 }, { "epoch": 0.6544178364987614, "grad_norm": 2.4282921305894676, "learning_rate": 3.2197188772147593e-06, "loss": 0.3869, "step": 1585 }, { "epoch": 0.6564822460776218, "grad_norm": 2.4682745998960827, "learning_rate": 3.186083612348434e-06, "loss": 0.4028, "step": 1590 }, { "epoch": 0.6585466556564823, "grad_norm": 2.376999029803135, "learning_rate": 3.1525426103561774e-06, "loss": 0.4064, "step": 1595 }, { "epoch": 0.6606110652353427, "grad_norm": 2.370838413822066, "learning_rate": 3.119097614246252e-06, "loss": 0.3978, "step": 1600 }, { "epoch": 0.6626754748142032, "grad_norm": 2.362260369144417, "learning_rate": 3.0857503620378284e-06, "loss": 0.3905, "step": 1605 }, { "epoch": 0.6647398843930635, "grad_norm": 2.421231306606393, "learning_rate": 3.052502586670669e-06, "loss": 0.4053, "step": 1610 }, { "epoch": 0.666804293971924, "grad_norm": 2.395970410200359, "learning_rate": 3.0193560159150747e-06, "loss": 0.3943, "step": 1615 }, { "epoch": 0.6688687035507844, "grad_norm": 2.5057121725436917, "learning_rate": 2.986312372282097e-06, "loss": 0.4182, "step": 1620 }, { "epoch": 0.6709331131296449, "grad_norm": 2.3776161687385207, "learning_rate": 2.9533733729340274e-06, "loss": 0.4077, "step": 1625 }, { "epoch": 0.6729975227085053, "grad_norm": 2.5915693629163217, "learning_rate": 2.920540729595159e-06, "loss": 0.4085, "step": 1630 }, { "epoch": 0.6750619322873658, "grad_norm": 2.443160283840937, "learning_rate": 2.8878161484628364e-06, "loss": 0.4028, "step": 1635 }, { "epoch": 0.6771263418662262, "grad_norm": 2.420398270818125, "learning_rate": 2.8552013301187898e-06, "loss": 0.3972, "step": 1640 }, { "epoch": 0.6791907514450867, "grad_norm": 2.4196367974051265, "learning_rate": 2.8226979694407657e-06, "loss": 0.4067, "step": 1645 }, { "epoch": 0.6812551610239471, "grad_norm": 2.438274119472032, "learning_rate": 2.790307755514442e-06, "loss": 0.4032, "step": 1650 }, { "epoch": 0.6833195706028076, "grad_norm": 3.0023110501423944, "learning_rate": 2.7580323715456564e-06, "loss": 0.4121, "step": 1655 }, { "epoch": 0.685383980181668, "grad_norm": 2.3597852522402474, "learning_rate": 2.7258734947729428e-06, "loss": 0.3938, "step": 1660 }, { "epoch": 0.6874483897605285, "grad_norm": 2.4413156249499854, "learning_rate": 2.6938327963803545e-06, "loss": 0.3925, "step": 1665 }, { "epoch": 0.6895127993393889, "grad_norm": 2.441149846551036, "learning_rate": 2.6619119414106385e-06, "loss": 0.3902, "step": 1670 }, { "epoch": 0.6915772089182494, "grad_norm": 2.4274483132890685, "learning_rate": 2.6301125886786965e-06, "loss": 0.3933, "step": 1675 }, { "epoch": 0.6936416184971098, "grad_norm": 2.514301355730002, "learning_rate": 2.5984363906853814e-06, "loss": 0.3847, "step": 1680 }, { "epoch": 0.6957060280759703, "grad_norm": 2.23025942549343, "learning_rate": 2.566884993531628e-06, "loss": 0.3868, "step": 1685 }, { "epoch": 0.6977704376548307, "grad_norm": 2.3910909958649764, "learning_rate": 2.5354600368329108e-06, "loss": 0.379, "step": 1690 }, { "epoch": 0.6998348472336912, "grad_norm": 2.5413951969823554, "learning_rate": 2.5041631536340374e-06, "loss": 0.3909, "step": 1695 }, { "epoch": 0.7018992568125516, "grad_norm": 2.360003867942285, "learning_rate": 2.472995970324282e-06, "loss": 0.3888, "step": 1700 }, { "epoch": 0.7039636663914121, "grad_norm": 2.4621248065886516, "learning_rate": 2.4419601065528703e-06, "loss": 0.3803, "step": 1705 }, { "epoch": 0.7060280759702725, "grad_norm": 2.4221633580536865, "learning_rate": 2.4110571751448154e-06, "loss": 0.3804, "step": 1710 }, { "epoch": 0.708092485549133, "grad_norm": 2.3044799343595685, "learning_rate": 2.3802887820171012e-06, "loss": 0.3709, "step": 1715 }, { "epoch": 0.7101568951279934, "grad_norm": 2.3101555840305585, "learning_rate": 2.349656526095224e-06, "loss": 0.382, "step": 1720 }, { "epoch": 0.7122213047068539, "grad_norm": 2.439524803446898, "learning_rate": 2.3191619992301144e-06, "loss": 0.3969, "step": 1725 }, { "epoch": 0.7142857142857143, "grad_norm": 2.3771584605894174, "learning_rate": 2.2888067861153983e-06, "loss": 0.3792, "step": 1730 }, { "epoch": 0.7163501238645748, "grad_norm": 2.5666012183468587, "learning_rate": 2.2585924642050638e-06, "loss": 0.3802, "step": 1735 }, { "epoch": 0.7184145334434352, "grad_norm": 2.3145141740784307, "learning_rate": 2.2285206036314684e-06, "loss": 0.3835, "step": 1740 }, { "epoch": 0.7204789430222956, "grad_norm": 2.520653602273766, "learning_rate": 2.1985927671237605e-06, "loss": 0.3766, "step": 1745 }, { "epoch": 0.7225433526011561, "grad_norm": 2.470460484865602, "learning_rate": 2.168810509926659e-06, "loss": 0.3897, "step": 1750 }, { "epoch": 0.7246077621800165, "grad_norm": 2.408249418201136, "learning_rate": 2.1391753797196328e-06, "loss": 0.3952, "step": 1755 }, { "epoch": 0.726672171758877, "grad_norm": 2.4402970653227247, "learning_rate": 2.1096889165364894e-06, "loss": 0.375, "step": 1760 }, { "epoch": 0.7287365813377374, "grad_norm": 2.4710285995197188, "learning_rate": 2.0803526526853206e-06, "loss": 0.3748, "step": 1765 }, { "epoch": 0.7308009909165979, "grad_norm": 2.2841746369286633, "learning_rate": 2.0511681126688883e-06, "loss": 0.3799, "step": 1770 }, { "epoch": 0.7328654004954583, "grad_norm": 2.46102171882109, "learning_rate": 2.022136813105401e-06, "loss": 0.3729, "step": 1775 }, { "epoch": 0.7349298100743188, "grad_norm": 2.2829340264643156, "learning_rate": 1.9932602626496994e-06, "loss": 0.3722, "step": 1780 }, { "epoch": 0.7369942196531792, "grad_norm": 2.4439441720908, "learning_rate": 1.9645399619148515e-06, "loss": 0.3774, "step": 1785 }, { "epoch": 0.7390586292320397, "grad_norm": 2.363486720634483, "learning_rate": 1.9359774033941757e-06, "loss": 0.3775, "step": 1790 }, { "epoch": 0.7411230388109001, "grad_norm": 2.2893262909558065, "learning_rate": 1.9075740713836854e-06, "loss": 0.3851, "step": 1795 }, { "epoch": 0.7431874483897605, "grad_norm": 2.381579584567696, "learning_rate": 1.8793314419049446e-06, "loss": 0.3736, "step": 1800 }, { "epoch": 0.7452518579686209, "grad_norm": 2.3588892144435714, "learning_rate": 1.8512509826283752e-06, "loss": 0.3647, "step": 1805 }, { "epoch": 0.7473162675474814, "grad_norm": 2.468444814623241, "learning_rate": 1.8233341527969777e-06, "loss": 0.3731, "step": 1810 }, { "epoch": 0.7493806771263418, "grad_norm": 2.4216200751888217, "learning_rate": 1.7955824031505097e-06, "loss": 0.3741, "step": 1815 }, { "epoch": 0.7514450867052023, "grad_norm": 2.3925423466955884, "learning_rate": 1.7679971758500836e-06, "loss": 0.3652, "step": 1820 }, { "epoch": 0.7535094962840627, "grad_norm": 2.386508421187067, "learning_rate": 1.7405799044032356e-06, "loss": 0.3685, "step": 1825 }, { "epoch": 0.7555739058629232, "grad_norm": 2.439989821363405, "learning_rate": 1.7133320135894233e-06, "loss": 0.3653, "step": 1830 }, { "epoch": 0.7576383154417836, "grad_norm": 2.3039350711469186, "learning_rate": 1.686254919385985e-06, "loss": 0.3783, "step": 1835 }, { "epoch": 0.7597027250206441, "grad_norm": 2.216200410510777, "learning_rate": 1.6593500288945574e-06, "loss": 0.3646, "step": 1840 }, { "epoch": 0.7617671345995045, "grad_norm": 2.4989495842816405, "learning_rate": 1.6326187402679577e-06, "loss": 0.3741, "step": 1845 }, { "epoch": 0.763831544178365, "grad_norm": 2.4154932431289673, "learning_rate": 1.6060624426375222e-06, "loss": 0.3698, "step": 1850 }, { "epoch": 0.7658959537572254, "grad_norm": 2.473649546182009, "learning_rate": 1.5796825160409168e-06, "loss": 0.3672, "step": 1855 }, { "epoch": 0.7679603633360859, "grad_norm": 2.355038901564694, "learning_rate": 1.5534803313504215e-06, "loss": 0.3566, "step": 1860 }, { "epoch": 0.7700247729149463, "grad_norm": 2.3588413583977625, "learning_rate": 1.5274572502016982e-06, "loss": 0.3711, "step": 1865 }, { "epoch": 0.7720891824938068, "grad_norm": 2.401065534914905, "learning_rate": 1.5016146249230229e-06, "loss": 0.3736, "step": 1870 }, { "epoch": 0.7741535920726672, "grad_norm": 2.342195370497388, "learning_rate": 1.4759537984650114e-06, "loss": 0.3626, "step": 1875 }, { "epoch": 0.7762180016515277, "grad_norm": 2.3601036223952634, "learning_rate": 1.4504761043308346e-06, "loss": 0.3534, "step": 1880 }, { "epoch": 0.7782824112303881, "grad_norm": 2.3058771526048405, "learning_rate": 1.425182866506918e-06, "loss": 0.3633, "step": 1885 }, { "epoch": 0.7803468208092486, "grad_norm": 2.308152081660379, "learning_rate": 1.4000753993941369e-06, "loss": 0.3723, "step": 1890 }, { "epoch": 0.782411230388109, "grad_norm": 2.4604302877257664, "learning_rate": 1.3751550077395181e-06, "loss": 0.3523, "step": 1895 }, { "epoch": 0.7844756399669695, "grad_norm": 2.551739458655361, "learning_rate": 1.3504229865684326e-06, "loss": 0.3706, "step": 1900 }, { "epoch": 0.7865400495458299, "grad_norm": 2.4143979686568104, "learning_rate": 1.3258806211172954e-06, "loss": 0.3588, "step": 1905 }, { "epoch": 0.7886044591246903, "grad_norm": 2.364426767604513, "learning_rate": 1.3015291867667757e-06, "loss": 0.3508, "step": 1910 }, { "epoch": 0.7906688687035508, "grad_norm": 2.3233651206895147, "learning_rate": 1.2773699489755343e-06, "loss": 0.3571, "step": 1915 }, { "epoch": 0.7927332782824112, "grad_norm": 2.3196733264987675, "learning_rate": 1.2534041632144407e-06, "loss": 0.3405, "step": 1920 }, { "epoch": 0.7947976878612717, "grad_norm": 2.347467115991222, "learning_rate": 1.2296330749013408e-06, "loss": 0.3502, "step": 1925 }, { "epoch": 0.7968620974401321, "grad_norm": 2.3004029534314174, "learning_rate": 1.2060579193363431e-06, "loss": 0.3564, "step": 1930 }, { "epoch": 0.7989265070189926, "grad_norm": 2.5374233309055083, "learning_rate": 1.1826799216376128e-06, "loss": 0.3778, "step": 1935 }, { "epoch": 0.800990916597853, "grad_norm": 2.4431483706896007, "learning_rate": 1.1595002966777152e-06, "loss": 0.356, "step": 1940 }, { "epoch": 0.8030553261767135, "grad_norm": 2.35423434746266, "learning_rate": 1.136520249020478e-06, "loss": 0.3491, "step": 1945 }, { "epoch": 0.8051197357555739, "grad_norm": 2.3767268596250672, "learning_rate": 1.113740972858398e-06, "loss": 0.348, "step": 1950 }, { "epoch": 0.8071841453344344, "grad_norm": 2.4631890684934965, "learning_rate": 1.0911636519505791e-06, "loss": 0.3671, "step": 1955 }, { "epoch": 0.8092485549132948, "grad_norm": 2.201423417502954, "learning_rate": 1.068789459561223e-06, "loss": 0.3459, "step": 1960 }, { "epoch": 0.8113129644921553, "grad_norm": 2.36077824921438, "learning_rate": 1.0466195583986487e-06, "loss": 0.3639, "step": 1965 }, { "epoch": 0.8133773740710157, "grad_norm": 2.354003763601073, "learning_rate": 1.024655100554882e-06, "loss": 0.3643, "step": 1970 }, { "epoch": 0.8154417836498762, "grad_norm": 2.4101822702917004, "learning_rate": 1.002897227445777e-06, "loss": 0.3617, "step": 1975 }, { "epoch": 0.8175061932287366, "grad_norm": 2.4833584786174163, "learning_rate": 9.813470697517009e-07, "loss": 0.3501, "step": 1980 }, { "epoch": 0.8195706028075971, "grad_norm": 2.3627937000377393, "learning_rate": 9.600057473587876e-07, "loss": 0.3671, "step": 1985 }, { "epoch": 0.8216350123864574, "grad_norm": 2.415763997145714, "learning_rate": 9.388743693007247e-07, "loss": 0.3541, "step": 1990 }, { "epoch": 0.8236994219653179, "grad_norm": 2.4671751077694406, "learning_rate": 9.1795403370113e-07, "loss": 0.3535, "step": 1995 }, { "epoch": 0.8257638315441783, "grad_norm": 2.3775799288896287, "learning_rate": 8.972458277164886e-07, "loss": 0.3516, "step": 2000 }, { "epoch": 0.8278282411230388, "grad_norm": 2.2938470233646604, "learning_rate": 8.76750827479651e-07, "loss": 0.3398, "step": 2005 }, { "epoch": 0.8298926507018992, "grad_norm": 2.2933382256010497, "learning_rate": 8.564700980439122e-07, "loss": 0.3608, "step": 2010 }, { "epoch": 0.8319570602807597, "grad_norm": 2.3382609787220785, "learning_rate": 8.364046933276642e-07, "loss": 0.3666, "step": 2015 }, { "epoch": 0.8340214698596201, "grad_norm": 2.44152050789561, "learning_rate": 8.165556560596316e-07, "loss": 0.3485, "step": 2020 }, { "epoch": 0.8360858794384806, "grad_norm": 2.3858823443159034, "learning_rate": 7.969240177246779e-07, "loss": 0.3547, "step": 2025 }, { "epoch": 0.838150289017341, "grad_norm": 2.453983334202833, "learning_rate": 7.775107985102087e-07, "loss": 0.3651, "step": 2030 }, { "epoch": 0.8402146985962015, "grad_norm": 2.5029504947887875, "learning_rate": 7.583170072531543e-07, "loss": 0.3509, "step": 2035 }, { "epoch": 0.8422791081750619, "grad_norm": 2.482629951124892, "learning_rate": 7.393436413875411e-07, "loss": 0.3476, "step": 2040 }, { "epoch": 0.8443435177539224, "grad_norm": 2.283216056409345, "learning_rate": 7.205916868926604e-07, "loss": 0.3526, "step": 2045 }, { "epoch": 0.8464079273327828, "grad_norm": 2.3590631737094845, "learning_rate": 7.020621182418347e-07, "loss": 0.34, "step": 2050 }, { "epoch": 0.8484723369116433, "grad_norm": 2.6201296500024553, "learning_rate": 6.837558983517723e-07, "loss": 0.3535, "step": 2055 }, { "epoch": 0.8505367464905037, "grad_norm": 2.4227114258373095, "learning_rate": 6.656739785325278e-07, "loss": 0.3507, "step": 2060 }, { "epoch": 0.8526011560693642, "grad_norm": 2.390954329942295, "learning_rate": 6.478172984380687e-07, "loss": 0.3532, "step": 2065 }, { "epoch": 0.8546655656482246, "grad_norm": 2.556923085001048, "learning_rate": 6.301867860174443e-07, "loss": 0.3434, "step": 2070 }, { "epoch": 0.856729975227085, "grad_norm": 2.3169295549044335, "learning_rate": 6.127833574665642e-07, "loss": 0.3459, "step": 2075 }, { "epoch": 0.8587943848059455, "grad_norm": 2.3497867089138644, "learning_rate": 5.956079171805818e-07, "loss": 0.3415, "step": 2080 }, { "epoch": 0.860858794384806, "grad_norm": 2.1735667040509137, "learning_rate": 5.78661357706905e-07, "loss": 0.3379, "step": 2085 }, { "epoch": 0.8629232039636664, "grad_norm": 2.5779842919745355, "learning_rate": 5.619445596988026e-07, "loss": 0.3425, "step": 2090 }, { "epoch": 0.8649876135425268, "grad_norm": 2.401992025146321, "learning_rate": 5.45458391869651e-07, "loss": 0.3489, "step": 2095 }, { "epoch": 0.8670520231213873, "grad_norm": 2.3105243423015587, "learning_rate": 5.29203710947781e-07, "loss": 0.335, "step": 2100 }, { "epoch": 0.8691164327002477, "grad_norm": 2.5027833045418943, "learning_rate": 5.131813616319641e-07, "loss": 0.35, "step": 2105 }, { "epoch": 0.8711808422791082, "grad_norm": 2.492666189587263, "learning_rate": 4.973921765475093e-07, "loss": 0.3419, "step": 2110 }, { "epoch": 0.8732452518579686, "grad_norm": 2.476326337718451, "learning_rate": 4.818369762030001e-07, "loss": 0.3437, "step": 2115 }, { "epoch": 0.8753096614368291, "grad_norm": 2.4895934072647923, "learning_rate": 4.6651656894765307e-07, "loss": 0.345, "step": 2120 }, { "epoch": 0.8773740710156895, "grad_norm": 2.432351776306209, "learning_rate": 4.5143175092931257e-07, "loss": 0.3457, "step": 2125 }, { "epoch": 0.87943848059455, "grad_norm": 2.5446256871769894, "learning_rate": 4.3658330605307496e-07, "loss": 0.3501, "step": 2130 }, { "epoch": 0.8815028901734104, "grad_norm": 2.400237240106358, "learning_rate": 4.2197200594055135e-07, "loss": 0.3417, "step": 2135 }, { "epoch": 0.8835672997522709, "grad_norm": 2.420863056343347, "learning_rate": 4.0759860988977715e-07, "loss": 0.3336, "step": 2140 }, { "epoch": 0.8856317093311313, "grad_norm": 2.5855831775345317, "learning_rate": 3.934638648357414e-07, "loss": 0.3473, "step": 2145 }, { "epoch": 0.8876961189099918, "grad_norm": 2.4571094742538633, "learning_rate": 3.795685053115805e-07, "loss": 0.3425, "step": 2150 }, { "epoch": 0.8897605284888522, "grad_norm": 2.318969751255526, "learning_rate": 3.659132534104054e-07, "loss": 0.339, "step": 2155 }, { "epoch": 0.8918249380677127, "grad_norm": 2.464561015161846, "learning_rate": 3.5249881874777235e-07, "loss": 0.3415, "step": 2160 }, { "epoch": 0.8938893476465731, "grad_norm": 2.5739185699856653, "learning_rate": 3.3932589842481434e-07, "loss": 0.3549, "step": 2165 }, { "epoch": 0.8959537572254336, "grad_norm": 2.3924860465104554, "learning_rate": 3.2639517699200765e-07, "loss": 0.3501, "step": 2170 }, { "epoch": 0.898018166804294, "grad_norm": 2.4561049545962983, "learning_rate": 3.1370732641360424e-07, "loss": 0.3382, "step": 2175 }, { "epoch": 0.9000825763831544, "grad_norm": 2.4531360622529985, "learning_rate": 3.0126300603270763e-07, "loss": 0.3386, "step": 2180 }, { "epoch": 0.9021469859620148, "grad_norm": 2.3726671856862698, "learning_rate": 2.89062862537009e-07, "loss": 0.3382, "step": 2185 }, { "epoch": 0.9042113955408753, "grad_norm": 2.4508531032953704, "learning_rate": 2.7710752992518906e-07, "loss": 0.3475, "step": 2190 }, { "epoch": 0.9062758051197357, "grad_norm": 2.344783384097571, "learning_rate": 2.653976294739596e-07, "loss": 0.3435, "step": 2195 }, { "epoch": 0.9083402146985962, "grad_norm": 2.3832943118751495, "learning_rate": 2.539337697057853e-07, "loss": 0.3386, "step": 2200 }, { "epoch": 0.9104046242774566, "grad_norm": 2.2656029102160917, "learning_rate": 2.4271654635726003e-07, "loss": 0.326, "step": 2205 }, { "epoch": 0.9124690338563171, "grad_norm": 2.531439971555802, "learning_rate": 2.3174654234814854e-07, "loss": 0.3336, "step": 2210 }, { "epoch": 0.9145334434351775, "grad_norm": 2.526736652879497, "learning_rate": 2.2102432775109151e-07, "loss": 0.3431, "step": 2215 }, { "epoch": 0.916597853014038, "grad_norm": 2.499234877151574, "learning_rate": 2.1055045976198286e-07, "loss": 0.3274, "step": 2220 }, { "epoch": 0.9186622625928984, "grad_norm": 2.347953826058282, "learning_rate": 2.0032548267101547e-07, "loss": 0.3415, "step": 2225 }, { "epoch": 0.9207266721717589, "grad_norm": 2.2781319110873923, "learning_rate": 1.9034992783439398e-07, "loss": 0.3313, "step": 2230 }, { "epoch": 0.9227910817506193, "grad_norm": 2.4985215803253777, "learning_rate": 1.8062431364672394e-07, "loss": 0.347, "step": 2235 }, { "epoch": 0.9248554913294798, "grad_norm": 2.3523550820055807, "learning_rate": 1.7114914551407125e-07, "loss": 0.3405, "step": 2240 }, { "epoch": 0.9269199009083402, "grad_norm": 2.35149253079431, "learning_rate": 1.619249158276981e-07, "loss": 0.3396, "step": 2245 }, { "epoch": 0.9289843104872006, "grad_norm": 2.5391046088636724, "learning_rate": 1.529521039384757e-07, "loss": 0.3382, "step": 2250 }, { "epoch": 0.9310487200660611, "grad_norm": 2.436611958199382, "learning_rate": 1.4423117613197535e-07, "loss": 0.3507, "step": 2255 }, { "epoch": 0.9331131296449215, "grad_norm": 2.368900058293189, "learning_rate": 1.357625856042344e-07, "loss": 0.3436, "step": 2260 }, { "epoch": 0.935177539223782, "grad_norm": 2.483713260867796, "learning_rate": 1.275467724382068e-07, "loss": 0.3493, "step": 2265 }, { "epoch": 0.9372419488026424, "grad_norm": 2.3378032035177556, "learning_rate": 1.1958416358089308e-07, "loss": 0.3366, "step": 2270 }, { "epoch": 0.9393063583815029, "grad_norm": 2.427445932737846, "learning_rate": 1.1187517282115478e-07, "loss": 0.3357, "step": 2275 }, { "epoch": 0.9413707679603633, "grad_norm": 2.45979849427157, "learning_rate": 1.0442020076820992e-07, "loss": 0.3432, "step": 2280 }, { "epoch": 0.9434351775392238, "grad_norm": 2.4102972284673485, "learning_rate": 9.72196348308152e-08, "loss": 0.3421, "step": 2285 }, { "epoch": 0.9454995871180842, "grad_norm": 2.4147926771751127, "learning_rate": 9.027384919713167e-08, "loss": 0.3449, "step": 2290 }, { "epoch": 0.9475639966969447, "grad_norm": 2.492908131419299, "learning_rate": 8.358320481528404e-08, "loss": 0.3349, "step": 2295 }, { "epoch": 0.9496284062758051, "grad_norm": 2.445696827488344, "learning_rate": 7.714804937459964e-08, "loss": 0.3327, "step": 2300 }, { "epoch": 0.9516928158546656, "grad_norm": 2.4859237981346434, "learning_rate": 7.096871728754173e-08, "loss": 0.3369, "step": 2305 }, { "epoch": 0.953757225433526, "grad_norm": 2.4755163895408483, "learning_rate": 6.504552967233014e-08, "loss": 0.3341, "step": 2310 }, { "epoch": 0.9558216350123865, "grad_norm": 2.4903780675355947, "learning_rate": 5.9378794336255595e-08, "loss": 0.3259, "step": 2315 }, { "epoch": 0.9578860445912469, "grad_norm": 2.3608897794698986, "learning_rate": 5.396880575968266e-08, "loss": 0.34, "step": 2320 }, { "epoch": 0.9599504541701074, "grad_norm": 2.349066629113589, "learning_rate": 4.881584508074688e-08, "loss": 0.3285, "step": 2325 }, { "epoch": 0.9620148637489678, "grad_norm": 2.475193042134275, "learning_rate": 4.39201800807465e-08, "loss": 0.339, "step": 2330 }, { "epoch": 0.9640792733278283, "grad_norm": 2.3656425365762384, "learning_rate": 3.92820651702247e-08, "loss": 0.3358, "step": 2335 }, { "epoch": 0.9661436829066887, "grad_norm": 2.441572724635035, "learning_rate": 3.4901741375747975e-08, "loss": 0.3471, "step": 2340 }, { "epoch": 0.9682080924855492, "grad_norm": 2.3678044079850413, "learning_rate": 3.077943632738556e-08, "loss": 0.3499, "step": 2345 }, { "epoch": 0.9702725020644096, "grad_norm": 2.630292946781589, "learning_rate": 2.6915364246875574e-08, "loss": 0.341, "step": 2350 }, { "epoch": 0.9723369116432701, "grad_norm": 2.388897601971072, "learning_rate": 2.3309725936493387e-08, "loss": 0.3437, "step": 2355 }, { "epoch": 0.9744013212221305, "grad_norm": 2.4665523794028785, "learning_rate": 1.9962708768620498e-08, "loss": 0.3453, "step": 2360 }, { "epoch": 0.976465730800991, "grad_norm": 2.448813469296534, "learning_rate": 1.687448667600067e-08, "loss": 0.3408, "step": 2365 }, { "epoch": 0.9785301403798513, "grad_norm": 2.3718128749961576, "learning_rate": 1.4045220142708816e-08, "loss": 0.35, "step": 2370 }, { "epoch": 0.9805945499587118, "grad_norm": 2.4841177332012325, "learning_rate": 1.1475056195804358e-08, "loss": 0.3518, "step": 2375 }, { "epoch": 0.9826589595375722, "grad_norm": 2.493744247830282, "learning_rate": 9.164128397695648e-09, "loss": 0.332, "step": 2380 }, { "epoch": 0.9847233691164327, "grad_norm": 2.2819847476640107, "learning_rate": 7.112556839196649e-09, "loss": 0.3301, "step": 2385 }, { "epoch": 0.9867877786952931, "grad_norm": 2.4247247484873173, "learning_rate": 5.3204481332880255e-09, "loss": 0.3433, "step": 2390 }, { "epoch": 0.9888521882741536, "grad_norm": 2.4373319136694285, "learning_rate": 3.787895409574915e-09, "loss": 0.3388, "step": 2395 }, { "epoch": 0.990916597853014, "grad_norm": 2.3711173222964037, "learning_rate": 2.514978309447469e-09, "loss": 0.3433, "step": 2400 }, { "epoch": 0.9929810074318745, "grad_norm": 2.3426000581900652, "learning_rate": 1.501762981945265e-09, "loss": 0.3368, "step": 2405 }, { "epoch": 0.9950454170107349, "grad_norm": 2.35822764782016, "learning_rate": 7.483020803145114e-10, "loss": 0.3396, "step": 2410 }, { "epoch": 0.9971098265895953, "grad_norm": 2.4725534953927233, "learning_rate": 2.5463475927745007e-10, "loss": 0.3401, "step": 2415 }, { "epoch": 0.9991742361684558, "grad_norm": 2.407515233988016, "learning_rate": 2.078667299343362e-11, "loss": 0.3298, "step": 2420 }, { "epoch": 1.0, "eval_runtime": 3.3946, "eval_samples_per_second": 2.946, "eval_steps_per_second": 0.884, "step": 2422 }, { "epoch": 1.0, "step": 2422, "total_flos": 253558763028480.0, "train_loss": 0.530042272082643, "train_runtime": 21564.4437, "train_samples_per_second": 1.797, "train_steps_per_second": 0.112 } ], "logging_steps": 5, "max_steps": 2422, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 253558763028480.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }