terry69's picture
Model save
50ff9c0 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 2422,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00041288191577208916,
"grad_norm": 24.94086805103786,
"learning_rate": 4.1152263374485605e-08,
"loss": 1.4345,
"step": 1
},
{
"epoch": 0.002064409578860446,
"grad_norm": 24.898038168161193,
"learning_rate": 2.05761316872428e-07,
"loss": 1.4101,
"step": 5
},
{
"epoch": 0.004128819157720892,
"grad_norm": 15.938300194174698,
"learning_rate": 4.11522633744856e-07,
"loss": 1.3943,
"step": 10
},
{
"epoch": 0.006193228736581338,
"grad_norm": 8.846886408597587,
"learning_rate": 6.17283950617284e-07,
"loss": 1.2771,
"step": 15
},
{
"epoch": 0.008257638315441783,
"grad_norm": 9.948610282503696,
"learning_rate": 8.23045267489712e-07,
"loss": 1.1572,
"step": 20
},
{
"epoch": 0.01032204789430223,
"grad_norm": 4.744946112205854,
"learning_rate": 1.02880658436214e-06,
"loss": 1.0575,
"step": 25
},
{
"epoch": 0.012386457473162676,
"grad_norm": 3.3793530185545926,
"learning_rate": 1.234567901234568e-06,
"loss": 0.9708,
"step": 30
},
{
"epoch": 0.014450867052023121,
"grad_norm": 3.4379702794812124,
"learning_rate": 1.440329218106996e-06,
"loss": 0.9668,
"step": 35
},
{
"epoch": 0.016515276630883566,
"grad_norm": 3.110466178670075,
"learning_rate": 1.646090534979424e-06,
"loss": 0.9508,
"step": 40
},
{
"epoch": 0.018579686209744012,
"grad_norm": 3.0934122499344614,
"learning_rate": 1.8518518518518519e-06,
"loss": 0.9496,
"step": 45
},
{
"epoch": 0.02064409578860446,
"grad_norm": 3.2046299727793572,
"learning_rate": 2.05761316872428e-06,
"loss": 0.9256,
"step": 50
},
{
"epoch": 0.022708505367464906,
"grad_norm": 2.9857050496874358,
"learning_rate": 2.263374485596708e-06,
"loss": 0.9146,
"step": 55
},
{
"epoch": 0.02477291494632535,
"grad_norm": 3.010697976532333,
"learning_rate": 2.469135802469136e-06,
"loss": 0.9185,
"step": 60
},
{
"epoch": 0.026837324525185797,
"grad_norm": 2.9986835294430274,
"learning_rate": 2.674897119341564e-06,
"loss": 0.9008,
"step": 65
},
{
"epoch": 0.028901734104046242,
"grad_norm": 3.2527007862069186,
"learning_rate": 2.880658436213992e-06,
"loss": 0.8855,
"step": 70
},
{
"epoch": 0.030966143682906688,
"grad_norm": 3.2007644630686833,
"learning_rate": 3.08641975308642e-06,
"loss": 0.8783,
"step": 75
},
{
"epoch": 0.03303055326176713,
"grad_norm": 3.196196143650837,
"learning_rate": 3.292181069958848e-06,
"loss": 0.8775,
"step": 80
},
{
"epoch": 0.03509496284062758,
"grad_norm": 3.004102067065696,
"learning_rate": 3.4979423868312762e-06,
"loss": 0.8808,
"step": 85
},
{
"epoch": 0.037159372419488024,
"grad_norm": 3.118135767316795,
"learning_rate": 3.7037037037037037e-06,
"loss": 0.8775,
"step": 90
},
{
"epoch": 0.03922378199834847,
"grad_norm": 3.0856528450327354,
"learning_rate": 3.909465020576132e-06,
"loss": 0.8634,
"step": 95
},
{
"epoch": 0.04128819157720892,
"grad_norm": 2.9700335979493167,
"learning_rate": 4.11522633744856e-06,
"loss": 0.872,
"step": 100
},
{
"epoch": 0.04335260115606936,
"grad_norm": 3.1350056589977533,
"learning_rate": 4.3209876543209875e-06,
"loss": 0.8706,
"step": 105
},
{
"epoch": 0.04541701073492981,
"grad_norm": 3.2497160658033852,
"learning_rate": 4.526748971193416e-06,
"loss": 0.8539,
"step": 110
},
{
"epoch": 0.047481420313790254,
"grad_norm": 3.382546616433861,
"learning_rate": 4.732510288065844e-06,
"loss": 0.8573,
"step": 115
},
{
"epoch": 0.0495458298926507,
"grad_norm": 3.1741955682207124,
"learning_rate": 4.938271604938272e-06,
"loss": 0.8433,
"step": 120
},
{
"epoch": 0.051610239471511145,
"grad_norm": 3.13507489217465,
"learning_rate": 5.1440329218107e-06,
"loss": 0.8546,
"step": 125
},
{
"epoch": 0.053674649050371594,
"grad_norm": 3.182938519722128,
"learning_rate": 5.349794238683128e-06,
"loss": 0.8505,
"step": 130
},
{
"epoch": 0.05573905862923204,
"grad_norm": 3.3061978280140916,
"learning_rate": 5.555555555555557e-06,
"loss": 0.8596,
"step": 135
},
{
"epoch": 0.057803468208092484,
"grad_norm": 3.172304485577032,
"learning_rate": 5.761316872427984e-06,
"loss": 0.8454,
"step": 140
},
{
"epoch": 0.05986787778695293,
"grad_norm": 3.222279073856588,
"learning_rate": 5.967078189300412e-06,
"loss": 0.8552,
"step": 145
},
{
"epoch": 0.061932287365813375,
"grad_norm": 3.026540545603275,
"learning_rate": 6.17283950617284e-06,
"loss": 0.8365,
"step": 150
},
{
"epoch": 0.06399669694467382,
"grad_norm": 3.1724013332685304,
"learning_rate": 6.3786008230452675e-06,
"loss": 0.8397,
"step": 155
},
{
"epoch": 0.06606110652353427,
"grad_norm": 3.4516704362207222,
"learning_rate": 6.584362139917696e-06,
"loss": 0.8346,
"step": 160
},
{
"epoch": 0.06812551610239471,
"grad_norm": 3.1199583778413285,
"learning_rate": 6.790123456790124e-06,
"loss": 0.8296,
"step": 165
},
{
"epoch": 0.07018992568125516,
"grad_norm": 3.448087429701013,
"learning_rate": 6.9958847736625525e-06,
"loss": 0.8384,
"step": 170
},
{
"epoch": 0.07225433526011561,
"grad_norm": 3.072968814322973,
"learning_rate": 7.201646090534981e-06,
"loss": 0.846,
"step": 175
},
{
"epoch": 0.07431874483897605,
"grad_norm": 3.1647169459938076,
"learning_rate": 7.4074074074074075e-06,
"loss": 0.8256,
"step": 180
},
{
"epoch": 0.0763831544178365,
"grad_norm": 3.105263189462399,
"learning_rate": 7.613168724279836e-06,
"loss": 0.8011,
"step": 185
},
{
"epoch": 0.07844756399669695,
"grad_norm": 2.9946543968073294,
"learning_rate": 7.818930041152263e-06,
"loss": 0.8214,
"step": 190
},
{
"epoch": 0.0805119735755574,
"grad_norm": 3.25501723509638,
"learning_rate": 8.024691358024692e-06,
"loss": 0.8201,
"step": 195
},
{
"epoch": 0.08257638315441784,
"grad_norm": 3.0087568189695237,
"learning_rate": 8.23045267489712e-06,
"loss": 0.8139,
"step": 200
},
{
"epoch": 0.08464079273327828,
"grad_norm": 3.0911976211351817,
"learning_rate": 8.43621399176955e-06,
"loss": 0.8242,
"step": 205
},
{
"epoch": 0.08670520231213873,
"grad_norm": 3.310286578361364,
"learning_rate": 8.641975308641975e-06,
"loss": 0.8364,
"step": 210
},
{
"epoch": 0.08876961189099918,
"grad_norm": 2.986473336622259,
"learning_rate": 8.847736625514404e-06,
"loss": 0.8316,
"step": 215
},
{
"epoch": 0.09083402146985962,
"grad_norm": 3.102015428130825,
"learning_rate": 9.053497942386832e-06,
"loss": 0.8128,
"step": 220
},
{
"epoch": 0.09289843104872006,
"grad_norm": 2.906374211555821,
"learning_rate": 9.25925925925926e-06,
"loss": 0.7945,
"step": 225
},
{
"epoch": 0.09496284062758051,
"grad_norm": 3.0506200713500826,
"learning_rate": 9.465020576131688e-06,
"loss": 0.8158,
"step": 230
},
{
"epoch": 0.09702725020644096,
"grad_norm": 3.0384361806454017,
"learning_rate": 9.670781893004116e-06,
"loss": 0.8228,
"step": 235
},
{
"epoch": 0.0990916597853014,
"grad_norm": 3.100843008960617,
"learning_rate": 9.876543209876543e-06,
"loss": 0.8099,
"step": 240
},
{
"epoch": 0.10115606936416185,
"grad_norm": 2.935521796714598,
"learning_rate": 9.999979213327006e-06,
"loss": 0.8108,
"step": 245
},
{
"epoch": 0.10322047894302229,
"grad_norm": 3.361172390200279,
"learning_rate": 9.999745365240722e-06,
"loss": 0.7923,
"step": 250
},
{
"epoch": 0.10528488852188274,
"grad_norm": 2.915556545903911,
"learning_rate": 9.999251697919687e-06,
"loss": 0.8045,
"step": 255
},
{
"epoch": 0.10734929810074319,
"grad_norm": 3.212286307862521,
"learning_rate": 9.998498237018056e-06,
"loss": 0.8112,
"step": 260
},
{
"epoch": 0.10941370767960364,
"grad_norm": 3.191428090938715,
"learning_rate": 9.997485021690553e-06,
"loss": 0.8001,
"step": 265
},
{
"epoch": 0.11147811725846409,
"grad_norm": 2.838764527991799,
"learning_rate": 9.996212104590426e-06,
"loss": 0.8074,
"step": 270
},
{
"epoch": 0.11354252683732452,
"grad_norm": 2.9644134550569943,
"learning_rate": 9.994679551866712e-06,
"loss": 0.8033,
"step": 275
},
{
"epoch": 0.11560693641618497,
"grad_norm": 3.118728479888768,
"learning_rate": 9.992887443160804e-06,
"loss": 0.7905,
"step": 280
},
{
"epoch": 0.11767134599504542,
"grad_norm": 2.8425828854985737,
"learning_rate": 9.990835871602305e-06,
"loss": 0.7987,
"step": 285
},
{
"epoch": 0.11973575557390587,
"grad_norm": 3.091935940986862,
"learning_rate": 9.988524943804196e-06,
"loss": 0.7919,
"step": 290
},
{
"epoch": 0.12180016515276632,
"grad_norm": 2.9471270363090976,
"learning_rate": 9.985954779857292e-06,
"loss": 0.7885,
"step": 295
},
{
"epoch": 0.12386457473162675,
"grad_norm": 2.761224162387467,
"learning_rate": 9.983125513324e-06,
"loss": 0.784,
"step": 300
},
{
"epoch": 0.1259289843104872,
"grad_norm": 2.779726805954528,
"learning_rate": 9.98003729123138e-06,
"loss": 0.7762,
"step": 305
},
{
"epoch": 0.12799339388934763,
"grad_norm": 2.8806127663541625,
"learning_rate": 9.976690274063507e-06,
"loss": 0.7995,
"step": 310
},
{
"epoch": 0.13005780346820808,
"grad_norm": 3.0861309652893163,
"learning_rate": 9.973084635753125e-06,
"loss": 0.8002,
"step": 315
},
{
"epoch": 0.13212221304706853,
"grad_norm": 2.886159344081087,
"learning_rate": 9.969220563672614e-06,
"loss": 0.7742,
"step": 320
},
{
"epoch": 0.13418662262592898,
"grad_norm": 2.9262415184999835,
"learning_rate": 9.965098258624253e-06,
"loss": 0.7807,
"step": 325
},
{
"epoch": 0.13625103220478943,
"grad_norm": 2.9922747460068324,
"learning_rate": 9.960717934829777e-06,
"loss": 0.7582,
"step": 330
},
{
"epoch": 0.13831544178364988,
"grad_norm": 2.9066701618612596,
"learning_rate": 9.956079819919254e-06,
"loss": 0.7849,
"step": 335
},
{
"epoch": 0.14037985136251033,
"grad_norm": 2.9577891790066766,
"learning_rate": 9.951184154919253e-06,
"loss": 0.7676,
"step": 340
},
{
"epoch": 0.14244426094137078,
"grad_norm": 2.7242674149455923,
"learning_rate": 9.946031194240317e-06,
"loss": 0.7613,
"step": 345
},
{
"epoch": 0.14450867052023122,
"grad_norm": 2.738187421683147,
"learning_rate": 9.940621205663745e-06,
"loss": 0.7476,
"step": 350
},
{
"epoch": 0.14657308009909165,
"grad_norm": 2.959447667783807,
"learning_rate": 9.934954470327671e-06,
"loss": 0.757,
"step": 355
},
{
"epoch": 0.1486374896779521,
"grad_norm": 2.9344152274329693,
"learning_rate": 9.92903128271246e-06,
"loss": 0.7559,
"step": 360
},
{
"epoch": 0.15070189925681254,
"grad_norm": 3.0358566235662554,
"learning_rate": 9.9228519506254e-06,
"loss": 0.7671,
"step": 365
},
{
"epoch": 0.152766308835673,
"grad_norm": 2.917481842444454,
"learning_rate": 9.916416795184716e-06,
"loss": 0.7423,
"step": 370
},
{
"epoch": 0.15483071841453344,
"grad_norm": 2.858632389269523,
"learning_rate": 9.90972615080287e-06,
"loss": 0.7625,
"step": 375
},
{
"epoch": 0.1568951279933939,
"grad_norm": 2.7940389389869997,
"learning_rate": 9.902780365169186e-06,
"loss": 0.7344,
"step": 380
},
{
"epoch": 0.15895953757225434,
"grad_norm": 2.966296190863466,
"learning_rate": 9.895579799231791e-06,
"loss": 0.7431,
"step": 385
},
{
"epoch": 0.1610239471511148,
"grad_norm": 2.734933032890328,
"learning_rate": 9.888124827178847e-06,
"loss": 0.7543,
"step": 390
},
{
"epoch": 0.16308835672997524,
"grad_norm": 3.0106930594709183,
"learning_rate": 9.880415836419108e-06,
"loss": 0.7322,
"step": 395
},
{
"epoch": 0.16515276630883569,
"grad_norm": 3.1160711879023277,
"learning_rate": 9.872453227561795e-06,
"loss": 0.7217,
"step": 400
},
{
"epoch": 0.1672171758876961,
"grad_norm": 3.025574483100628,
"learning_rate": 9.864237414395768e-06,
"loss": 0.7234,
"step": 405
},
{
"epoch": 0.16928158546655656,
"grad_norm": 3.0800133917199024,
"learning_rate": 9.855768823868026e-06,
"loss": 0.7451,
"step": 410
},
{
"epoch": 0.171345995045417,
"grad_norm": 2.878566581474389,
"learning_rate": 9.847047896061526e-06,
"loss": 0.7479,
"step": 415
},
{
"epoch": 0.17341040462427745,
"grad_norm": 3.168491396422218,
"learning_rate": 9.838075084172303e-06,
"loss": 0.7258,
"step": 420
},
{
"epoch": 0.1754748142031379,
"grad_norm": 2.815061772988165,
"learning_rate": 9.828850854485929e-06,
"loss": 0.7055,
"step": 425
},
{
"epoch": 0.17753922378199835,
"grad_norm": 2.6873037207867356,
"learning_rate": 9.819375686353277e-06,
"loss": 0.7253,
"step": 430
},
{
"epoch": 0.1796036333608588,
"grad_norm": 2.8648079803750077,
"learning_rate": 9.809650072165606e-06,
"loss": 0.7218,
"step": 435
},
{
"epoch": 0.18166804293971925,
"grad_norm": 2.6485478231791446,
"learning_rate": 9.799674517328985e-06,
"loss": 0.7211,
"step": 440
},
{
"epoch": 0.1837324525185797,
"grad_norm": 2.9086971891718365,
"learning_rate": 9.78944954023802e-06,
"loss": 0.7113,
"step": 445
},
{
"epoch": 0.18579686209744012,
"grad_norm": 2.777331603392391,
"learning_rate": 9.778975672248909e-06,
"loss": 0.7236,
"step": 450
},
{
"epoch": 0.18786127167630057,
"grad_norm": 2.896862579927988,
"learning_rate": 9.768253457651853e-06,
"loss": 0.7239,
"step": 455
},
{
"epoch": 0.18992568125516102,
"grad_norm": 2.788383349290103,
"learning_rate": 9.757283453642741e-06,
"loss": 0.7083,
"step": 460
},
{
"epoch": 0.19199009083402147,
"grad_norm": 2.7799217861819816,
"learning_rate": 9.746066230294216e-06,
"loss": 0.7008,
"step": 465
},
{
"epoch": 0.19405450041288191,
"grad_norm": 3.0828748770941603,
"learning_rate": 9.734602370526042e-06,
"loss": 0.6949,
"step": 470
},
{
"epoch": 0.19611890999174236,
"grad_norm": 2.8583233928158145,
"learning_rate": 9.722892470074811e-06,
"loss": 0.6923,
"step": 475
},
{
"epoch": 0.1981833195706028,
"grad_norm": 2.850888921108626,
"learning_rate": 9.710937137462992e-06,
"loss": 0.7056,
"step": 480
},
{
"epoch": 0.20024772914946326,
"grad_norm": 2.8804381008997697,
"learning_rate": 9.698736993967295e-06,
"loss": 0.7134,
"step": 485
},
{
"epoch": 0.2023121387283237,
"grad_norm": 2.5675849510127473,
"learning_rate": 9.686292673586397e-06,
"loss": 0.6854,
"step": 490
},
{
"epoch": 0.20437654830718416,
"grad_norm": 2.752726728070582,
"learning_rate": 9.673604823007994e-06,
"loss": 0.6944,
"step": 495
},
{
"epoch": 0.20644095788604458,
"grad_norm": 2.8340105200706507,
"learning_rate": 9.660674101575188e-06,
"loss": 0.6959,
"step": 500
},
{
"epoch": 0.20850536746490503,
"grad_norm": 2.6749836979567325,
"learning_rate": 9.647501181252228e-06,
"loss": 0.7181,
"step": 505
},
{
"epoch": 0.21056977704376548,
"grad_norm": 2.871887495235821,
"learning_rate": 9.634086746589597e-06,
"loss": 0.6855,
"step": 510
},
{
"epoch": 0.21263418662262593,
"grad_norm": 3.022510499989922,
"learning_rate": 9.62043149468842e-06,
"loss": 0.6786,
"step": 515
},
{
"epoch": 0.21469859620148637,
"grad_norm": 2.798631786402729,
"learning_rate": 9.60653613516426e-06,
"loss": 0.6878,
"step": 520
},
{
"epoch": 0.21676300578034682,
"grad_norm": 2.7665584034808557,
"learning_rate": 9.592401390110224e-06,
"loss": 0.6689,
"step": 525
},
{
"epoch": 0.21882741535920727,
"grad_norm": 2.6972597849127995,
"learning_rate": 9.578027994059448e-06,
"loss": 0.6938,
"step": 530
},
{
"epoch": 0.22089182493806772,
"grad_norm": 2.876624388604322,
"learning_rate": 9.563416693946927e-06,
"loss": 0.6831,
"step": 535
},
{
"epoch": 0.22295623451692817,
"grad_norm": 2.760614779991045,
"learning_rate": 9.548568249070688e-06,
"loss": 0.6721,
"step": 540
},
{
"epoch": 0.2250206440957886,
"grad_norm": 2.711332679097759,
"learning_rate": 9.533483431052347e-06,
"loss": 0.6734,
"step": 545
},
{
"epoch": 0.22708505367464904,
"grad_norm": 2.9419038213760573,
"learning_rate": 9.518163023797001e-06,
"loss": 0.6826,
"step": 550
},
{
"epoch": 0.2291494632535095,
"grad_norm": 2.708449916455307,
"learning_rate": 9.502607823452492e-06,
"loss": 0.6909,
"step": 555
},
{
"epoch": 0.23121387283236994,
"grad_norm": 2.6744723555279712,
"learning_rate": 9.486818638368037e-06,
"loss": 0.6727,
"step": 560
},
{
"epoch": 0.2332782824112304,
"grad_norm": 2.740519766192663,
"learning_rate": 9.47079628905222e-06,
"loss": 0.6643,
"step": 565
},
{
"epoch": 0.23534269199009084,
"grad_norm": 2.6127044855477544,
"learning_rate": 9.45454160813035e-06,
"loss": 0.641,
"step": 570
},
{
"epoch": 0.23740710156895128,
"grad_norm": 2.5898360039465547,
"learning_rate": 9.438055440301198e-06,
"loss": 0.654,
"step": 575
},
{
"epoch": 0.23947151114781173,
"grad_norm": 2.594315428105211,
"learning_rate": 9.421338642293096e-06,
"loss": 0.6612,
"step": 580
},
{
"epoch": 0.24153592072667218,
"grad_norm": 2.675279309210325,
"learning_rate": 9.404392082819418e-06,
"loss": 0.6835,
"step": 585
},
{
"epoch": 0.24360033030553263,
"grad_norm": 2.720481221371884,
"learning_rate": 9.387216642533436e-06,
"loss": 0.638,
"step": 590
},
{
"epoch": 0.24566473988439305,
"grad_norm": 2.7274539164197567,
"learning_rate": 9.369813213982557e-06,
"loss": 0.6742,
"step": 595
},
{
"epoch": 0.2477291494632535,
"grad_norm": 2.881159963630694,
"learning_rate": 9.352182701561932e-06,
"loss": 0.643,
"step": 600
},
{
"epoch": 0.24979355904211395,
"grad_norm": 2.672785271513205,
"learning_rate": 9.334326021467473e-06,
"loss": 0.6518,
"step": 605
},
{
"epoch": 0.2518579686209744,
"grad_norm": 2.6806455710314365,
"learning_rate": 9.316244101648229e-06,
"loss": 0.644,
"step": 610
},
{
"epoch": 0.2539223781998348,
"grad_norm": 2.5531072290720296,
"learning_rate": 9.297937881758165e-06,
"loss": 0.6738,
"step": 615
},
{
"epoch": 0.25598678777869527,
"grad_norm": 2.8562519058356335,
"learning_rate": 9.279408313107342e-06,
"loss": 0.6497,
"step": 620
},
{
"epoch": 0.2580511973575557,
"grad_norm": 2.9118129566241273,
"learning_rate": 9.260656358612461e-06,
"loss": 0.6544,
"step": 625
},
{
"epoch": 0.26011560693641617,
"grad_norm": 2.6582669873812117,
"learning_rate": 9.241682992746847e-06,
"loss": 0.6492,
"step": 630
},
{
"epoch": 0.2621800165152766,
"grad_norm": 3.135088950824475,
"learning_rate": 9.222489201489792e-06,
"loss": 0.6324,
"step": 635
},
{
"epoch": 0.26424442609413706,
"grad_norm": 2.5674033607777282,
"learning_rate": 9.203075982275323e-06,
"loss": 0.6461,
"step": 640
},
{
"epoch": 0.2663088356729975,
"grad_norm": 2.615329482313224,
"learning_rate": 9.18344434394037e-06,
"loss": 0.6322,
"step": 645
},
{
"epoch": 0.26837324525185796,
"grad_norm": 2.6662637432037144,
"learning_rate": 9.163595306672337e-06,
"loss": 0.6245,
"step": 650
},
{
"epoch": 0.2704376548307184,
"grad_norm": 2.788831071229617,
"learning_rate": 9.143529901956089e-06,
"loss": 0.6395,
"step": 655
},
{
"epoch": 0.27250206440957886,
"grad_norm": 2.7354638041171344,
"learning_rate": 9.12324917252035e-06,
"loss": 0.6239,
"step": 660
},
{
"epoch": 0.2745664739884393,
"grad_norm": 2.6074213157412682,
"learning_rate": 9.102754172283513e-06,
"loss": 0.6146,
"step": 665
},
{
"epoch": 0.27663088356729976,
"grad_norm": 2.6642140176865485,
"learning_rate": 9.082045966298871e-06,
"loss": 0.6207,
"step": 670
},
{
"epoch": 0.2786952931461602,
"grad_norm": 2.7423993871208547,
"learning_rate": 9.061125630699276e-06,
"loss": 0.6244,
"step": 675
},
{
"epoch": 0.28075970272502065,
"grad_norm": 2.639661943202539,
"learning_rate": 9.039994252641214e-06,
"loss": 0.6212,
"step": 680
},
{
"epoch": 0.2828241123038811,
"grad_norm": 2.7917609886441443,
"learning_rate": 9.018652930248299e-06,
"loss": 0.6279,
"step": 685
},
{
"epoch": 0.28488852188274155,
"grad_norm": 2.629871903545716,
"learning_rate": 8.997102772554226e-06,
"loss": 0.6229,
"step": 690
},
{
"epoch": 0.286952931461602,
"grad_norm": 2.625664182045412,
"learning_rate": 8.975344899445119e-06,
"loss": 0.6177,
"step": 695
},
{
"epoch": 0.28901734104046245,
"grad_norm": 2.80144553747866,
"learning_rate": 8.953380441601352e-06,
"loss": 0.6225,
"step": 700
},
{
"epoch": 0.2910817506193229,
"grad_norm": 2.672640679384628,
"learning_rate": 8.931210540438778e-06,
"loss": 0.6169,
"step": 705
},
{
"epoch": 0.2931461601981833,
"grad_norm": 2.709692648262891,
"learning_rate": 8.908836348049421e-06,
"loss": 0.6198,
"step": 710
},
{
"epoch": 0.29521056977704374,
"grad_norm": 2.681528317767069,
"learning_rate": 8.886259027141603e-06,
"loss": 0.6149,
"step": 715
},
{
"epoch": 0.2972749793559042,
"grad_norm": 3.03293644975451,
"learning_rate": 8.863479750979523e-06,
"loss": 0.6048,
"step": 720
},
{
"epoch": 0.29933938893476464,
"grad_norm": 2.670868153191048,
"learning_rate": 8.840499703322286e-06,
"loss": 0.6244,
"step": 725
},
{
"epoch": 0.3014037985136251,
"grad_norm": 2.7608395704880646,
"learning_rate": 8.817320078362388e-06,
"loss": 0.6035,
"step": 730
},
{
"epoch": 0.30346820809248554,
"grad_norm": 2.62588384982977,
"learning_rate": 8.793942080663658e-06,
"loss": 0.6172,
"step": 735
},
{
"epoch": 0.305532617671346,
"grad_norm": 2.5889068256927006,
"learning_rate": 8.77036692509866e-06,
"loss": 0.6128,
"step": 740
},
{
"epoch": 0.30759702725020643,
"grad_norm": 2.8092080143766216,
"learning_rate": 8.746595836785561e-06,
"loss": 0.6218,
"step": 745
},
{
"epoch": 0.3096614368290669,
"grad_norm": 2.6419072436310205,
"learning_rate": 8.722630051024467e-06,
"loss": 0.5877,
"step": 750
},
{
"epoch": 0.31172584640792733,
"grad_norm": 2.4997349328580363,
"learning_rate": 8.698470813233223e-06,
"loss": 0.6091,
"step": 755
},
{
"epoch": 0.3137902559867878,
"grad_norm": 2.6466710278909074,
"learning_rate": 8.674119378882707e-06,
"loss": 0.6087,
"step": 760
},
{
"epoch": 0.31585466556564823,
"grad_norm": 2.9297154437714363,
"learning_rate": 8.649577013431569e-06,
"loss": 0.592,
"step": 765
},
{
"epoch": 0.3179190751445087,
"grad_norm": 2.549997409403058,
"learning_rate": 8.624844992260483e-06,
"loss": 0.6023,
"step": 770
},
{
"epoch": 0.3199834847233691,
"grad_norm": 2.7080008761741827,
"learning_rate": 8.599924600605865e-06,
"loss": 0.6087,
"step": 775
},
{
"epoch": 0.3220478943022296,
"grad_norm": 2.567143331934158,
"learning_rate": 8.574817133493085e-06,
"loss": 0.5847,
"step": 780
},
{
"epoch": 0.32411230388109,
"grad_norm": 3.0113192932757418,
"learning_rate": 8.549523895669167e-06,
"loss": 0.5908,
"step": 785
},
{
"epoch": 0.3261767134599505,
"grad_norm": 2.7819521209578206,
"learning_rate": 8.52404620153499e-06,
"loss": 0.5953,
"step": 790
},
{
"epoch": 0.3282411230388109,
"grad_norm": 2.747216748021648,
"learning_rate": 8.498385375076979e-06,
"loss": 0.6114,
"step": 795
},
{
"epoch": 0.33030553261767137,
"grad_norm": 2.8476886670194523,
"learning_rate": 8.472542749798302e-06,
"loss": 0.5923,
"step": 800
},
{
"epoch": 0.33236994219653176,
"grad_norm": 3.078148326221217,
"learning_rate": 8.44651966864958e-06,
"loss": 0.5854,
"step": 805
},
{
"epoch": 0.3344343517753922,
"grad_norm": 2.7069355571781553,
"learning_rate": 8.420317483959086e-06,
"loss": 0.5963,
"step": 810
},
{
"epoch": 0.33649876135425266,
"grad_norm": 2.651998023029158,
"learning_rate": 8.39393755736248e-06,
"loss": 0.578,
"step": 815
},
{
"epoch": 0.3385631709331131,
"grad_norm": 2.6251615353567117,
"learning_rate": 8.367381259732042e-06,
"loss": 0.587,
"step": 820
},
{
"epoch": 0.34062758051197356,
"grad_norm": 2.5308341360997595,
"learning_rate": 8.340649971105444e-06,
"loss": 0.5732,
"step": 825
},
{
"epoch": 0.342691990090834,
"grad_norm": 2.823077213728183,
"learning_rate": 8.313745080614017e-06,
"loss": 0.5672,
"step": 830
},
{
"epoch": 0.34475639966969446,
"grad_norm": 2.659469137057744,
"learning_rate": 8.286667986410578e-06,
"loss": 0.5858,
"step": 835
},
{
"epoch": 0.3468208092485549,
"grad_norm": 2.845018948853141,
"learning_rate": 8.259420095596766e-06,
"loss": 0.5662,
"step": 840
},
{
"epoch": 0.34888521882741536,
"grad_norm": 2.6028478895400116,
"learning_rate": 8.232002824149917e-06,
"loss": 0.5606,
"step": 845
},
{
"epoch": 0.3509496284062758,
"grad_norm": 2.664709732227379,
"learning_rate": 8.204417596849492e-06,
"loss": 0.5807,
"step": 850
},
{
"epoch": 0.35301403798513625,
"grad_norm": 2.6109659834823478,
"learning_rate": 8.176665847203023e-06,
"loss": 0.5678,
"step": 855
},
{
"epoch": 0.3550784475639967,
"grad_norm": 2.576771155175357,
"learning_rate": 8.148749017371627e-06,
"loss": 0.5656,
"step": 860
},
{
"epoch": 0.35714285714285715,
"grad_norm": 2.6688801593883817,
"learning_rate": 8.120668558095056e-06,
"loss": 0.5749,
"step": 865
},
{
"epoch": 0.3592072667217176,
"grad_norm": 2.801881409304358,
"learning_rate": 8.092425928616317e-06,
"loss": 0.5603,
"step": 870
},
{
"epoch": 0.36127167630057805,
"grad_norm": 2.7158675577145512,
"learning_rate": 8.064022596605825e-06,
"loss": 0.5878,
"step": 875
},
{
"epoch": 0.3633360858794385,
"grad_norm": 2.7112475864205208,
"learning_rate": 8.03546003808515e-06,
"loss": 0.5697,
"step": 880
},
{
"epoch": 0.36540049545829895,
"grad_norm": 2.757081302105068,
"learning_rate": 8.006739737350303e-06,
"loss": 0.5549,
"step": 885
},
{
"epoch": 0.3674649050371594,
"grad_norm": 2.6798695533176,
"learning_rate": 7.977863186894598e-06,
"loss": 0.5621,
"step": 890
},
{
"epoch": 0.36952931461601984,
"grad_norm": 2.901794116790705,
"learning_rate": 7.948831887331113e-06,
"loss": 0.5531,
"step": 895
},
{
"epoch": 0.37159372419488024,
"grad_norm": 2.6209628611019333,
"learning_rate": 7.91964734731468e-06,
"loss": 0.5529,
"step": 900
},
{
"epoch": 0.3736581337737407,
"grad_norm": 2.631776999408232,
"learning_rate": 7.890311083463511e-06,
"loss": 0.5341,
"step": 905
},
{
"epoch": 0.37572254335260113,
"grad_norm": 2.6460839052940424,
"learning_rate": 7.860824620280367e-06,
"loss": 0.5589,
"step": 910
},
{
"epoch": 0.3777869529314616,
"grad_norm": 2.7027858769480604,
"learning_rate": 7.831189490073343e-06,
"loss": 0.5662,
"step": 915
},
{
"epoch": 0.37985136251032203,
"grad_norm": 2.686204866691535,
"learning_rate": 7.80140723287624e-06,
"loss": 0.5557,
"step": 920
},
{
"epoch": 0.3819157720891825,
"grad_norm": 2.537669294604416,
"learning_rate": 7.771479396368533e-06,
"loss": 0.5392,
"step": 925
},
{
"epoch": 0.38398018166804293,
"grad_norm": 2.634245727933428,
"learning_rate": 7.741407535794939e-06,
"loss": 0.5476,
"step": 930
},
{
"epoch": 0.3860445912469034,
"grad_norm": 2.63998365984595,
"learning_rate": 7.711193213884602e-06,
"loss": 0.5411,
"step": 935
},
{
"epoch": 0.38810900082576383,
"grad_norm": 2.5066641406139163,
"learning_rate": 7.680838000769889e-06,
"loss": 0.5436,
"step": 940
},
{
"epoch": 0.3901734104046243,
"grad_norm": 2.4698510563626384,
"learning_rate": 7.650343473904776e-06,
"loss": 0.536,
"step": 945
},
{
"epoch": 0.3922378199834847,
"grad_norm": 2.5102808206533753,
"learning_rate": 7.619711217982899e-06,
"loss": 0.5391,
"step": 950
},
{
"epoch": 0.3943022295623452,
"grad_norm": 2.648139467398325,
"learning_rate": 7.5889428248551854e-06,
"loss": 0.5368,
"step": 955
},
{
"epoch": 0.3963666391412056,
"grad_norm": 2.6648738415553765,
"learning_rate": 7.558039893447131e-06,
"loss": 0.5496,
"step": 960
},
{
"epoch": 0.39843104872006607,
"grad_norm": 2.6157509498551526,
"learning_rate": 7.527004029675722e-06,
"loss": 0.5318,
"step": 965
},
{
"epoch": 0.4004954582989265,
"grad_norm": 2.5975829607560574,
"learning_rate": 7.495836846365965e-06,
"loss": 0.5352,
"step": 970
},
{
"epoch": 0.40255986787778697,
"grad_norm": 2.92450937390704,
"learning_rate": 7.464539963167091e-06,
"loss": 0.5177,
"step": 975
},
{
"epoch": 0.4046242774566474,
"grad_norm": 2.7355952122295264,
"learning_rate": 7.433115006468373e-06,
"loss": 0.5322,
"step": 980
},
{
"epoch": 0.40668868703550787,
"grad_norm": 2.5829736114469184,
"learning_rate": 7.40156360931462e-06,
"loss": 0.5298,
"step": 985
},
{
"epoch": 0.4087530966143683,
"grad_norm": 2.705958846647064,
"learning_rate": 7.369887411321305e-06,
"loss": 0.5477,
"step": 990
},
{
"epoch": 0.4108175061932287,
"grad_norm": 2.4846959730474145,
"learning_rate": 7.338088058589362e-06,
"loss": 0.5324,
"step": 995
},
{
"epoch": 0.41288191577208916,
"grad_norm": 2.5252160694956762,
"learning_rate": 7.306167203619647e-06,
"loss": 0.5335,
"step": 1000
},
{
"epoch": 0.4149463253509496,
"grad_norm": 2.509240890701763,
"learning_rate": 7.27412650522706e-06,
"loss": 0.5219,
"step": 1005
},
{
"epoch": 0.41701073492981006,
"grad_norm": 2.567378202198639,
"learning_rate": 7.241967628454345e-06,
"loss": 0.5142,
"step": 1010
},
{
"epoch": 0.4190751445086705,
"grad_norm": 2.479116521394592,
"learning_rate": 7.20969224448556e-06,
"loss": 0.5198,
"step": 1015
},
{
"epoch": 0.42113955408753095,
"grad_norm": 2.456783739959803,
"learning_rate": 7.1773020305592355e-06,
"loss": 0.5141,
"step": 1020
},
{
"epoch": 0.4232039636663914,
"grad_norm": 2.6855797799019356,
"learning_rate": 7.1447986698812115e-06,
"loss": 0.5307,
"step": 1025
},
{
"epoch": 0.42526837324525185,
"grad_norm": 2.5323199215752203,
"learning_rate": 7.112183851537166e-06,
"loss": 0.5256,
"step": 1030
},
{
"epoch": 0.4273327828241123,
"grad_norm": 2.735418738176186,
"learning_rate": 7.079459270404842e-06,
"loss": 0.5112,
"step": 1035
},
{
"epoch": 0.42939719240297275,
"grad_norm": 2.6703449875287735,
"learning_rate": 7.046626627065974e-06,
"loss": 0.5099,
"step": 1040
},
{
"epoch": 0.4314616019818332,
"grad_norm": 2.4631394786044747,
"learning_rate": 7.013687627717905e-06,
"loss": 0.4978,
"step": 1045
},
{
"epoch": 0.43352601156069365,
"grad_norm": 2.8328555379260196,
"learning_rate": 6.980643984084927e-06,
"loss": 0.5179,
"step": 1050
},
{
"epoch": 0.4355904211395541,
"grad_norm": 2.504430665948721,
"learning_rate": 6.947497413329333e-06,
"loss": 0.5199,
"step": 1055
},
{
"epoch": 0.43765483071841454,
"grad_norm": 2.5711442943268663,
"learning_rate": 6.914249637962174e-06,
"loss": 0.5133,
"step": 1060
},
{
"epoch": 0.439719240297275,
"grad_norm": 2.5576807334846747,
"learning_rate": 6.880902385753749e-06,
"loss": 0.51,
"step": 1065
},
{
"epoch": 0.44178364987613544,
"grad_norm": 2.4378114750971833,
"learning_rate": 6.847457389643823e-06,
"loss": 0.4966,
"step": 1070
},
{
"epoch": 0.4438480594549959,
"grad_norm": 2.57922332477375,
"learning_rate": 6.8139163876515694e-06,
"loss": 0.5018,
"step": 1075
},
{
"epoch": 0.44591246903385634,
"grad_norm": 2.4923563722799145,
"learning_rate": 6.780281122785243e-06,
"loss": 0.5089,
"step": 1080
},
{
"epoch": 0.4479768786127168,
"grad_norm": 2.380488675820629,
"learning_rate": 6.74655334295162e-06,
"loss": 0.4969,
"step": 1085
},
{
"epoch": 0.4500412881915772,
"grad_norm": 2.622744707083321,
"learning_rate": 6.712734800865146e-06,
"loss": 0.5006,
"step": 1090
},
{
"epoch": 0.45210569777043763,
"grad_norm": 2.5279298037448514,
"learning_rate": 6.67882725395687e-06,
"loss": 0.4919,
"step": 1095
},
{
"epoch": 0.4541701073492981,
"grad_norm": 2.5696353167432076,
"learning_rate": 6.644832464283105e-06,
"loss": 0.5006,
"step": 1100
},
{
"epoch": 0.45623451692815853,
"grad_norm": 2.6472120693562675,
"learning_rate": 6.6107521984338654e-06,
"loss": 0.5222,
"step": 1105
},
{
"epoch": 0.458298926507019,
"grad_norm": 2.5641259833492915,
"learning_rate": 6.576588227441063e-06,
"loss": 0.5021,
"step": 1110
},
{
"epoch": 0.4603633360858794,
"grad_norm": 2.508674684822535,
"learning_rate": 6.5423423266864715e-06,
"loss": 0.5019,
"step": 1115
},
{
"epoch": 0.4624277456647399,
"grad_norm": 2.5803067400324546,
"learning_rate": 6.508016275809472e-06,
"loss": 0.5116,
"step": 1120
},
{
"epoch": 0.4644921552436003,
"grad_norm": 2.4783246238450296,
"learning_rate": 6.473611858614557e-06,
"loss": 0.4939,
"step": 1125
},
{
"epoch": 0.4665565648224608,
"grad_norm": 2.3961470840043986,
"learning_rate": 6.439130862978653e-06,
"loss": 0.4964,
"step": 1130
},
{
"epoch": 0.4686209744013212,
"grad_norm": 2.523268688947089,
"learning_rate": 6.4045750807581875e-06,
"loss": 0.4878,
"step": 1135
},
{
"epoch": 0.47068538398018167,
"grad_norm": 2.464908065215536,
"learning_rate": 6.369946307695994e-06,
"loss": 0.5004,
"step": 1140
},
{
"epoch": 0.4727497935590421,
"grad_norm": 2.4978854529114556,
"learning_rate": 6.335246343327983e-06,
"loss": 0.4928,
"step": 1145
},
{
"epoch": 0.47481420313790257,
"grad_norm": 2.5910220265362405,
"learning_rate": 6.300476990889622e-06,
"loss": 0.4859,
"step": 1150
},
{
"epoch": 0.476878612716763,
"grad_norm": 2.6474515998209314,
"learning_rate": 6.2656400572222374e-06,
"loss": 0.4968,
"step": 1155
},
{
"epoch": 0.47894302229562347,
"grad_norm": 2.5111508875593627,
"learning_rate": 6.230737352679115e-06,
"loss": 0.4825,
"step": 1160
},
{
"epoch": 0.4810074318744839,
"grad_norm": 2.6001140755401067,
"learning_rate": 6.1957706910314195e-06,
"loss": 0.4903,
"step": 1165
},
{
"epoch": 0.48307184145334436,
"grad_norm": 2.4405093084108853,
"learning_rate": 6.160741889373948e-06,
"loss": 0.4738,
"step": 1170
},
{
"epoch": 0.4851362510322048,
"grad_norm": 2.6122999506914817,
"learning_rate": 6.12565276803069e-06,
"loss": 0.4851,
"step": 1175
},
{
"epoch": 0.48720066061106526,
"grad_norm": 2.5491172182391284,
"learning_rate": 6.090505150460239e-06,
"loss": 0.4969,
"step": 1180
},
{
"epoch": 0.48926507018992565,
"grad_norm": 2.5615374385855256,
"learning_rate": 6.055300863161035e-06,
"loss": 0.4778,
"step": 1185
},
{
"epoch": 0.4913294797687861,
"grad_norm": 2.4469903993231514,
"learning_rate": 6.020041735576444e-06,
"loss": 0.4787,
"step": 1190
},
{
"epoch": 0.49339388934764655,
"grad_norm": 2.4860974782023346,
"learning_rate": 5.9847295999996835e-06,
"loss": 0.4914,
"step": 1195
},
{
"epoch": 0.495458298926507,
"grad_norm": 2.4900103589842395,
"learning_rate": 5.949366291478621e-06,
"loss": 0.4906,
"step": 1200
},
{
"epoch": 0.49752270850536745,
"grad_norm": 2.5283147963068964,
"learning_rate": 5.9139536477203945e-06,
"loss": 0.4759,
"step": 1205
},
{
"epoch": 0.4995871180842279,
"grad_norm": 2.573137486004109,
"learning_rate": 5.878493508995924e-06,
"loss": 0.4768,
"step": 1210
},
{
"epoch": 0.5016515276630884,
"grad_norm": 2.579476813837634,
"learning_rate": 5.842987718044275e-06,
"loss": 0.4679,
"step": 1215
},
{
"epoch": 0.5037159372419489,
"grad_norm": 2.497492462831134,
"learning_rate": 5.807438119976904e-06,
"loss": 0.4698,
"step": 1220
},
{
"epoch": 0.5057803468208093,
"grad_norm": 2.590655156263669,
"learning_rate": 5.771846562181761e-06,
"loss": 0.4728,
"step": 1225
},
{
"epoch": 0.5078447563996696,
"grad_norm": 2.5177297962486493,
"learning_rate": 5.736214894227304e-06,
"loss": 0.473,
"step": 1230
},
{
"epoch": 0.5099091659785301,
"grad_norm": 2.6021365054723775,
"learning_rate": 5.700544967766373e-06,
"loss": 0.4753,
"step": 1235
},
{
"epoch": 0.5119735755573905,
"grad_norm": 2.604951347166713,
"learning_rate": 5.6648386364399645e-06,
"loss": 0.47,
"step": 1240
},
{
"epoch": 0.514037985136251,
"grad_norm": 2.4157020679627794,
"learning_rate": 5.629097755780913e-06,
"loss": 0.4711,
"step": 1245
},
{
"epoch": 0.5161023947151114,
"grad_norm": 2.3659395254457167,
"learning_rate": 5.59332418311746e-06,
"loss": 0.4773,
"step": 1250
},
{
"epoch": 0.5181668042939719,
"grad_norm": 2.499361991412955,
"learning_rate": 5.557519777476733e-06,
"loss": 0.4527,
"step": 1255
},
{
"epoch": 0.5202312138728323,
"grad_norm": 2.5585105365426513,
"learning_rate": 5.521686399488145e-06,
"loss": 0.4573,
"step": 1260
},
{
"epoch": 0.5222956234516928,
"grad_norm": 2.4814508055210567,
"learning_rate": 5.485825911286697e-06,
"loss": 0.4691,
"step": 1265
},
{
"epoch": 0.5243600330305532,
"grad_norm": 2.6081291748684716,
"learning_rate": 5.449940176416213e-06,
"loss": 0.4526,
"step": 1270
},
{
"epoch": 0.5264244426094137,
"grad_norm": 2.587126119981003,
"learning_rate": 5.414031059732502e-06,
"loss": 0.4605,
"step": 1275
},
{
"epoch": 0.5284888521882741,
"grad_norm": 2.5458768564010765,
"learning_rate": 5.378100427306439e-06,
"loss": 0.4467,
"step": 1280
},
{
"epoch": 0.5305532617671346,
"grad_norm": 2.422344978356131,
"learning_rate": 5.3421501463269976e-06,
"loss": 0.4563,
"step": 1285
},
{
"epoch": 0.532617671345995,
"grad_norm": 2.481246363328885,
"learning_rate": 5.30618208500422e-06,
"loss": 0.4496,
"step": 1290
},
{
"epoch": 0.5346820809248555,
"grad_norm": 2.5964265890691274,
"learning_rate": 5.270198112472128e-06,
"loss": 0.4487,
"step": 1295
},
{
"epoch": 0.5367464905037159,
"grad_norm": 2.375375823310664,
"learning_rate": 5.234200098691597e-06,
"loss": 0.4492,
"step": 1300
},
{
"epoch": 0.5388109000825764,
"grad_norm": 2.482312531712164,
"learning_rate": 5.19818991435317e-06,
"loss": 0.4524,
"step": 1305
},
{
"epoch": 0.5408753096614368,
"grad_norm": 2.6268917122173154,
"learning_rate": 5.162169430779861e-06,
"loss": 0.4343,
"step": 1310
},
{
"epoch": 0.5429397192402973,
"grad_norm": 2.543083790336233,
"learning_rate": 5.126140519829888e-06,
"loss": 0.4738,
"step": 1315
},
{
"epoch": 0.5450041288191577,
"grad_norm": 2.4547121862030865,
"learning_rate": 5.090105053799418e-06,
"loss": 0.4604,
"step": 1320
},
{
"epoch": 0.5470685383980182,
"grad_norm": 2.5806078709462295,
"learning_rate": 5.054064905325262e-06,
"loss": 0.4571,
"step": 1325
},
{
"epoch": 0.5491329479768786,
"grad_norm": 2.538503365518751,
"learning_rate": 5.018021947287556e-06,
"loss": 0.4432,
"step": 1330
},
{
"epoch": 0.5511973575557391,
"grad_norm": 2.4205037183193157,
"learning_rate": 4.981978052712446e-06,
"loss": 0.4545,
"step": 1335
},
{
"epoch": 0.5532617671345995,
"grad_norm": 2.5910957054221817,
"learning_rate": 4.9459350946747405e-06,
"loss": 0.4584,
"step": 1340
},
{
"epoch": 0.55532617671346,
"grad_norm": 2.423593643872453,
"learning_rate": 4.909894946200583e-06,
"loss": 0.454,
"step": 1345
},
{
"epoch": 0.5573905862923204,
"grad_norm": 2.491439793223575,
"learning_rate": 4.873859480170113e-06,
"loss": 0.4535,
"step": 1350
},
{
"epoch": 0.5594549958711809,
"grad_norm": 2.6251023268801106,
"learning_rate": 4.83783056922014e-06,
"loss": 0.4475,
"step": 1355
},
{
"epoch": 0.5615194054500413,
"grad_norm": 2.379012672081149,
"learning_rate": 4.801810085646831e-06,
"loss": 0.4427,
"step": 1360
},
{
"epoch": 0.5635838150289018,
"grad_norm": 2.438307211574731,
"learning_rate": 4.765799901308405e-06,
"loss": 0.4392,
"step": 1365
},
{
"epoch": 0.5656482246077622,
"grad_norm": 2.4743966890129534,
"learning_rate": 4.729801887527872e-06,
"loss": 0.4376,
"step": 1370
},
{
"epoch": 0.5677126341866227,
"grad_norm": 2.4438402043185636,
"learning_rate": 4.6938179149957815e-06,
"loss": 0.4448,
"step": 1375
},
{
"epoch": 0.5697770437654831,
"grad_norm": 2.3233363056641587,
"learning_rate": 4.657849853673004e-06,
"loss": 0.4332,
"step": 1380
},
{
"epoch": 0.5718414533443436,
"grad_norm": 2.3513159156694536,
"learning_rate": 4.621899572693564e-06,
"loss": 0.4286,
"step": 1385
},
{
"epoch": 0.573905862923204,
"grad_norm": 2.452790907750448,
"learning_rate": 4.585968940267499e-06,
"loss": 0.4329,
"step": 1390
},
{
"epoch": 0.5759702725020645,
"grad_norm": 2.4876165011978544,
"learning_rate": 4.550059823583788e-06,
"loss": 0.4368,
"step": 1395
},
{
"epoch": 0.5780346820809249,
"grad_norm": 2.4110190508295926,
"learning_rate": 4.514174088713305e-06,
"loss": 0.4294,
"step": 1400
},
{
"epoch": 0.5800990916597853,
"grad_norm": 2.4790636104654724,
"learning_rate": 4.478313600511856e-06,
"loss": 0.438,
"step": 1405
},
{
"epoch": 0.5821635012386458,
"grad_norm": 2.5304633815274062,
"learning_rate": 4.442480222523269e-06,
"loss": 0.4383,
"step": 1410
},
{
"epoch": 0.5842279108175062,
"grad_norm": 2.365309103378621,
"learning_rate": 4.4066758168825414e-06,
"loss": 0.4383,
"step": 1415
},
{
"epoch": 0.5862923203963666,
"grad_norm": 2.4670833952729687,
"learning_rate": 4.370902244219088e-06,
"loss": 0.4389,
"step": 1420
},
{
"epoch": 0.588356729975227,
"grad_norm": 2.3390938676909734,
"learning_rate": 4.335161363560037e-06,
"loss": 0.4291,
"step": 1425
},
{
"epoch": 0.5904211395540875,
"grad_norm": 2.362644766829775,
"learning_rate": 4.299455032233631e-06,
"loss": 0.4288,
"step": 1430
},
{
"epoch": 0.5924855491329479,
"grad_norm": 2.445730007800526,
"learning_rate": 4.263785105772696e-06,
"loss": 0.4335,
"step": 1435
},
{
"epoch": 0.5945499587118084,
"grad_norm": 2.303973634970439,
"learning_rate": 4.22815343781824e-06,
"loss": 0.4366,
"step": 1440
},
{
"epoch": 0.5966143682906688,
"grad_norm": 2.391887494831502,
"learning_rate": 4.192561880023099e-06,
"loss": 0.4261,
"step": 1445
},
{
"epoch": 0.5986787778695293,
"grad_norm": 2.293302291662098,
"learning_rate": 4.157012281955726e-06,
"loss": 0.424,
"step": 1450
},
{
"epoch": 0.6007431874483897,
"grad_norm": 2.3712447663578655,
"learning_rate": 4.1215064910040795e-06,
"loss": 0.4314,
"step": 1455
},
{
"epoch": 0.6028075970272502,
"grad_norm": 2.381812200033987,
"learning_rate": 4.086046352279606e-06,
"loss": 0.4227,
"step": 1460
},
{
"epoch": 0.6048720066061106,
"grad_norm": 2.4743735981326114,
"learning_rate": 4.050633708521381e-06,
"loss": 0.4158,
"step": 1465
},
{
"epoch": 0.6069364161849711,
"grad_norm": 2.396322395044398,
"learning_rate": 4.015270400000317e-06,
"loss": 0.42,
"step": 1470
},
{
"epoch": 0.6090008257638315,
"grad_norm": 2.565114051952279,
"learning_rate": 3.979958264423558e-06,
"loss": 0.4155,
"step": 1475
},
{
"epoch": 0.611065235342692,
"grad_norm": 2.4934554445972954,
"learning_rate": 3.944699136838965e-06,
"loss": 0.4126,
"step": 1480
},
{
"epoch": 0.6131296449215524,
"grad_norm": 2.4162398419293187,
"learning_rate": 3.909494849539761e-06,
"loss": 0.4213,
"step": 1485
},
{
"epoch": 0.6151940545004129,
"grad_norm": 2.464180248251729,
"learning_rate": 3.874347231969312e-06,
"loss": 0.4235,
"step": 1490
},
{
"epoch": 0.6172584640792733,
"grad_norm": 2.4941582294474434,
"learning_rate": 3.839258110626053e-06,
"loss": 0.4307,
"step": 1495
},
{
"epoch": 0.6193228736581338,
"grad_norm": 2.276754213405921,
"learning_rate": 3.8042293089685813e-06,
"loss": 0.4065,
"step": 1500
},
{
"epoch": 0.6213872832369942,
"grad_norm": 2.324659425092549,
"learning_rate": 3.769262647320886e-06,
"loss": 0.4179,
"step": 1505
},
{
"epoch": 0.6234516928158547,
"grad_norm": 2.3013999943076815,
"learning_rate": 3.7343599427777634e-06,
"loss": 0.4157,
"step": 1510
},
{
"epoch": 0.6255161023947151,
"grad_norm": 2.3853366654087633,
"learning_rate": 3.69952300911038e-06,
"loss": 0.4087,
"step": 1515
},
{
"epoch": 0.6275805119735756,
"grad_norm": 2.3274220274703206,
"learning_rate": 3.6647536566720186e-06,
"loss": 0.4203,
"step": 1520
},
{
"epoch": 0.629644921552436,
"grad_norm": 2.4392433402643947,
"learning_rate": 3.630053692304005e-06,
"loss": 0.4091,
"step": 1525
},
{
"epoch": 0.6317093311312965,
"grad_norm": 2.4428540121878894,
"learning_rate": 3.5954249192418133e-06,
"loss": 0.4255,
"step": 1530
},
{
"epoch": 0.6337737407101569,
"grad_norm": 2.3810770043048692,
"learning_rate": 3.560869137021349e-06,
"loss": 0.4103,
"step": 1535
},
{
"epoch": 0.6358381502890174,
"grad_norm": 2.301863999851205,
"learning_rate": 3.526388141385444e-06,
"loss": 0.4155,
"step": 1540
},
{
"epoch": 0.6379025598678778,
"grad_norm": 2.400363519907276,
"learning_rate": 3.4919837241905308e-06,
"loss": 0.4224,
"step": 1545
},
{
"epoch": 0.6399669694467383,
"grad_norm": 2.477498090820749,
"learning_rate": 3.4576576733135284e-06,
"loss": 0.3938,
"step": 1550
},
{
"epoch": 0.6420313790255987,
"grad_norm": 2.707387721012179,
"learning_rate": 3.4234117725589385e-06,
"loss": 0.4111,
"step": 1555
},
{
"epoch": 0.6440957886044592,
"grad_norm": 2.341459106750255,
"learning_rate": 3.3892478015661362e-06,
"loss": 0.4139,
"step": 1560
},
{
"epoch": 0.6461601981833196,
"grad_norm": 2.4202740830209137,
"learning_rate": 3.355167535716897e-06,
"loss": 0.4127,
"step": 1565
},
{
"epoch": 0.64822460776218,
"grad_norm": 2.3816575196973577,
"learning_rate": 3.321172746043131e-06,
"loss": 0.4002,
"step": 1570
},
{
"epoch": 0.6502890173410405,
"grad_norm": 2.454435864594131,
"learning_rate": 3.2872651991348548e-06,
"loss": 0.403,
"step": 1575
},
{
"epoch": 0.652353426919901,
"grad_norm": 2.591163202406022,
"learning_rate": 3.253446657048382e-06,
"loss": 0.4072,
"step": 1580
},
{
"epoch": 0.6544178364987614,
"grad_norm": 2.4282921305894676,
"learning_rate": 3.2197188772147593e-06,
"loss": 0.3869,
"step": 1585
},
{
"epoch": 0.6564822460776218,
"grad_norm": 2.4682745998960827,
"learning_rate": 3.186083612348434e-06,
"loss": 0.4028,
"step": 1590
},
{
"epoch": 0.6585466556564823,
"grad_norm": 2.376999029803135,
"learning_rate": 3.1525426103561774e-06,
"loss": 0.4064,
"step": 1595
},
{
"epoch": 0.6606110652353427,
"grad_norm": 2.370838413822066,
"learning_rate": 3.119097614246252e-06,
"loss": 0.3978,
"step": 1600
},
{
"epoch": 0.6626754748142032,
"grad_norm": 2.362260369144417,
"learning_rate": 3.0857503620378284e-06,
"loss": 0.3905,
"step": 1605
},
{
"epoch": 0.6647398843930635,
"grad_norm": 2.421231306606393,
"learning_rate": 3.052502586670669e-06,
"loss": 0.4053,
"step": 1610
},
{
"epoch": 0.666804293971924,
"grad_norm": 2.395970410200359,
"learning_rate": 3.0193560159150747e-06,
"loss": 0.3943,
"step": 1615
},
{
"epoch": 0.6688687035507844,
"grad_norm": 2.5057121725436917,
"learning_rate": 2.986312372282097e-06,
"loss": 0.4182,
"step": 1620
},
{
"epoch": 0.6709331131296449,
"grad_norm": 2.3776161687385207,
"learning_rate": 2.9533733729340274e-06,
"loss": 0.4077,
"step": 1625
},
{
"epoch": 0.6729975227085053,
"grad_norm": 2.5915693629163217,
"learning_rate": 2.920540729595159e-06,
"loss": 0.4085,
"step": 1630
},
{
"epoch": 0.6750619322873658,
"grad_norm": 2.443160283840937,
"learning_rate": 2.8878161484628364e-06,
"loss": 0.4028,
"step": 1635
},
{
"epoch": 0.6771263418662262,
"grad_norm": 2.420398270818125,
"learning_rate": 2.8552013301187898e-06,
"loss": 0.3972,
"step": 1640
},
{
"epoch": 0.6791907514450867,
"grad_norm": 2.4196367974051265,
"learning_rate": 2.8226979694407657e-06,
"loss": 0.4067,
"step": 1645
},
{
"epoch": 0.6812551610239471,
"grad_norm": 2.438274119472032,
"learning_rate": 2.790307755514442e-06,
"loss": 0.4032,
"step": 1650
},
{
"epoch": 0.6833195706028076,
"grad_norm": 3.0023110501423944,
"learning_rate": 2.7580323715456564e-06,
"loss": 0.4121,
"step": 1655
},
{
"epoch": 0.685383980181668,
"grad_norm": 2.3597852522402474,
"learning_rate": 2.7258734947729428e-06,
"loss": 0.3938,
"step": 1660
},
{
"epoch": 0.6874483897605285,
"grad_norm": 2.4413156249499854,
"learning_rate": 2.6938327963803545e-06,
"loss": 0.3925,
"step": 1665
},
{
"epoch": 0.6895127993393889,
"grad_norm": 2.441149846551036,
"learning_rate": 2.6619119414106385e-06,
"loss": 0.3902,
"step": 1670
},
{
"epoch": 0.6915772089182494,
"grad_norm": 2.4274483132890685,
"learning_rate": 2.6301125886786965e-06,
"loss": 0.3933,
"step": 1675
},
{
"epoch": 0.6936416184971098,
"grad_norm": 2.514301355730002,
"learning_rate": 2.5984363906853814e-06,
"loss": 0.3847,
"step": 1680
},
{
"epoch": 0.6957060280759703,
"grad_norm": 2.23025942549343,
"learning_rate": 2.566884993531628e-06,
"loss": 0.3868,
"step": 1685
},
{
"epoch": 0.6977704376548307,
"grad_norm": 2.3910909958649764,
"learning_rate": 2.5354600368329108e-06,
"loss": 0.379,
"step": 1690
},
{
"epoch": 0.6998348472336912,
"grad_norm": 2.5413951969823554,
"learning_rate": 2.5041631536340374e-06,
"loss": 0.3909,
"step": 1695
},
{
"epoch": 0.7018992568125516,
"grad_norm": 2.360003867942285,
"learning_rate": 2.472995970324282e-06,
"loss": 0.3888,
"step": 1700
},
{
"epoch": 0.7039636663914121,
"grad_norm": 2.4621248065886516,
"learning_rate": 2.4419601065528703e-06,
"loss": 0.3803,
"step": 1705
},
{
"epoch": 0.7060280759702725,
"grad_norm": 2.4221633580536865,
"learning_rate": 2.4110571751448154e-06,
"loss": 0.3804,
"step": 1710
},
{
"epoch": 0.708092485549133,
"grad_norm": 2.3044799343595685,
"learning_rate": 2.3802887820171012e-06,
"loss": 0.3709,
"step": 1715
},
{
"epoch": 0.7101568951279934,
"grad_norm": 2.3101555840305585,
"learning_rate": 2.349656526095224e-06,
"loss": 0.382,
"step": 1720
},
{
"epoch": 0.7122213047068539,
"grad_norm": 2.439524803446898,
"learning_rate": 2.3191619992301144e-06,
"loss": 0.3969,
"step": 1725
},
{
"epoch": 0.7142857142857143,
"grad_norm": 2.3771584605894174,
"learning_rate": 2.2888067861153983e-06,
"loss": 0.3792,
"step": 1730
},
{
"epoch": 0.7163501238645748,
"grad_norm": 2.5666012183468587,
"learning_rate": 2.2585924642050638e-06,
"loss": 0.3802,
"step": 1735
},
{
"epoch": 0.7184145334434352,
"grad_norm": 2.3145141740784307,
"learning_rate": 2.2285206036314684e-06,
"loss": 0.3835,
"step": 1740
},
{
"epoch": 0.7204789430222956,
"grad_norm": 2.520653602273766,
"learning_rate": 2.1985927671237605e-06,
"loss": 0.3766,
"step": 1745
},
{
"epoch": 0.7225433526011561,
"grad_norm": 2.470460484865602,
"learning_rate": 2.168810509926659e-06,
"loss": 0.3897,
"step": 1750
},
{
"epoch": 0.7246077621800165,
"grad_norm": 2.408249418201136,
"learning_rate": 2.1391753797196328e-06,
"loss": 0.3952,
"step": 1755
},
{
"epoch": 0.726672171758877,
"grad_norm": 2.4402970653227247,
"learning_rate": 2.1096889165364894e-06,
"loss": 0.375,
"step": 1760
},
{
"epoch": 0.7287365813377374,
"grad_norm": 2.4710285995197188,
"learning_rate": 2.0803526526853206e-06,
"loss": 0.3748,
"step": 1765
},
{
"epoch": 0.7308009909165979,
"grad_norm": 2.2841746369286633,
"learning_rate": 2.0511681126688883e-06,
"loss": 0.3799,
"step": 1770
},
{
"epoch": 0.7328654004954583,
"grad_norm": 2.46102171882109,
"learning_rate": 2.022136813105401e-06,
"loss": 0.3729,
"step": 1775
},
{
"epoch": 0.7349298100743188,
"grad_norm": 2.2829340264643156,
"learning_rate": 1.9932602626496994e-06,
"loss": 0.3722,
"step": 1780
},
{
"epoch": 0.7369942196531792,
"grad_norm": 2.4439441720908,
"learning_rate": 1.9645399619148515e-06,
"loss": 0.3774,
"step": 1785
},
{
"epoch": 0.7390586292320397,
"grad_norm": 2.363486720634483,
"learning_rate": 1.9359774033941757e-06,
"loss": 0.3775,
"step": 1790
},
{
"epoch": 0.7411230388109001,
"grad_norm": 2.2893262909558065,
"learning_rate": 1.9075740713836854e-06,
"loss": 0.3851,
"step": 1795
},
{
"epoch": 0.7431874483897605,
"grad_norm": 2.381579584567696,
"learning_rate": 1.8793314419049446e-06,
"loss": 0.3736,
"step": 1800
},
{
"epoch": 0.7452518579686209,
"grad_norm": 2.3588892144435714,
"learning_rate": 1.8512509826283752e-06,
"loss": 0.3647,
"step": 1805
},
{
"epoch": 0.7473162675474814,
"grad_norm": 2.468444814623241,
"learning_rate": 1.8233341527969777e-06,
"loss": 0.3731,
"step": 1810
},
{
"epoch": 0.7493806771263418,
"grad_norm": 2.4216200751888217,
"learning_rate": 1.7955824031505097e-06,
"loss": 0.3741,
"step": 1815
},
{
"epoch": 0.7514450867052023,
"grad_norm": 2.3925423466955884,
"learning_rate": 1.7679971758500836e-06,
"loss": 0.3652,
"step": 1820
},
{
"epoch": 0.7535094962840627,
"grad_norm": 2.386508421187067,
"learning_rate": 1.7405799044032356e-06,
"loss": 0.3685,
"step": 1825
},
{
"epoch": 0.7555739058629232,
"grad_norm": 2.439989821363405,
"learning_rate": 1.7133320135894233e-06,
"loss": 0.3653,
"step": 1830
},
{
"epoch": 0.7576383154417836,
"grad_norm": 2.3039350711469186,
"learning_rate": 1.686254919385985e-06,
"loss": 0.3783,
"step": 1835
},
{
"epoch": 0.7597027250206441,
"grad_norm": 2.216200410510777,
"learning_rate": 1.6593500288945574e-06,
"loss": 0.3646,
"step": 1840
},
{
"epoch": 0.7617671345995045,
"grad_norm": 2.4989495842816405,
"learning_rate": 1.6326187402679577e-06,
"loss": 0.3741,
"step": 1845
},
{
"epoch": 0.763831544178365,
"grad_norm": 2.4154932431289673,
"learning_rate": 1.6060624426375222e-06,
"loss": 0.3698,
"step": 1850
},
{
"epoch": 0.7658959537572254,
"grad_norm": 2.473649546182009,
"learning_rate": 1.5796825160409168e-06,
"loss": 0.3672,
"step": 1855
},
{
"epoch": 0.7679603633360859,
"grad_norm": 2.355038901564694,
"learning_rate": 1.5534803313504215e-06,
"loss": 0.3566,
"step": 1860
},
{
"epoch": 0.7700247729149463,
"grad_norm": 2.3588413583977625,
"learning_rate": 1.5274572502016982e-06,
"loss": 0.3711,
"step": 1865
},
{
"epoch": 0.7720891824938068,
"grad_norm": 2.401065534914905,
"learning_rate": 1.5016146249230229e-06,
"loss": 0.3736,
"step": 1870
},
{
"epoch": 0.7741535920726672,
"grad_norm": 2.342195370497388,
"learning_rate": 1.4759537984650114e-06,
"loss": 0.3626,
"step": 1875
},
{
"epoch": 0.7762180016515277,
"grad_norm": 2.3601036223952634,
"learning_rate": 1.4504761043308346e-06,
"loss": 0.3534,
"step": 1880
},
{
"epoch": 0.7782824112303881,
"grad_norm": 2.3058771526048405,
"learning_rate": 1.425182866506918e-06,
"loss": 0.3633,
"step": 1885
},
{
"epoch": 0.7803468208092486,
"grad_norm": 2.308152081660379,
"learning_rate": 1.4000753993941369e-06,
"loss": 0.3723,
"step": 1890
},
{
"epoch": 0.782411230388109,
"grad_norm": 2.4604302877257664,
"learning_rate": 1.3751550077395181e-06,
"loss": 0.3523,
"step": 1895
},
{
"epoch": 0.7844756399669695,
"grad_norm": 2.551739458655361,
"learning_rate": 1.3504229865684326e-06,
"loss": 0.3706,
"step": 1900
},
{
"epoch": 0.7865400495458299,
"grad_norm": 2.4143979686568104,
"learning_rate": 1.3258806211172954e-06,
"loss": 0.3588,
"step": 1905
},
{
"epoch": 0.7886044591246903,
"grad_norm": 2.364426767604513,
"learning_rate": 1.3015291867667757e-06,
"loss": 0.3508,
"step": 1910
},
{
"epoch": 0.7906688687035508,
"grad_norm": 2.3233651206895147,
"learning_rate": 1.2773699489755343e-06,
"loss": 0.3571,
"step": 1915
},
{
"epoch": 0.7927332782824112,
"grad_norm": 2.3196733264987675,
"learning_rate": 1.2534041632144407e-06,
"loss": 0.3405,
"step": 1920
},
{
"epoch": 0.7947976878612717,
"grad_norm": 2.347467115991222,
"learning_rate": 1.2296330749013408e-06,
"loss": 0.3502,
"step": 1925
},
{
"epoch": 0.7968620974401321,
"grad_norm": 2.3004029534314174,
"learning_rate": 1.2060579193363431e-06,
"loss": 0.3564,
"step": 1930
},
{
"epoch": 0.7989265070189926,
"grad_norm": 2.5374233309055083,
"learning_rate": 1.1826799216376128e-06,
"loss": 0.3778,
"step": 1935
},
{
"epoch": 0.800990916597853,
"grad_norm": 2.4431483706896007,
"learning_rate": 1.1595002966777152e-06,
"loss": 0.356,
"step": 1940
},
{
"epoch": 0.8030553261767135,
"grad_norm": 2.35423434746266,
"learning_rate": 1.136520249020478e-06,
"loss": 0.3491,
"step": 1945
},
{
"epoch": 0.8051197357555739,
"grad_norm": 2.3767268596250672,
"learning_rate": 1.113740972858398e-06,
"loss": 0.348,
"step": 1950
},
{
"epoch": 0.8071841453344344,
"grad_norm": 2.4631890684934965,
"learning_rate": 1.0911636519505791e-06,
"loss": 0.3671,
"step": 1955
},
{
"epoch": 0.8092485549132948,
"grad_norm": 2.201423417502954,
"learning_rate": 1.068789459561223e-06,
"loss": 0.3459,
"step": 1960
},
{
"epoch": 0.8113129644921553,
"grad_norm": 2.36077824921438,
"learning_rate": 1.0466195583986487e-06,
"loss": 0.3639,
"step": 1965
},
{
"epoch": 0.8133773740710157,
"grad_norm": 2.354003763601073,
"learning_rate": 1.024655100554882e-06,
"loss": 0.3643,
"step": 1970
},
{
"epoch": 0.8154417836498762,
"grad_norm": 2.4101822702917004,
"learning_rate": 1.002897227445777e-06,
"loss": 0.3617,
"step": 1975
},
{
"epoch": 0.8175061932287366,
"grad_norm": 2.4833584786174163,
"learning_rate": 9.813470697517009e-07,
"loss": 0.3501,
"step": 1980
},
{
"epoch": 0.8195706028075971,
"grad_norm": 2.3627937000377393,
"learning_rate": 9.600057473587876e-07,
"loss": 0.3671,
"step": 1985
},
{
"epoch": 0.8216350123864574,
"grad_norm": 2.415763997145714,
"learning_rate": 9.388743693007247e-07,
"loss": 0.3541,
"step": 1990
},
{
"epoch": 0.8236994219653179,
"grad_norm": 2.4671751077694406,
"learning_rate": 9.1795403370113e-07,
"loss": 0.3535,
"step": 1995
},
{
"epoch": 0.8257638315441783,
"grad_norm": 2.3775799288896287,
"learning_rate": 8.972458277164886e-07,
"loss": 0.3516,
"step": 2000
},
{
"epoch": 0.8278282411230388,
"grad_norm": 2.2938470233646604,
"learning_rate": 8.76750827479651e-07,
"loss": 0.3398,
"step": 2005
},
{
"epoch": 0.8298926507018992,
"grad_norm": 2.2933382256010497,
"learning_rate": 8.564700980439122e-07,
"loss": 0.3608,
"step": 2010
},
{
"epoch": 0.8319570602807597,
"grad_norm": 2.3382609787220785,
"learning_rate": 8.364046933276642e-07,
"loss": 0.3666,
"step": 2015
},
{
"epoch": 0.8340214698596201,
"grad_norm": 2.44152050789561,
"learning_rate": 8.165556560596316e-07,
"loss": 0.3485,
"step": 2020
},
{
"epoch": 0.8360858794384806,
"grad_norm": 2.3858823443159034,
"learning_rate": 7.969240177246779e-07,
"loss": 0.3547,
"step": 2025
},
{
"epoch": 0.838150289017341,
"grad_norm": 2.453983334202833,
"learning_rate": 7.775107985102087e-07,
"loss": 0.3651,
"step": 2030
},
{
"epoch": 0.8402146985962015,
"grad_norm": 2.5029504947887875,
"learning_rate": 7.583170072531543e-07,
"loss": 0.3509,
"step": 2035
},
{
"epoch": 0.8422791081750619,
"grad_norm": 2.482629951124892,
"learning_rate": 7.393436413875411e-07,
"loss": 0.3476,
"step": 2040
},
{
"epoch": 0.8443435177539224,
"grad_norm": 2.283216056409345,
"learning_rate": 7.205916868926604e-07,
"loss": 0.3526,
"step": 2045
},
{
"epoch": 0.8464079273327828,
"grad_norm": 2.3590631737094845,
"learning_rate": 7.020621182418347e-07,
"loss": 0.34,
"step": 2050
},
{
"epoch": 0.8484723369116433,
"grad_norm": 2.6201296500024553,
"learning_rate": 6.837558983517723e-07,
"loss": 0.3535,
"step": 2055
},
{
"epoch": 0.8505367464905037,
"grad_norm": 2.4227114258373095,
"learning_rate": 6.656739785325278e-07,
"loss": 0.3507,
"step": 2060
},
{
"epoch": 0.8526011560693642,
"grad_norm": 2.390954329942295,
"learning_rate": 6.478172984380687e-07,
"loss": 0.3532,
"step": 2065
},
{
"epoch": 0.8546655656482246,
"grad_norm": 2.556923085001048,
"learning_rate": 6.301867860174443e-07,
"loss": 0.3434,
"step": 2070
},
{
"epoch": 0.856729975227085,
"grad_norm": 2.3169295549044335,
"learning_rate": 6.127833574665642e-07,
"loss": 0.3459,
"step": 2075
},
{
"epoch": 0.8587943848059455,
"grad_norm": 2.3497867089138644,
"learning_rate": 5.956079171805818e-07,
"loss": 0.3415,
"step": 2080
},
{
"epoch": 0.860858794384806,
"grad_norm": 2.1735667040509137,
"learning_rate": 5.78661357706905e-07,
"loss": 0.3379,
"step": 2085
},
{
"epoch": 0.8629232039636664,
"grad_norm": 2.5779842919745355,
"learning_rate": 5.619445596988026e-07,
"loss": 0.3425,
"step": 2090
},
{
"epoch": 0.8649876135425268,
"grad_norm": 2.401992025146321,
"learning_rate": 5.45458391869651e-07,
"loss": 0.3489,
"step": 2095
},
{
"epoch": 0.8670520231213873,
"grad_norm": 2.3105243423015587,
"learning_rate": 5.29203710947781e-07,
"loss": 0.335,
"step": 2100
},
{
"epoch": 0.8691164327002477,
"grad_norm": 2.5027833045418943,
"learning_rate": 5.131813616319641e-07,
"loss": 0.35,
"step": 2105
},
{
"epoch": 0.8711808422791082,
"grad_norm": 2.492666189587263,
"learning_rate": 4.973921765475093e-07,
"loss": 0.3419,
"step": 2110
},
{
"epoch": 0.8732452518579686,
"grad_norm": 2.476326337718451,
"learning_rate": 4.818369762030001e-07,
"loss": 0.3437,
"step": 2115
},
{
"epoch": 0.8753096614368291,
"grad_norm": 2.4895934072647923,
"learning_rate": 4.6651656894765307e-07,
"loss": 0.345,
"step": 2120
},
{
"epoch": 0.8773740710156895,
"grad_norm": 2.432351776306209,
"learning_rate": 4.5143175092931257e-07,
"loss": 0.3457,
"step": 2125
},
{
"epoch": 0.87943848059455,
"grad_norm": 2.5446256871769894,
"learning_rate": 4.3658330605307496e-07,
"loss": 0.3501,
"step": 2130
},
{
"epoch": 0.8815028901734104,
"grad_norm": 2.400237240106358,
"learning_rate": 4.2197200594055135e-07,
"loss": 0.3417,
"step": 2135
},
{
"epoch": 0.8835672997522709,
"grad_norm": 2.420863056343347,
"learning_rate": 4.0759860988977715e-07,
"loss": 0.3336,
"step": 2140
},
{
"epoch": 0.8856317093311313,
"grad_norm": 2.5855831775345317,
"learning_rate": 3.934638648357414e-07,
"loss": 0.3473,
"step": 2145
},
{
"epoch": 0.8876961189099918,
"grad_norm": 2.4571094742538633,
"learning_rate": 3.795685053115805e-07,
"loss": 0.3425,
"step": 2150
},
{
"epoch": 0.8897605284888522,
"grad_norm": 2.318969751255526,
"learning_rate": 3.659132534104054e-07,
"loss": 0.339,
"step": 2155
},
{
"epoch": 0.8918249380677127,
"grad_norm": 2.464561015161846,
"learning_rate": 3.5249881874777235e-07,
"loss": 0.3415,
"step": 2160
},
{
"epoch": 0.8938893476465731,
"grad_norm": 2.5739185699856653,
"learning_rate": 3.3932589842481434e-07,
"loss": 0.3549,
"step": 2165
},
{
"epoch": 0.8959537572254336,
"grad_norm": 2.3924860465104554,
"learning_rate": 3.2639517699200765e-07,
"loss": 0.3501,
"step": 2170
},
{
"epoch": 0.898018166804294,
"grad_norm": 2.4561049545962983,
"learning_rate": 3.1370732641360424e-07,
"loss": 0.3382,
"step": 2175
},
{
"epoch": 0.9000825763831544,
"grad_norm": 2.4531360622529985,
"learning_rate": 3.0126300603270763e-07,
"loss": 0.3386,
"step": 2180
},
{
"epoch": 0.9021469859620148,
"grad_norm": 2.3726671856862698,
"learning_rate": 2.89062862537009e-07,
"loss": 0.3382,
"step": 2185
},
{
"epoch": 0.9042113955408753,
"grad_norm": 2.4508531032953704,
"learning_rate": 2.7710752992518906e-07,
"loss": 0.3475,
"step": 2190
},
{
"epoch": 0.9062758051197357,
"grad_norm": 2.344783384097571,
"learning_rate": 2.653976294739596e-07,
"loss": 0.3435,
"step": 2195
},
{
"epoch": 0.9083402146985962,
"grad_norm": 2.3832943118751495,
"learning_rate": 2.539337697057853e-07,
"loss": 0.3386,
"step": 2200
},
{
"epoch": 0.9104046242774566,
"grad_norm": 2.2656029102160917,
"learning_rate": 2.4271654635726003e-07,
"loss": 0.326,
"step": 2205
},
{
"epoch": 0.9124690338563171,
"grad_norm": 2.531439971555802,
"learning_rate": 2.3174654234814854e-07,
"loss": 0.3336,
"step": 2210
},
{
"epoch": 0.9145334434351775,
"grad_norm": 2.526736652879497,
"learning_rate": 2.2102432775109151e-07,
"loss": 0.3431,
"step": 2215
},
{
"epoch": 0.916597853014038,
"grad_norm": 2.499234877151574,
"learning_rate": 2.1055045976198286e-07,
"loss": 0.3274,
"step": 2220
},
{
"epoch": 0.9186622625928984,
"grad_norm": 2.347953826058282,
"learning_rate": 2.0032548267101547e-07,
"loss": 0.3415,
"step": 2225
},
{
"epoch": 0.9207266721717589,
"grad_norm": 2.2781319110873923,
"learning_rate": 1.9034992783439398e-07,
"loss": 0.3313,
"step": 2230
},
{
"epoch": 0.9227910817506193,
"grad_norm": 2.4985215803253777,
"learning_rate": 1.8062431364672394e-07,
"loss": 0.347,
"step": 2235
},
{
"epoch": 0.9248554913294798,
"grad_norm": 2.3523550820055807,
"learning_rate": 1.7114914551407125e-07,
"loss": 0.3405,
"step": 2240
},
{
"epoch": 0.9269199009083402,
"grad_norm": 2.35149253079431,
"learning_rate": 1.619249158276981e-07,
"loss": 0.3396,
"step": 2245
},
{
"epoch": 0.9289843104872006,
"grad_norm": 2.5391046088636724,
"learning_rate": 1.529521039384757e-07,
"loss": 0.3382,
"step": 2250
},
{
"epoch": 0.9310487200660611,
"grad_norm": 2.436611958199382,
"learning_rate": 1.4423117613197535e-07,
"loss": 0.3507,
"step": 2255
},
{
"epoch": 0.9331131296449215,
"grad_norm": 2.368900058293189,
"learning_rate": 1.357625856042344e-07,
"loss": 0.3436,
"step": 2260
},
{
"epoch": 0.935177539223782,
"grad_norm": 2.483713260867796,
"learning_rate": 1.275467724382068e-07,
"loss": 0.3493,
"step": 2265
},
{
"epoch": 0.9372419488026424,
"grad_norm": 2.3378032035177556,
"learning_rate": 1.1958416358089308e-07,
"loss": 0.3366,
"step": 2270
},
{
"epoch": 0.9393063583815029,
"grad_norm": 2.427445932737846,
"learning_rate": 1.1187517282115478e-07,
"loss": 0.3357,
"step": 2275
},
{
"epoch": 0.9413707679603633,
"grad_norm": 2.45979849427157,
"learning_rate": 1.0442020076820992e-07,
"loss": 0.3432,
"step": 2280
},
{
"epoch": 0.9434351775392238,
"grad_norm": 2.4102972284673485,
"learning_rate": 9.72196348308152e-08,
"loss": 0.3421,
"step": 2285
},
{
"epoch": 0.9454995871180842,
"grad_norm": 2.4147926771751127,
"learning_rate": 9.027384919713167e-08,
"loss": 0.3449,
"step": 2290
},
{
"epoch": 0.9475639966969447,
"grad_norm": 2.492908131419299,
"learning_rate": 8.358320481528404e-08,
"loss": 0.3349,
"step": 2295
},
{
"epoch": 0.9496284062758051,
"grad_norm": 2.445696827488344,
"learning_rate": 7.714804937459964e-08,
"loss": 0.3327,
"step": 2300
},
{
"epoch": 0.9516928158546656,
"grad_norm": 2.4859237981346434,
"learning_rate": 7.096871728754173e-08,
"loss": 0.3369,
"step": 2305
},
{
"epoch": 0.953757225433526,
"grad_norm": 2.4755163895408483,
"learning_rate": 6.504552967233014e-08,
"loss": 0.3341,
"step": 2310
},
{
"epoch": 0.9558216350123865,
"grad_norm": 2.4903780675355947,
"learning_rate": 5.9378794336255595e-08,
"loss": 0.3259,
"step": 2315
},
{
"epoch": 0.9578860445912469,
"grad_norm": 2.3608897794698986,
"learning_rate": 5.396880575968266e-08,
"loss": 0.34,
"step": 2320
},
{
"epoch": 0.9599504541701074,
"grad_norm": 2.349066629113589,
"learning_rate": 4.881584508074688e-08,
"loss": 0.3285,
"step": 2325
},
{
"epoch": 0.9620148637489678,
"grad_norm": 2.475193042134275,
"learning_rate": 4.39201800807465e-08,
"loss": 0.339,
"step": 2330
},
{
"epoch": 0.9640792733278283,
"grad_norm": 2.3656425365762384,
"learning_rate": 3.92820651702247e-08,
"loss": 0.3358,
"step": 2335
},
{
"epoch": 0.9661436829066887,
"grad_norm": 2.441572724635035,
"learning_rate": 3.4901741375747975e-08,
"loss": 0.3471,
"step": 2340
},
{
"epoch": 0.9682080924855492,
"grad_norm": 2.3678044079850413,
"learning_rate": 3.077943632738556e-08,
"loss": 0.3499,
"step": 2345
},
{
"epoch": 0.9702725020644096,
"grad_norm": 2.630292946781589,
"learning_rate": 2.6915364246875574e-08,
"loss": 0.341,
"step": 2350
},
{
"epoch": 0.9723369116432701,
"grad_norm": 2.388897601971072,
"learning_rate": 2.3309725936493387e-08,
"loss": 0.3437,
"step": 2355
},
{
"epoch": 0.9744013212221305,
"grad_norm": 2.4665523794028785,
"learning_rate": 1.9962708768620498e-08,
"loss": 0.3453,
"step": 2360
},
{
"epoch": 0.976465730800991,
"grad_norm": 2.448813469296534,
"learning_rate": 1.687448667600067e-08,
"loss": 0.3408,
"step": 2365
},
{
"epoch": 0.9785301403798513,
"grad_norm": 2.3718128749961576,
"learning_rate": 1.4045220142708816e-08,
"loss": 0.35,
"step": 2370
},
{
"epoch": 0.9805945499587118,
"grad_norm": 2.4841177332012325,
"learning_rate": 1.1475056195804358e-08,
"loss": 0.3518,
"step": 2375
},
{
"epoch": 0.9826589595375722,
"grad_norm": 2.493744247830282,
"learning_rate": 9.164128397695648e-09,
"loss": 0.332,
"step": 2380
},
{
"epoch": 0.9847233691164327,
"grad_norm": 2.2819847476640107,
"learning_rate": 7.112556839196649e-09,
"loss": 0.3301,
"step": 2385
},
{
"epoch": 0.9867877786952931,
"grad_norm": 2.4247247484873173,
"learning_rate": 5.3204481332880255e-09,
"loss": 0.3433,
"step": 2390
},
{
"epoch": 0.9888521882741536,
"grad_norm": 2.4373319136694285,
"learning_rate": 3.787895409574915e-09,
"loss": 0.3388,
"step": 2395
},
{
"epoch": 0.990916597853014,
"grad_norm": 2.3711173222964037,
"learning_rate": 2.514978309447469e-09,
"loss": 0.3433,
"step": 2400
},
{
"epoch": 0.9929810074318745,
"grad_norm": 2.3426000581900652,
"learning_rate": 1.501762981945265e-09,
"loss": 0.3368,
"step": 2405
},
{
"epoch": 0.9950454170107349,
"grad_norm": 2.35822764782016,
"learning_rate": 7.483020803145114e-10,
"loss": 0.3396,
"step": 2410
},
{
"epoch": 0.9971098265895953,
"grad_norm": 2.4725534953927233,
"learning_rate": 2.5463475927745007e-10,
"loss": 0.3401,
"step": 2415
},
{
"epoch": 0.9991742361684558,
"grad_norm": 2.407515233988016,
"learning_rate": 2.078667299343362e-11,
"loss": 0.3298,
"step": 2420
},
{
"epoch": 1.0,
"eval_runtime": 3.3946,
"eval_samples_per_second": 2.946,
"eval_steps_per_second": 0.884,
"step": 2422
},
{
"epoch": 1.0,
"step": 2422,
"total_flos": 253558763028480.0,
"train_loss": 0.530042272082643,
"train_runtime": 21564.4437,
"train_samples_per_second": 1.797,
"train_steps_per_second": 0.112
}
],
"logging_steps": 5,
"max_steps": 2422,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 253558763028480.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}