nthakur's picture
Model save
7a1a66b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 200,
"global_step": 4852,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00020610057708161583,
"grad_norm": 0.7938007287198758,
"learning_rate": 4.1152263374485604e-07,
"loss": 1.2403,
"step": 1
},
{
"epoch": 0.001030502885408079,
"grad_norm": 0.9829139995457005,
"learning_rate": 2.05761316872428e-06,
"loss": 1.5592,
"step": 5
},
{
"epoch": 0.002061005770816158,
"grad_norm": 1.3104221596563024,
"learning_rate": 4.11522633744856e-06,
"loss": 1.9165,
"step": 10
},
{
"epoch": 0.0030915086562242375,
"grad_norm": 0.8263152262533461,
"learning_rate": 6.172839506172839e-06,
"loss": 1.6148,
"step": 15
},
{
"epoch": 0.004122011541632316,
"grad_norm": 1.2188669223330737,
"learning_rate": 8.23045267489712e-06,
"loss": 1.7383,
"step": 20
},
{
"epoch": 0.005152514427040396,
"grad_norm": 1.3669452124425536,
"learning_rate": 1.02880658436214e-05,
"loss": 1.8777,
"step": 25
},
{
"epoch": 0.006183017312448475,
"grad_norm": 0.6877557296913395,
"learning_rate": 1.2345679012345678e-05,
"loss": 1.3855,
"step": 30
},
{
"epoch": 0.007213520197856554,
"grad_norm": 0.6292369266751937,
"learning_rate": 1.440329218106996e-05,
"loss": 1.7622,
"step": 35
},
{
"epoch": 0.008244023083264633,
"grad_norm": 0.43117531258376596,
"learning_rate": 1.646090534979424e-05,
"loss": 1.4525,
"step": 40
},
{
"epoch": 0.009274525968672712,
"grad_norm": 0.3651450490233128,
"learning_rate": 1.8518518518518518e-05,
"loss": 1.546,
"step": 45
},
{
"epoch": 0.010305028854080791,
"grad_norm": 0.4844380918063832,
"learning_rate": 2.05761316872428e-05,
"loss": 1.6499,
"step": 50
},
{
"epoch": 0.01133553173948887,
"grad_norm": 0.33393658192356523,
"learning_rate": 2.2633744855967078e-05,
"loss": 1.1817,
"step": 55
},
{
"epoch": 0.01236603462489695,
"grad_norm": 0.34761501749088797,
"learning_rate": 2.4691358024691357e-05,
"loss": 1.5655,
"step": 60
},
{
"epoch": 0.01339653751030503,
"grad_norm": 0.3221134242168732,
"learning_rate": 2.6748971193415638e-05,
"loss": 1.2608,
"step": 65
},
{
"epoch": 0.014427040395713108,
"grad_norm": 0.31264722399213424,
"learning_rate": 2.880658436213992e-05,
"loss": 1.4104,
"step": 70
},
{
"epoch": 0.015457543281121188,
"grad_norm": 0.3955664926406516,
"learning_rate": 3.08641975308642e-05,
"loss": 1.4622,
"step": 75
},
{
"epoch": 0.016488046166529265,
"grad_norm": 0.3149436282257214,
"learning_rate": 3.292181069958848e-05,
"loss": 1.159,
"step": 80
},
{
"epoch": 0.017518549051937345,
"grad_norm": 0.2854140231011414,
"learning_rate": 3.497942386831276e-05,
"loss": 1.4829,
"step": 85
},
{
"epoch": 0.018549051937345424,
"grad_norm": 0.3326334208099821,
"learning_rate": 3.7037037037037037e-05,
"loss": 1.1516,
"step": 90
},
{
"epoch": 0.019579554822753503,
"grad_norm": 0.27192427116328566,
"learning_rate": 3.909465020576132e-05,
"loss": 1.3785,
"step": 95
},
{
"epoch": 0.020610057708161583,
"grad_norm": 0.3152208635246729,
"learning_rate": 4.11522633744856e-05,
"loss": 1.3886,
"step": 100
},
{
"epoch": 0.021640560593569662,
"grad_norm": 0.27663290557612785,
"learning_rate": 4.3209876543209875e-05,
"loss": 1.1185,
"step": 105
},
{
"epoch": 0.02267106347897774,
"grad_norm": 0.27185356829855356,
"learning_rate": 4.5267489711934157e-05,
"loss": 1.4773,
"step": 110
},
{
"epoch": 0.02370156636438582,
"grad_norm": 0.2935081378811297,
"learning_rate": 4.732510288065844e-05,
"loss": 1.261,
"step": 115
},
{
"epoch": 0.0247320692497939,
"grad_norm": 0.29236576081560334,
"learning_rate": 4.938271604938271e-05,
"loss": 1.3387,
"step": 120
},
{
"epoch": 0.02576257213520198,
"grad_norm": 0.38637732181078555,
"learning_rate": 5.1440329218106995e-05,
"loss": 1.4305,
"step": 125
},
{
"epoch": 0.02679307502061006,
"grad_norm": 0.2982040676995298,
"learning_rate": 5.3497942386831277e-05,
"loss": 1.0162,
"step": 130
},
{
"epoch": 0.027823577906018138,
"grad_norm": 0.2699066491657892,
"learning_rate": 5.555555555555556e-05,
"loss": 1.4066,
"step": 135
},
{
"epoch": 0.028854080791426217,
"grad_norm": 0.33203812944453787,
"learning_rate": 5.761316872427984e-05,
"loss": 1.1751,
"step": 140
},
{
"epoch": 0.029884583676834296,
"grad_norm": 0.3312647784314266,
"learning_rate": 5.9670781893004115e-05,
"loss": 1.2269,
"step": 145
},
{
"epoch": 0.030915086562242376,
"grad_norm": 0.541956195146165,
"learning_rate": 6.17283950617284e-05,
"loss": 1.354,
"step": 150
},
{
"epoch": 0.031945589447650455,
"grad_norm": 0.3172355233573687,
"learning_rate": 6.378600823045267e-05,
"loss": 1.0666,
"step": 155
},
{
"epoch": 0.03297609233305853,
"grad_norm": 0.32272337262102746,
"learning_rate": 6.584362139917696e-05,
"loss": 1.4928,
"step": 160
},
{
"epoch": 0.03400659521846661,
"grad_norm": 0.4252299718796606,
"learning_rate": 6.790123456790123e-05,
"loss": 1.1979,
"step": 165
},
{
"epoch": 0.03503709810387469,
"grad_norm": 0.2816942875302398,
"learning_rate": 6.995884773662552e-05,
"loss": 1.3223,
"step": 170
},
{
"epoch": 0.03606760098928277,
"grad_norm": 0.34311165874954624,
"learning_rate": 7.20164609053498e-05,
"loss": 1.4296,
"step": 175
},
{
"epoch": 0.03709810387469085,
"grad_norm": 0.297834971078222,
"learning_rate": 7.407407407407407e-05,
"loss": 1.087,
"step": 180
},
{
"epoch": 0.03812860676009893,
"grad_norm": 0.2602802208757066,
"learning_rate": 7.613168724279836e-05,
"loss": 1.4754,
"step": 185
},
{
"epoch": 0.039159109645507006,
"grad_norm": 0.40680201193562965,
"learning_rate": 7.818930041152264e-05,
"loss": 1.1889,
"step": 190
},
{
"epoch": 0.04018961253091509,
"grad_norm": 0.45086574163739,
"learning_rate": 8.024691358024692e-05,
"loss": 1.2851,
"step": 195
},
{
"epoch": 0.041220115416323165,
"grad_norm": 0.4308650820975275,
"learning_rate": 8.23045267489712e-05,
"loss": 1.3478,
"step": 200
},
{
"epoch": 0.041220115416323165,
"eval_loss": 1.2309638261795044,
"eval_runtime": 2469.2141,
"eval_samples_per_second": 3.24,
"eval_steps_per_second": 0.202,
"step": 200
},
{
"epoch": 0.04225061830173125,
"grad_norm": 0.37845098126001203,
"learning_rate": 8.436213991769549e-05,
"loss": 1.0707,
"step": 205
},
{
"epoch": 0.043281121187139324,
"grad_norm": 0.2851215824173796,
"learning_rate": 8.641975308641975e-05,
"loss": 1.4131,
"step": 210
},
{
"epoch": 0.044311624072547406,
"grad_norm": 0.3309919261494515,
"learning_rate": 8.847736625514404e-05,
"loss": 1.1951,
"step": 215
},
{
"epoch": 0.04534212695795548,
"grad_norm": 0.3065911094811643,
"learning_rate": 9.053497942386831e-05,
"loss": 1.2554,
"step": 220
},
{
"epoch": 0.04637262984336356,
"grad_norm": 0.32586015367310234,
"learning_rate": 9.25925925925926e-05,
"loss": 1.4263,
"step": 225
},
{
"epoch": 0.04740313272877164,
"grad_norm": 0.277075233563208,
"learning_rate": 9.465020576131688e-05,
"loss": 1.0216,
"step": 230
},
{
"epoch": 0.04843363561417972,
"grad_norm": 0.2769072933842902,
"learning_rate": 9.670781893004116e-05,
"loss": 1.3689,
"step": 235
},
{
"epoch": 0.0494641384995878,
"grad_norm": 0.27922329154610587,
"learning_rate": 9.876543209876543e-05,
"loss": 1.2181,
"step": 240
},
{
"epoch": 0.050494641384995875,
"grad_norm": 0.30982282181292814,
"learning_rate": 0.00010082304526748971,
"loss": 1.2183,
"step": 245
},
{
"epoch": 0.05152514427040396,
"grad_norm": 0.33257211954241994,
"learning_rate": 0.00010288065843621399,
"loss": 1.4305,
"step": 250
},
{
"epoch": 0.052555647155812034,
"grad_norm": 0.2838288504939334,
"learning_rate": 0.00010493827160493828,
"loss": 1.0051,
"step": 255
},
{
"epoch": 0.05358615004122012,
"grad_norm": 0.28623162042613864,
"learning_rate": 0.00010699588477366255,
"loss": 1.436,
"step": 260
},
{
"epoch": 0.05461665292662819,
"grad_norm": 0.2720219498268959,
"learning_rate": 0.00010905349794238684,
"loss": 1.1381,
"step": 265
},
{
"epoch": 0.055647155812036275,
"grad_norm": 0.30404224920143047,
"learning_rate": 0.00011111111111111112,
"loss": 1.1905,
"step": 270
},
{
"epoch": 0.05667765869744435,
"grad_norm": 0.3469618797327675,
"learning_rate": 0.0001131687242798354,
"loss": 1.3797,
"step": 275
},
{
"epoch": 0.057708161582852434,
"grad_norm": 0.30848130199190793,
"learning_rate": 0.00011522633744855968,
"loss": 1.0163,
"step": 280
},
{
"epoch": 0.05873866446826051,
"grad_norm": 0.2587435053896794,
"learning_rate": 0.00011728395061728397,
"loss": 1.401,
"step": 285
},
{
"epoch": 0.05976916735366859,
"grad_norm": 0.30665664697045764,
"learning_rate": 0.00011934156378600823,
"loss": 1.197,
"step": 290
},
{
"epoch": 0.06079967023907667,
"grad_norm": 0.28688303187237163,
"learning_rate": 0.0001213991769547325,
"loss": 1.2817,
"step": 295
},
{
"epoch": 0.06183017312448475,
"grad_norm": 0.32502880249151705,
"learning_rate": 0.0001234567901234568,
"loss": 1.388,
"step": 300
},
{
"epoch": 0.06286067600989283,
"grad_norm": 0.30699533228385806,
"learning_rate": 0.00012551440329218108,
"loss": 1.0403,
"step": 305
},
{
"epoch": 0.06389117889530091,
"grad_norm": 0.29902009304688737,
"learning_rate": 0.00012757201646090534,
"loss": 1.3604,
"step": 310
},
{
"epoch": 0.06492168178070899,
"grad_norm": 0.26224004465269324,
"learning_rate": 0.00012962962962962963,
"loss": 1.12,
"step": 315
},
{
"epoch": 0.06595218466611706,
"grad_norm": 0.2697856637362638,
"learning_rate": 0.00013168724279835392,
"loss": 1.295,
"step": 320
},
{
"epoch": 0.06698268755152514,
"grad_norm": 0.3118309921774858,
"learning_rate": 0.0001337448559670782,
"loss": 1.3498,
"step": 325
},
{
"epoch": 0.06801319043693323,
"grad_norm": 0.3774735054690596,
"learning_rate": 0.00013580246913580247,
"loss": 1.1412,
"step": 330
},
{
"epoch": 0.0690436933223413,
"grad_norm": 0.27963249445347155,
"learning_rate": 0.00013786008230452676,
"loss": 1.4175,
"step": 335
},
{
"epoch": 0.07007419620774938,
"grad_norm": 0.25343833022804746,
"learning_rate": 0.00013991769547325105,
"loss": 1.1228,
"step": 340
},
{
"epoch": 0.07110469909315746,
"grad_norm": 0.2676785839608282,
"learning_rate": 0.00014197530864197534,
"loss": 1.2235,
"step": 345
},
{
"epoch": 0.07213520197856554,
"grad_norm": 0.33130141183815665,
"learning_rate": 0.0001440329218106996,
"loss": 1.3623,
"step": 350
},
{
"epoch": 0.07316570486397361,
"grad_norm": 0.2676095204791478,
"learning_rate": 0.00014609053497942386,
"loss": 1.0355,
"step": 355
},
{
"epoch": 0.0741962077493817,
"grad_norm": 0.31354711887470305,
"learning_rate": 0.00014814814814814815,
"loss": 1.3388,
"step": 360
},
{
"epoch": 0.07522671063478978,
"grad_norm": 0.26048475689786116,
"learning_rate": 0.00015020576131687243,
"loss": 1.1111,
"step": 365
},
{
"epoch": 0.07625721352019786,
"grad_norm": 0.3339375055235073,
"learning_rate": 0.00015226337448559672,
"loss": 1.1734,
"step": 370
},
{
"epoch": 0.07728771640560593,
"grad_norm": 0.327943444067688,
"learning_rate": 0.00015432098765432098,
"loss": 1.3208,
"step": 375
},
{
"epoch": 0.07831821929101401,
"grad_norm": 0.2830129836407308,
"learning_rate": 0.00015637860082304527,
"loss": 1.0787,
"step": 380
},
{
"epoch": 0.0793487221764221,
"grad_norm": 0.2751640321887791,
"learning_rate": 0.00015843621399176956,
"loss": 1.3772,
"step": 385
},
{
"epoch": 0.08037922506183018,
"grad_norm": 0.27545532948015855,
"learning_rate": 0.00016049382716049385,
"loss": 1.1672,
"step": 390
},
{
"epoch": 0.08140972794723825,
"grad_norm": 0.300224560861019,
"learning_rate": 0.0001625514403292181,
"loss": 1.3224,
"step": 395
},
{
"epoch": 0.08244023083264633,
"grad_norm": 0.3227759504569782,
"learning_rate": 0.0001646090534979424,
"loss": 1.3495,
"step": 400
},
{
"epoch": 0.08244023083264633,
"eval_loss": 1.1826071739196777,
"eval_runtime": 2476.8142,
"eval_samples_per_second": 3.23,
"eval_steps_per_second": 0.202,
"step": 400
},
{
"epoch": 0.08347073371805441,
"grad_norm": 0.27853687771921737,
"learning_rate": 0.0001666666666666667,
"loss": 1.1022,
"step": 405
},
{
"epoch": 0.0845012366034625,
"grad_norm": 0.2685951286639512,
"learning_rate": 0.00016872427983539098,
"loss": 1.4061,
"step": 410
},
{
"epoch": 0.08553173948887056,
"grad_norm": 0.2608899320776935,
"learning_rate": 0.00017078189300411524,
"loss": 1.1174,
"step": 415
},
{
"epoch": 0.08656224237427865,
"grad_norm": 0.3397550472709015,
"learning_rate": 0.0001728395061728395,
"loss": 1.2198,
"step": 420
},
{
"epoch": 0.08759274525968673,
"grad_norm": 0.25033631170694937,
"learning_rate": 0.0001748971193415638,
"loss": 1.4029,
"step": 425
},
{
"epoch": 0.08862324814509481,
"grad_norm": 0.3049472388523693,
"learning_rate": 0.00017695473251028808,
"loss": 1.0265,
"step": 430
},
{
"epoch": 0.08965375103050288,
"grad_norm": 0.29598713943285077,
"learning_rate": 0.00017901234567901234,
"loss": 1.3719,
"step": 435
},
{
"epoch": 0.09068425391591096,
"grad_norm": 0.2656526738754943,
"learning_rate": 0.00018106995884773663,
"loss": 1.1238,
"step": 440
},
{
"epoch": 0.09171475680131905,
"grad_norm": 0.2546218728984369,
"learning_rate": 0.00018312757201646091,
"loss": 1.1853,
"step": 445
},
{
"epoch": 0.09274525968672712,
"grad_norm": 0.309485029099833,
"learning_rate": 0.0001851851851851852,
"loss": 1.3064,
"step": 450
},
{
"epoch": 0.0937757625721352,
"grad_norm": 0.28184488141448555,
"learning_rate": 0.00018724279835390946,
"loss": 1.0605,
"step": 455
},
{
"epoch": 0.09480626545754328,
"grad_norm": 0.2913075285039461,
"learning_rate": 0.00018930041152263375,
"loss": 1.4311,
"step": 460
},
{
"epoch": 0.09583676834295136,
"grad_norm": 0.7186836779179142,
"learning_rate": 0.00019135802469135804,
"loss": 1.1263,
"step": 465
},
{
"epoch": 0.09686727122835943,
"grad_norm": 0.29590695855715343,
"learning_rate": 0.00019341563786008233,
"loss": 1.2589,
"step": 470
},
{
"epoch": 0.09789777411376752,
"grad_norm": 0.31832697133586063,
"learning_rate": 0.0001954732510288066,
"loss": 1.395,
"step": 475
},
{
"epoch": 0.0989282769991756,
"grad_norm": 0.23627896192430478,
"learning_rate": 0.00019753086419753085,
"loss": 1.0703,
"step": 480
},
{
"epoch": 0.09995877988458368,
"grad_norm": 0.27354848483986705,
"learning_rate": 0.00019958847736625514,
"loss": 1.3205,
"step": 485
},
{
"epoch": 0.10098928276999175,
"grad_norm": 0.4895153568779187,
"learning_rate": 0.00019999958578867407,
"loss": 1.1488,
"step": 490
},
{
"epoch": 0.10201978565539983,
"grad_norm": 0.2811706808521417,
"learning_rate": 0.00019999790306104336,
"loss": 1.2442,
"step": 495
},
{
"epoch": 0.10305028854080792,
"grad_norm": 0.2572161259717821,
"learning_rate": 0.0001999949259506647,
"loss": 1.3355,
"step": 500
},
{
"epoch": 0.104080791426216,
"grad_norm": 0.41643641721819225,
"learning_rate": 0.00019999065449607402,
"loss": 1.0238,
"step": 505
},
{
"epoch": 0.10511129431162407,
"grad_norm": 0.2871689949893079,
"learning_rate": 0.00019998508875256158,
"loss": 1.3608,
"step": 510
},
{
"epoch": 0.10614179719703215,
"grad_norm": 0.3219574868373838,
"learning_rate": 0.0001999782287921708,
"loss": 1.0774,
"step": 515
},
{
"epoch": 0.10717230008244023,
"grad_norm": 0.23385146626580391,
"learning_rate": 0.00019997007470369773,
"loss": 1.184,
"step": 520
},
{
"epoch": 0.10820280296784832,
"grad_norm": 0.3585994112684759,
"learning_rate": 0.0001999606265926897,
"loss": 1.3426,
"step": 525
},
{
"epoch": 0.10923330585325638,
"grad_norm": 0.2894131584925935,
"learning_rate": 0.000199949884581444,
"loss": 1.099,
"step": 530
},
{
"epoch": 0.11026380873866447,
"grad_norm": 0.2824735567473504,
"learning_rate": 0.00019993784880900623,
"loss": 1.3997,
"step": 535
},
{
"epoch": 0.11129431162407255,
"grad_norm": 0.23322094258169807,
"learning_rate": 0.0001999245194311687,
"loss": 1.1019,
"step": 540
},
{
"epoch": 0.11232481450948063,
"grad_norm": 0.30614162605617107,
"learning_rate": 0.00019990989662046818,
"loss": 1.2576,
"step": 545
},
{
"epoch": 0.1133553173948887,
"grad_norm": 0.3389548425122287,
"learning_rate": 0.0001998939805661837,
"loss": 1.301,
"step": 550
},
{
"epoch": 0.11438582028029678,
"grad_norm": 0.25179179489194975,
"learning_rate": 0.00019987677147433432,
"loss": 0.9871,
"step": 555
},
{
"epoch": 0.11541632316570487,
"grad_norm": 0.24522921674032325,
"learning_rate": 0.0001998582695676762,
"loss": 1.4205,
"step": 560
},
{
"epoch": 0.11644682605111294,
"grad_norm": 0.2528881733347297,
"learning_rate": 0.00019983847508569987,
"loss": 1.1283,
"step": 565
},
{
"epoch": 0.11747732893652102,
"grad_norm": 0.28075131598249065,
"learning_rate": 0.00019981738828462703,
"loss": 1.2467,
"step": 570
},
{
"epoch": 0.1185078318219291,
"grad_norm": 0.3223992578484773,
"learning_rate": 0.00019979500943740735,
"loss": 1.3673,
"step": 575
},
{
"epoch": 0.11953833470733718,
"grad_norm": 0.2804260590851507,
"learning_rate": 0.00019977133883371478,
"loss": 1.1133,
"step": 580
},
{
"epoch": 0.12056883759274525,
"grad_norm": 0.25183674337041767,
"learning_rate": 0.00019974637677994404,
"loss": 1.3493,
"step": 585
},
{
"epoch": 0.12159934047815334,
"grad_norm": 0.23352143639236672,
"learning_rate": 0.00019972012359920638,
"loss": 1.1183,
"step": 590
},
{
"epoch": 0.12262984336356142,
"grad_norm": 0.2597060461259452,
"learning_rate": 0.0001996925796313256,
"loss": 1.2021,
"step": 595
},
{
"epoch": 0.1236603462489695,
"grad_norm": 0.31429525704141625,
"learning_rate": 0.00019966374523283347,
"loss": 1.3753,
"step": 600
},
{
"epoch": 0.1236603462489695,
"eval_loss": 1.1556562185287476,
"eval_runtime": 2465.6561,
"eval_samples_per_second": 3.245,
"eval_steps_per_second": 0.203,
"step": 600
},
{
"epoch": 0.12469084913437757,
"grad_norm": 0.23970501230526484,
"learning_rate": 0.00019963362077696537,
"loss": 0.9978,
"step": 605
},
{
"epoch": 0.12572135201978565,
"grad_norm": 0.24126751266896648,
"learning_rate": 0.00019960220665365518,
"loss": 1.3576,
"step": 610
},
{
"epoch": 0.12675185490519372,
"grad_norm": 0.29806682020411596,
"learning_rate": 0.0001995695032695305,
"loss": 1.0721,
"step": 615
},
{
"epoch": 0.12778235779060182,
"grad_norm": 0.3121199170110613,
"learning_rate": 0.0001995355110479071,
"loss": 1.2159,
"step": 620
},
{
"epoch": 0.1288128606760099,
"grad_norm": 0.3534465377015386,
"learning_rate": 0.00019950023042878366,
"loss": 1.3058,
"step": 625
},
{
"epoch": 0.12984336356141798,
"grad_norm": 0.23203831442914624,
"learning_rate": 0.00019946366186883604,
"loss": 0.9527,
"step": 630
},
{
"epoch": 0.13087386644682605,
"grad_norm": 0.2720965062058241,
"learning_rate": 0.00019942580584141127,
"loss": 1.3558,
"step": 635
},
{
"epoch": 0.13190436933223412,
"grad_norm": 0.2629592480398592,
"learning_rate": 0.0001993866628365215,
"loss": 1.1318,
"step": 640
},
{
"epoch": 0.13293487221764222,
"grad_norm": 0.2920226997136243,
"learning_rate": 0.00019934623336083772,
"loss": 1.122,
"step": 645
},
{
"epoch": 0.1339653751030503,
"grad_norm": 0.27168430341439026,
"learning_rate": 0.00019930451793768298,
"loss": 1.3241,
"step": 650
},
{
"epoch": 0.13499587798845836,
"grad_norm": 0.2845184130718667,
"learning_rate": 0.00019926151710702588,
"loss": 0.9734,
"step": 655
},
{
"epoch": 0.13602638087386645,
"grad_norm": 0.2888242420563682,
"learning_rate": 0.00019921723142547347,
"loss": 1.3656,
"step": 660
},
{
"epoch": 0.13705688375927452,
"grad_norm": 0.27299177987397777,
"learning_rate": 0.00019917166146626392,
"loss": 1.0691,
"step": 665
},
{
"epoch": 0.1380873866446826,
"grad_norm": 0.35236952647741737,
"learning_rate": 0.0001991248078192593,
"loss": 1.2263,
"step": 670
},
{
"epoch": 0.1391178895300907,
"grad_norm": 0.2681417629469303,
"learning_rate": 0.00019907667109093794,
"loss": 1.354,
"step": 675
},
{
"epoch": 0.14014839241549876,
"grad_norm": 0.310226199539697,
"learning_rate": 0.00019902725190438627,
"loss": 1.0365,
"step": 680
},
{
"epoch": 0.14117889530090685,
"grad_norm": 0.3036115548984946,
"learning_rate": 0.00019897655089929126,
"loss": 1.3279,
"step": 685
},
{
"epoch": 0.14220939818631492,
"grad_norm": 0.24182676057547745,
"learning_rate": 0.00019892456873193165,
"loss": 1.0803,
"step": 690
},
{
"epoch": 0.143239901071723,
"grad_norm": 0.27410693799977515,
"learning_rate": 0.00019887130607516978,
"loss": 1.2236,
"step": 695
},
{
"epoch": 0.1442704039571311,
"grad_norm": 0.3158493448762225,
"learning_rate": 0.00019881676361844275,
"loss": 1.3796,
"step": 700
},
{
"epoch": 0.14530090684253916,
"grad_norm": 0.48562045991681607,
"learning_rate": 0.0001987609420677535,
"loss": 1.0014,
"step": 705
},
{
"epoch": 0.14633140972794723,
"grad_norm": 0.24730942495184344,
"learning_rate": 0.00019870384214566174,
"loss": 1.3463,
"step": 710
},
{
"epoch": 0.14736191261335532,
"grad_norm": 0.3153226513268582,
"learning_rate": 0.00019864546459127448,
"loss": 1.0977,
"step": 715
},
{
"epoch": 0.1483924154987634,
"grad_norm": 0.249510122134111,
"learning_rate": 0.0001985858101602366,
"loss": 1.2687,
"step": 720
},
{
"epoch": 0.1494229183841715,
"grad_norm": 0.30842159531538627,
"learning_rate": 0.0001985248796247209,
"loss": 1.2815,
"step": 725
},
{
"epoch": 0.15045342126957956,
"grad_norm": 0.26379223642130784,
"learning_rate": 0.00019846267377341827,
"loss": 0.9403,
"step": 730
},
{
"epoch": 0.15148392415498763,
"grad_norm": 0.23891641481355658,
"learning_rate": 0.00019839919341152742,
"loss": 1.3505,
"step": 735
},
{
"epoch": 0.15251442704039572,
"grad_norm": 0.3238473098861904,
"learning_rate": 0.00019833443936074442,
"loss": 1.0409,
"step": 740
},
{
"epoch": 0.1535449299258038,
"grad_norm": 0.2746647114109431,
"learning_rate": 0.00019826841245925212,
"loss": 1.0984,
"step": 745
},
{
"epoch": 0.15457543281121186,
"grad_norm": 0.37674302298987083,
"learning_rate": 0.00019820111356170923,
"loss": 1.3147,
"step": 750
},
{
"epoch": 0.15560593569661996,
"grad_norm": 0.25219340224453657,
"learning_rate": 0.00019813254353923937,
"loss": 1.0348,
"step": 755
},
{
"epoch": 0.15663643858202803,
"grad_norm": 0.26735510810071406,
"learning_rate": 0.00019806270327941971,
"loss": 1.3468,
"step": 760
},
{
"epoch": 0.15766694146743612,
"grad_norm": 0.25452976189367954,
"learning_rate": 0.00019799159368626945,
"loss": 1.0699,
"step": 765
},
{
"epoch": 0.1586974443528442,
"grad_norm": 0.2863967455600455,
"learning_rate": 0.00019791921568023822,
"loss": 1.1045,
"step": 770
},
{
"epoch": 0.15972794723825226,
"grad_norm": 0.32564915808035255,
"learning_rate": 0.00019784557019819404,
"loss": 1.301,
"step": 775
},
{
"epoch": 0.16075845012366036,
"grad_norm": 0.2598862519671672,
"learning_rate": 0.00019777065819341137,
"loss": 0.9559,
"step": 780
},
{
"epoch": 0.16178895300906843,
"grad_norm": 0.2694147600296894,
"learning_rate": 0.00019769448063555856,
"loss": 1.346,
"step": 785
},
{
"epoch": 0.1628194558944765,
"grad_norm": 0.2464767798391986,
"learning_rate": 0.00019761703851068553,
"loss": 1.102,
"step": 790
},
{
"epoch": 0.1638499587798846,
"grad_norm": 0.2753953016773011,
"learning_rate": 0.0001975383328212107,
"loss": 1.2188,
"step": 795
},
{
"epoch": 0.16488046166529266,
"grad_norm": 0.38049304343681456,
"learning_rate": 0.00019745836458590836,
"loss": 1.3454,
"step": 800
},
{
"epoch": 0.16488046166529266,
"eval_loss": 1.1297026872634888,
"eval_runtime": 2464.1297,
"eval_samples_per_second": 3.247,
"eval_steps_per_second": 0.203,
"step": 800
},
{
"epoch": 0.16591096455070073,
"grad_norm": 0.25205784909199425,
"learning_rate": 0.0001973771348398953,
"loss": 0.9905,
"step": 805
},
{
"epoch": 0.16694146743610883,
"grad_norm": 0.2719665322536544,
"learning_rate": 0.0001972946446346173,
"loss": 1.3772,
"step": 810
},
{
"epoch": 0.1679719703215169,
"grad_norm": 0.23804058554763297,
"learning_rate": 0.00019721089503783577,
"loss": 1.0834,
"step": 815
},
{
"epoch": 0.169002473206925,
"grad_norm": 0.28963422898407165,
"learning_rate": 0.00019712588713361378,
"loss": 1.2064,
"step": 820
},
{
"epoch": 0.17003297609233306,
"grad_norm": 0.3442324930908095,
"learning_rate": 0.00019703962202230203,
"loss": 1.2967,
"step": 825
},
{
"epoch": 0.17106347897774113,
"grad_norm": 0.2547770116672794,
"learning_rate": 0.00019695210082052472,
"loss": 0.949,
"step": 830
},
{
"epoch": 0.17209398186314923,
"grad_norm": 0.2832243503202021,
"learning_rate": 0.00019686332466116487,
"loss": 1.2852,
"step": 835
},
{
"epoch": 0.1731244847485573,
"grad_norm": 0.2332582520614932,
"learning_rate": 0.0001967732946933499,
"loss": 0.979,
"step": 840
},
{
"epoch": 0.17415498763396536,
"grad_norm": 0.29363618940675,
"learning_rate": 0.00019668201208243658,
"loss": 1.223,
"step": 845
},
{
"epoch": 0.17518549051937346,
"grad_norm": 0.28680318445015207,
"learning_rate": 0.0001965894780099961,
"loss": 1.3119,
"step": 850
},
{
"epoch": 0.17621599340478153,
"grad_norm": 0.2415531891983295,
"learning_rate": 0.00019649569367379867,
"loss": 0.9811,
"step": 855
},
{
"epoch": 0.17724649629018963,
"grad_norm": 0.2766629334909396,
"learning_rate": 0.00019640066028779794,
"loss": 1.2464,
"step": 860
},
{
"epoch": 0.1782769991755977,
"grad_norm": 0.24460494296561755,
"learning_rate": 0.00019630437908211548,
"loss": 1.0896,
"step": 865
},
{
"epoch": 0.17930750206100576,
"grad_norm": 0.28580998835949145,
"learning_rate": 0.00019620685130302478,
"loss": 1.1636,
"step": 870
},
{
"epoch": 0.18033800494641386,
"grad_norm": 0.29222255776500833,
"learning_rate": 0.00019610807821293503,
"loss": 1.3137,
"step": 875
},
{
"epoch": 0.18136850783182193,
"grad_norm": 0.2684207256549492,
"learning_rate": 0.00019600806109037485,
"loss": 0.9964,
"step": 880
},
{
"epoch": 0.18239901071723,
"grad_norm": 0.2559298498236225,
"learning_rate": 0.00019590680122997582,
"loss": 1.3378,
"step": 885
},
{
"epoch": 0.1834295136026381,
"grad_norm": 0.2349865929838881,
"learning_rate": 0.00019580429994245555,
"loss": 1.0436,
"step": 890
},
{
"epoch": 0.18446001648804616,
"grad_norm": 0.29736696922743316,
"learning_rate": 0.0001957005585546009,
"loss": 1.1778,
"step": 895
},
{
"epoch": 0.18549051937345423,
"grad_norm": 0.34641352119598734,
"learning_rate": 0.00019559557840925055,
"loss": 1.289,
"step": 900
},
{
"epoch": 0.18652102225886233,
"grad_norm": 0.2589416225643458,
"learning_rate": 0.00019548936086527798,
"loss": 0.9491,
"step": 905
},
{
"epoch": 0.1875515251442704,
"grad_norm": 0.2778855204059982,
"learning_rate": 0.00019538190729757356,
"loss": 1.3701,
"step": 910
},
{
"epoch": 0.1885820280296785,
"grad_norm": 0.2433946170800378,
"learning_rate": 0.00019527321909702688,
"loss": 1.0454,
"step": 915
},
{
"epoch": 0.18961253091508656,
"grad_norm": 0.28491470830869087,
"learning_rate": 0.00019516329767050878,
"loss": 1.2001,
"step": 920
},
{
"epoch": 0.19064303380049463,
"grad_norm": 0.30637698470530034,
"learning_rate": 0.00019505214444085308,
"loss": 1.3096,
"step": 925
},
{
"epoch": 0.19167353668590273,
"grad_norm": 0.24736218088176834,
"learning_rate": 0.00019493976084683813,
"loss": 0.9663,
"step": 930
},
{
"epoch": 0.1927040395713108,
"grad_norm": 0.24983643400010644,
"learning_rate": 0.00019482614834316836,
"loss": 1.3698,
"step": 935
},
{
"epoch": 0.19373454245671887,
"grad_norm": 0.2420833122771942,
"learning_rate": 0.00019471130840045518,
"loss": 1.0335,
"step": 940
},
{
"epoch": 0.19476504534212696,
"grad_norm": 0.2573293132603339,
"learning_rate": 0.00019459524250519826,
"loss": 1.1905,
"step": 945
},
{
"epoch": 0.19579554822753503,
"grad_norm": 0.3710740261808263,
"learning_rate": 0.00019447795215976594,
"loss": 1.3102,
"step": 950
},
{
"epoch": 0.19682605111294313,
"grad_norm": 0.2569919144815,
"learning_rate": 0.0001943594388823761,
"loss": 0.8868,
"step": 955
},
{
"epoch": 0.1978565539983512,
"grad_norm": 0.3020196940157676,
"learning_rate": 0.00019423970420707627,
"loss": 1.3362,
"step": 960
},
{
"epoch": 0.19888705688375927,
"grad_norm": 0.23638629002825132,
"learning_rate": 0.00019411874968372402,
"loss": 1.0545,
"step": 965
},
{
"epoch": 0.19991755976916736,
"grad_norm": 0.2690903988670187,
"learning_rate": 0.00019399657687796658,
"loss": 1.1581,
"step": 970
},
{
"epoch": 0.20094806265457543,
"grad_norm": 0.3219945588976219,
"learning_rate": 0.00019387318737122092,
"loss": 1.2816,
"step": 975
},
{
"epoch": 0.2019785655399835,
"grad_norm": 0.2510582302230213,
"learning_rate": 0.0001937485827606529,
"loss": 0.9363,
"step": 980
},
{
"epoch": 0.2030090684253916,
"grad_norm": 0.3075547394305651,
"learning_rate": 0.00019362276465915702,
"loss": 1.2897,
"step": 985
},
{
"epoch": 0.20403957131079967,
"grad_norm": 0.24639351565359302,
"learning_rate": 0.0001934957346953352,
"loss": 1.0314,
"step": 990
},
{
"epoch": 0.20507007419620774,
"grad_norm": 0.31059587986354126,
"learning_rate": 0.00019336749451347586,
"loss": 1.2041,
"step": 995
},
{
"epoch": 0.20610057708161583,
"grad_norm": 0.2842021215906897,
"learning_rate": 0.0001932380457735326,
"loss": 1.2731,
"step": 1000
},
{
"epoch": 0.20610057708161583,
"eval_loss": 1.1071134805679321,
"eval_runtime": 2476.6697,
"eval_samples_per_second": 3.23,
"eval_steps_per_second": 0.202,
"step": 1000
},
{
"epoch": 0.2071310799670239,
"grad_norm": 0.25497821921270564,
"learning_rate": 0.00019310739015110267,
"loss": 0.9285,
"step": 1005
},
{
"epoch": 0.208161582852432,
"grad_norm": 0.2793650388701531,
"learning_rate": 0.00019297552933740547,
"loss": 1.3337,
"step": 1010
},
{
"epoch": 0.20919208573784007,
"grad_norm": 0.2516697285323032,
"learning_rate": 0.0001928424650392603,
"loss": 1.0655,
"step": 1015
},
{
"epoch": 0.21022258862324814,
"grad_norm": 0.30212273915620885,
"learning_rate": 0.00019270819897906468,
"loss": 1.2153,
"step": 1020
},
{
"epoch": 0.21125309150865623,
"grad_norm": 0.33419010602150057,
"learning_rate": 0.00019257273289477174,
"loss": 1.3252,
"step": 1025
},
{
"epoch": 0.2122835943940643,
"grad_norm": 0.27421309631436125,
"learning_rate": 0.00019243606853986786,
"loss": 1.0082,
"step": 1030
},
{
"epoch": 0.21331409727947237,
"grad_norm": 0.34717245381281975,
"learning_rate": 0.00019229820768335,
"loss": 1.3901,
"step": 1035
},
{
"epoch": 0.21434460016488047,
"grad_norm": 0.23427202649137227,
"learning_rate": 0.00019215915210970267,
"loss": 0.9838,
"step": 1040
},
{
"epoch": 0.21537510305028854,
"grad_norm": 0.2923332885069089,
"learning_rate": 0.00019201890361887506,
"loss": 1.1746,
"step": 1045
},
{
"epoch": 0.21640560593569663,
"grad_norm": 0.34553332711604556,
"learning_rate": 0.0001918774640262574,
"loss": 1.2803,
"step": 1050
},
{
"epoch": 0.2174361088211047,
"grad_norm": 0.282937796924011,
"learning_rate": 0.00019173483516265788,
"loss": 1.0297,
"step": 1055
},
{
"epoch": 0.21846661170651277,
"grad_norm": 0.3064500972204145,
"learning_rate": 0.00019159101887427854,
"loss": 1.3082,
"step": 1060
},
{
"epoch": 0.21949711459192087,
"grad_norm": 0.24140887151281765,
"learning_rate": 0.00019144601702269162,
"loss": 1.0166,
"step": 1065
},
{
"epoch": 0.22052761747732894,
"grad_norm": 0.2751668782987028,
"learning_rate": 0.00019129983148481552,
"loss": 1.149,
"step": 1070
},
{
"epoch": 0.221558120362737,
"grad_norm": 0.32570210618301154,
"learning_rate": 0.0001911524641528902,
"loss": 1.287,
"step": 1075
},
{
"epoch": 0.2225886232481451,
"grad_norm": 0.24698689309459554,
"learning_rate": 0.00019100391693445306,
"loss": 0.9291,
"step": 1080
},
{
"epoch": 0.22361912613355317,
"grad_norm": 0.26397978494526464,
"learning_rate": 0.00019085419175231394,
"loss": 1.2573,
"step": 1085
},
{
"epoch": 0.22464962901896127,
"grad_norm": 0.23809647134840975,
"learning_rate": 0.00019070329054453046,
"loss": 1.0226,
"step": 1090
},
{
"epoch": 0.22568013190436934,
"grad_norm": 0.31058086961600134,
"learning_rate": 0.00019055121526438272,
"loss": 1.1561,
"step": 1095
},
{
"epoch": 0.2267106347897774,
"grad_norm": 0.2778718312695914,
"learning_rate": 0.00019039796788034822,
"loss": 1.3034,
"step": 1100
},
{
"epoch": 0.2277411376751855,
"grad_norm": 0.2417800787196308,
"learning_rate": 0.00019024355037607622,
"loss": 0.8994,
"step": 1105
},
{
"epoch": 0.22877164056059357,
"grad_norm": 0.28904453048763723,
"learning_rate": 0.00019008796475036222,
"loss": 1.2439,
"step": 1110
},
{
"epoch": 0.22980214344600164,
"grad_norm": 0.23624741886843423,
"learning_rate": 0.00018993121301712193,
"loss": 1.0735,
"step": 1115
},
{
"epoch": 0.23083264633140974,
"grad_norm": 0.27985038336320134,
"learning_rate": 0.00018977329720536529,
"loss": 1.1098,
"step": 1120
},
{
"epoch": 0.2318631492168178,
"grad_norm": 0.2891758603238283,
"learning_rate": 0.00018961421935917016,
"loss": 1.2739,
"step": 1125
},
{
"epoch": 0.23289365210222587,
"grad_norm": 0.2635168830732569,
"learning_rate": 0.00018945398153765597,
"loss": 1.0068,
"step": 1130
},
{
"epoch": 0.23392415498763397,
"grad_norm": 0.2773479010066933,
"learning_rate": 0.00018929258581495685,
"loss": 1.3011,
"step": 1135
},
{
"epoch": 0.23495465787304204,
"grad_norm": 0.2560481495589164,
"learning_rate": 0.00018913003428019506,
"loss": 1.0617,
"step": 1140
},
{
"epoch": 0.23598516075845014,
"grad_norm": 0.277847323563774,
"learning_rate": 0.00018896632903745374,
"loss": 1.0843,
"step": 1145
},
{
"epoch": 0.2370156636438582,
"grad_norm": 0.3618488575426716,
"learning_rate": 0.00018880147220574976,
"loss": 1.2909,
"step": 1150
},
{
"epoch": 0.23804616652926627,
"grad_norm": 0.2840442439659897,
"learning_rate": 0.00018863546591900622,
"loss": 0.9916,
"step": 1155
},
{
"epoch": 0.23907666941467437,
"grad_norm": 0.2904562870162801,
"learning_rate": 0.00018846831232602492,
"loss": 1.3014,
"step": 1160
},
{
"epoch": 0.24010717230008244,
"grad_norm": 0.24724715741223574,
"learning_rate": 0.00018830001359045845,
"loss": 0.9967,
"step": 1165
},
{
"epoch": 0.2411376751854905,
"grad_norm": 0.26593630753000524,
"learning_rate": 0.00018813057189078243,
"loss": 1.0626,
"step": 1170
},
{
"epoch": 0.2421681780708986,
"grad_norm": 0.3276447476001242,
"learning_rate": 0.00018795998942026685,
"loss": 1.2979,
"step": 1175
},
{
"epoch": 0.24319868095630667,
"grad_norm": 0.26299318912598185,
"learning_rate": 0.00018778826838694812,
"loss": 0.9544,
"step": 1180
},
{
"epoch": 0.24422918384171477,
"grad_norm": 0.281736858661113,
"learning_rate": 0.0001876154110136003,
"loss": 1.3208,
"step": 1185
},
{
"epoch": 0.24525968672712284,
"grad_norm": 0.2855862979125554,
"learning_rate": 0.0001874414195377063,
"loss": 1.073,
"step": 1190
},
{
"epoch": 0.2462901896125309,
"grad_norm": 0.29014069310260776,
"learning_rate": 0.000187266296211429,
"loss": 1.1681,
"step": 1195
},
{
"epoch": 0.247320692497939,
"grad_norm": 0.3159103313554715,
"learning_rate": 0.0001870900433015821,
"loss": 1.3863,
"step": 1200
},
{
"epoch": 0.247320692497939,
"eval_loss": 1.0878444910049438,
"eval_runtime": 2463.5391,
"eval_samples_per_second": 3.247,
"eval_steps_per_second": 0.203,
"step": 1200
},
{
"epoch": 0.24835119538334707,
"grad_norm": 0.2980434519288696,
"learning_rate": 0.00018691266308960066,
"loss": 1.0099,
"step": 1205
},
{
"epoch": 0.24938169826875514,
"grad_norm": 0.2786508426961495,
"learning_rate": 0.00018673415787151166,
"loss": 1.3381,
"step": 1210
},
{
"epoch": 0.25041220115416324,
"grad_norm": 0.2676203832313212,
"learning_rate": 0.00018655452995790435,
"loss": 1.0199,
"step": 1215
},
{
"epoch": 0.2514427040395713,
"grad_norm": 0.28577827654986465,
"learning_rate": 0.00018637378167390018,
"loss": 1.2001,
"step": 1220
},
{
"epoch": 0.2524732069249794,
"grad_norm": 0.3398905155731309,
"learning_rate": 0.0001861919153591228,
"loss": 1.2629,
"step": 1225
},
{
"epoch": 0.25350370981038745,
"grad_norm": 0.26851195221770446,
"learning_rate": 0.00018600893336766786,
"loss": 0.9409,
"step": 1230
},
{
"epoch": 0.25453421269579557,
"grad_norm": 0.26580580512721763,
"learning_rate": 0.00018582483806807228,
"loss": 1.3071,
"step": 1235
},
{
"epoch": 0.25556471558120364,
"grad_norm": 0.2571726332870005,
"learning_rate": 0.0001856396318432838,
"loss": 1.0674,
"step": 1240
},
{
"epoch": 0.2565952184666117,
"grad_norm": 0.28205467317170896,
"learning_rate": 0.0001854533170906302,
"loss": 1.1248,
"step": 1245
},
{
"epoch": 0.2576257213520198,
"grad_norm": 0.3312983996287865,
"learning_rate": 0.00018526589622178802,
"loss": 1.2796,
"step": 1250
},
{
"epoch": 0.25865622423742785,
"grad_norm": 0.2757071364626265,
"learning_rate": 0.00018507737166275154,
"loss": 0.9658,
"step": 1255
},
{
"epoch": 0.25968672712283597,
"grad_norm": 0.3459316658489597,
"learning_rate": 0.00018488774585380125,
"loss": 1.3143,
"step": 1260
},
{
"epoch": 0.26071723000824404,
"grad_norm": 0.2521379464850173,
"learning_rate": 0.00018469702124947245,
"loss": 0.9855,
"step": 1265
},
{
"epoch": 0.2617477328936521,
"grad_norm": 0.28337200244179356,
"learning_rate": 0.00018450520031852325,
"loss": 1.2029,
"step": 1270
},
{
"epoch": 0.2627782357790602,
"grad_norm": 0.3562233737662029,
"learning_rate": 0.0001843122855439027,
"loss": 1.2867,
"step": 1275
},
{
"epoch": 0.26380873866446825,
"grad_norm": 0.2883121950988055,
"learning_rate": 0.00018411827942271884,
"loss": 0.978,
"step": 1280
},
{
"epoch": 0.2648392415498763,
"grad_norm": 0.3157197550829857,
"learning_rate": 0.000183923184466206,
"loss": 1.2767,
"step": 1285
},
{
"epoch": 0.26586974443528444,
"grad_norm": 0.24013448810718416,
"learning_rate": 0.0001837270031996926,
"loss": 1.0749,
"step": 1290
},
{
"epoch": 0.2669002473206925,
"grad_norm": 0.32239639052379476,
"learning_rate": 0.00018352973816256838,
"loss": 1.1699,
"step": 1295
},
{
"epoch": 0.2679307502061006,
"grad_norm": 0.29188168243683993,
"learning_rate": 0.0001833313919082515,
"loss": 1.3297,
"step": 1300
},
{
"epoch": 0.26896125309150865,
"grad_norm": 0.2710883490462726,
"learning_rate": 0.0001831319670041555,
"loss": 0.9246,
"step": 1305
},
{
"epoch": 0.2699917559769167,
"grad_norm": 0.2949025846368701,
"learning_rate": 0.00018293146603165603,
"loss": 1.339,
"step": 1310
},
{
"epoch": 0.27102225886232484,
"grad_norm": 0.24960673732959246,
"learning_rate": 0.00018272989158605752,
"loss": 1.0767,
"step": 1315
},
{
"epoch": 0.2720527617477329,
"grad_norm": 0.31392425706558585,
"learning_rate": 0.00018252724627655954,
"loss": 1.1564,
"step": 1320
},
{
"epoch": 0.273083264633141,
"grad_norm": 0.3755895664237009,
"learning_rate": 0.00018232353272622302,
"loss": 1.2677,
"step": 1325
},
{
"epoch": 0.27411376751854905,
"grad_norm": 0.2711084044553726,
"learning_rate": 0.00018211875357193632,
"loss": 0.9061,
"step": 1330
},
{
"epoch": 0.2751442704039571,
"grad_norm": 0.33175445691564986,
"learning_rate": 0.00018191291146438105,
"loss": 1.273,
"step": 1335
},
{
"epoch": 0.2761747732893652,
"grad_norm": 0.2537090735923007,
"learning_rate": 0.0001817060090679978,
"loss": 1.0512,
"step": 1340
},
{
"epoch": 0.2772052761747733,
"grad_norm": 0.36552128539895057,
"learning_rate": 0.00018149804906095163,
"loss": 1.2273,
"step": 1345
},
{
"epoch": 0.2782357790601814,
"grad_norm": 0.3485807735804365,
"learning_rate": 0.00018128903413509756,
"loss": 1.2532,
"step": 1350
},
{
"epoch": 0.27926628194558945,
"grad_norm": 0.2737156223424568,
"learning_rate": 0.0001810789669959453,
"loss": 1.0208,
"step": 1355
},
{
"epoch": 0.2802967848309975,
"grad_norm": 0.3091305751030036,
"learning_rate": 0.0001808678503626248,
"loss": 1.3024,
"step": 1360
},
{
"epoch": 0.2813272877164056,
"grad_norm": 0.2448701581737246,
"learning_rate": 0.00018065568696785058,
"loss": 1.0065,
"step": 1365
},
{
"epoch": 0.2823577906018137,
"grad_norm": 0.33520109656163827,
"learning_rate": 0.00018044247955788662,
"loss": 1.1409,
"step": 1370
},
{
"epoch": 0.2833882934872218,
"grad_norm": 0.36094563501069604,
"learning_rate": 0.00018022823089251073,
"loss": 1.2687,
"step": 1375
},
{
"epoch": 0.28441879637262985,
"grad_norm": 0.29785608686411563,
"learning_rate": 0.00018001294374497882,
"loss": 0.9357,
"step": 1380
},
{
"epoch": 0.2854492992580379,
"grad_norm": 0.3875323214985989,
"learning_rate": 0.00017979662090198906,
"loss": 1.2922,
"step": 1385
},
{
"epoch": 0.286479802143446,
"grad_norm": 0.3058769686259136,
"learning_rate": 0.00017957926516364565,
"loss": 1.0047,
"step": 1390
},
{
"epoch": 0.2875103050288541,
"grad_norm": 0.3030215125843954,
"learning_rate": 0.00017936087934342283,
"loss": 1.0829,
"step": 1395
},
{
"epoch": 0.2885408079142622,
"grad_norm": 0.32726489328282893,
"learning_rate": 0.00017914146626812823,
"loss": 1.2567,
"step": 1400
},
{
"epoch": 0.2885408079142622,
"eval_loss": 1.077731728553772,
"eval_runtime": 2463.5195,
"eval_samples_per_second": 3.247,
"eval_steps_per_second": 0.203,
"step": 1400
},
{
"epoch": 0.28957131079967025,
"grad_norm": 0.2992981015754632,
"learning_rate": 0.0001789210287778664,
"loss": 0.9792,
"step": 1405
},
{
"epoch": 0.2906018136850783,
"grad_norm": 0.3418337934197468,
"learning_rate": 0.00017869956972600202,
"loss": 1.3454,
"step": 1410
},
{
"epoch": 0.2916323165704864,
"grad_norm": 0.2526294763625045,
"learning_rate": 0.00017847709197912296,
"loss": 0.9715,
"step": 1415
},
{
"epoch": 0.29266281945589445,
"grad_norm": 0.3445050915146703,
"learning_rate": 0.0001782535984170032,
"loss": 1.113,
"step": 1420
},
{
"epoch": 0.2936933223413026,
"grad_norm": 0.32389102430563055,
"learning_rate": 0.00017802909193256547,
"loss": 1.2316,
"step": 1425
},
{
"epoch": 0.29472382522671065,
"grad_norm": 0.2613877323741928,
"learning_rate": 0.00017780357543184397,
"loss": 0.9269,
"step": 1430
},
{
"epoch": 0.2957543281121187,
"grad_norm": 0.2752040315178045,
"learning_rate": 0.00017757705183394653,
"loss": 1.2481,
"step": 1435
},
{
"epoch": 0.2967848309975268,
"grad_norm": 0.2535871144663709,
"learning_rate": 0.00017734952407101706,
"loss": 1.0301,
"step": 1440
},
{
"epoch": 0.29781533388293485,
"grad_norm": 0.28927983501411936,
"learning_rate": 0.0001771209950881974,
"loss": 1.1218,
"step": 1445
},
{
"epoch": 0.298845836768343,
"grad_norm": 0.37005769485403384,
"learning_rate": 0.00017689146784358927,
"loss": 1.2699,
"step": 1450
},
{
"epoch": 0.29987633965375105,
"grad_norm": 0.257690573166382,
"learning_rate": 0.0001766609453082161,
"loss": 0.9097,
"step": 1455
},
{
"epoch": 0.3009068425391591,
"grad_norm": 0.2708622421638633,
"learning_rate": 0.00017642943046598436,
"loss": 1.2685,
"step": 1460
},
{
"epoch": 0.3019373454245672,
"grad_norm": 0.2553661493662262,
"learning_rate": 0.00017619692631364506,
"loss": 1.062,
"step": 1465
},
{
"epoch": 0.30296784830997525,
"grad_norm": 0.3268803054784382,
"learning_rate": 0.00017596343586075497,
"loss": 1.1081,
"step": 1470
},
{
"epoch": 0.3039983511953833,
"grad_norm": 0.36412223215382217,
"learning_rate": 0.00017572896212963754,
"loss": 1.3221,
"step": 1475
},
{
"epoch": 0.30502885408079145,
"grad_norm": 0.2624897168675397,
"learning_rate": 0.000175493508155344,
"loss": 0.854,
"step": 1480
},
{
"epoch": 0.3060593569661995,
"grad_norm": 0.2910755462081272,
"learning_rate": 0.00017525707698561385,
"loss": 1.2961,
"step": 1485
},
{
"epoch": 0.3070898598516076,
"grad_norm": 0.26020570413888594,
"learning_rate": 0.00017501967168083557,
"loss": 1.0475,
"step": 1490
},
{
"epoch": 0.30812036273701565,
"grad_norm": 0.31355952905674644,
"learning_rate": 0.00017478129531400688,
"loss": 1.1383,
"step": 1495
},
{
"epoch": 0.3091508656224237,
"grad_norm": 0.3394529288865745,
"learning_rate": 0.00017454195097069505,
"loss": 1.3237,
"step": 1500
},
{
"epoch": 0.31018136850783184,
"grad_norm": 0.2623708218727371,
"learning_rate": 0.00017430164174899696,
"loss": 0.8949,
"step": 1505
},
{
"epoch": 0.3112118713932399,
"grad_norm": 0.3129719424005796,
"learning_rate": 0.0001740603707594989,
"loss": 1.2513,
"step": 1510
},
{
"epoch": 0.312242374278648,
"grad_norm": 0.2457430728630137,
"learning_rate": 0.00017381814112523648,
"loss": 1.0458,
"step": 1515
},
{
"epoch": 0.31327287716405605,
"grad_norm": 0.3235157301586437,
"learning_rate": 0.000173574955981654,
"loss": 1.1418,
"step": 1520
},
{
"epoch": 0.3143033800494641,
"grad_norm": 0.38068266969796094,
"learning_rate": 0.00017333081847656397,
"loss": 1.2703,
"step": 1525
},
{
"epoch": 0.31533388293487224,
"grad_norm": 0.2845050833721142,
"learning_rate": 0.00017308573177010652,
"loss": 0.8513,
"step": 1530
},
{
"epoch": 0.3163643858202803,
"grad_norm": 0.3011956997232069,
"learning_rate": 0.00017283969903470815,
"loss": 1.2671,
"step": 1535
},
{
"epoch": 0.3173948887056884,
"grad_norm": 0.2548678800376021,
"learning_rate": 0.0001725927234550409,
"loss": 0.9834,
"step": 1540
},
{
"epoch": 0.31842539159109645,
"grad_norm": 0.33070817321371104,
"learning_rate": 0.00017234480822798113,
"loss": 1.204,
"step": 1545
},
{
"epoch": 0.3194558944765045,
"grad_norm": 0.3739990601669961,
"learning_rate": 0.00017209595656256807,
"loss": 1.2874,
"step": 1550
},
{
"epoch": 0.3204863973619126,
"grad_norm": 0.28630568916559335,
"learning_rate": 0.0001718461716799623,
"loss": 0.9404,
"step": 1555
},
{
"epoch": 0.3215169002473207,
"grad_norm": 0.2984332733087229,
"learning_rate": 0.000171595456813404,
"loss": 1.3065,
"step": 1560
},
{
"epoch": 0.3225474031327288,
"grad_norm": 0.25297494310840185,
"learning_rate": 0.00017134381520817127,
"loss": 0.9533,
"step": 1565
},
{
"epoch": 0.32357790601813685,
"grad_norm": 0.3035622251707662,
"learning_rate": 0.00017109125012153783,
"loss": 1.1281,
"step": 1570
},
{
"epoch": 0.3246084089035449,
"grad_norm": 0.3401968875424911,
"learning_rate": 0.00017083776482273126,
"loss": 1.2439,
"step": 1575
},
{
"epoch": 0.325638911788953,
"grad_norm": 0.2742633328984466,
"learning_rate": 0.00017058336259289026,
"loss": 0.9067,
"step": 1580
},
{
"epoch": 0.3266694146743611,
"grad_norm": 0.33201726031181766,
"learning_rate": 0.0001703280467250225,
"loss": 1.2643,
"step": 1585
},
{
"epoch": 0.3276999175597692,
"grad_norm": 0.2740668506157443,
"learning_rate": 0.0001700718205239618,
"loss": 0.9757,
"step": 1590
},
{
"epoch": 0.32873042044517725,
"grad_norm": 0.28447062660977135,
"learning_rate": 0.0001698146873063255,
"loss": 1.0629,
"step": 1595
},
{
"epoch": 0.3297609233305853,
"grad_norm": 0.3855059276442332,
"learning_rate": 0.00016955665040047134,
"loss": 1.257,
"step": 1600
},
{
"epoch": 0.3297609233305853,
"eval_loss": 1.0629926919937134,
"eval_runtime": 2467.5248,
"eval_samples_per_second": 3.242,
"eval_steps_per_second": 0.203,
"step": 1600
},
{
"epoch": 0.3307914262159934,
"grad_norm": 0.30766662381135884,
"learning_rate": 0.00016929771314645454,
"loss": 0.9716,
"step": 1605
},
{
"epoch": 0.33182192910140146,
"grad_norm": 0.3070232969296392,
"learning_rate": 0.00016903787889598458,
"loss": 1.3027,
"step": 1610
},
{
"epoch": 0.3328524319868096,
"grad_norm": 0.29718017467704144,
"learning_rate": 0.00016877715101238172,
"loss": 1.053,
"step": 1615
},
{
"epoch": 0.33388293487221765,
"grad_norm": 0.29124927902233755,
"learning_rate": 0.00016851553287053342,
"loss": 1.1027,
"step": 1620
},
{
"epoch": 0.3349134377576257,
"grad_norm": 0.37490768596522844,
"learning_rate": 0.00016825302785685077,
"loss": 1.2816,
"step": 1625
},
{
"epoch": 0.3359439406430338,
"grad_norm": 0.2909222245308242,
"learning_rate": 0.00016798963936922467,
"loss": 0.9477,
"step": 1630
},
{
"epoch": 0.33697444352844186,
"grad_norm": 0.369759262778725,
"learning_rate": 0.00016772537081698175,
"loss": 1.2659,
"step": 1635
},
{
"epoch": 0.33800494641385,
"grad_norm": 0.25527189055893823,
"learning_rate": 0.00016746022562084026,
"loss": 1.0374,
"step": 1640
},
{
"epoch": 0.33903544929925805,
"grad_norm": 0.332617684071012,
"learning_rate": 0.0001671942072128659,
"loss": 1.1066,
"step": 1645
},
{
"epoch": 0.3400659521846661,
"grad_norm": 0.3184886656626706,
"learning_rate": 0.00016692731903642725,
"loss": 1.2649,
"step": 1650
},
{
"epoch": 0.3410964550700742,
"grad_norm": 0.2799293838329678,
"learning_rate": 0.0001666595645461512,
"loss": 0.9379,
"step": 1655
},
{
"epoch": 0.34212695795548226,
"grad_norm": 0.3517762307499943,
"learning_rate": 0.0001663909472078784,
"loss": 1.252,
"step": 1660
},
{
"epoch": 0.3431574608408903,
"grad_norm": 0.25384775837778967,
"learning_rate": 0.0001661214704986182,
"loss": 0.9847,
"step": 1665
},
{
"epoch": 0.34418796372629845,
"grad_norm": 0.29743299437921455,
"learning_rate": 0.00016585113790650388,
"loss": 1.1201,
"step": 1670
},
{
"epoch": 0.3452184666117065,
"grad_norm": 0.35589431095248486,
"learning_rate": 0.00016557995293074715,
"loss": 1.2206,
"step": 1675
},
{
"epoch": 0.3462489694971146,
"grad_norm": 0.2911495340019745,
"learning_rate": 0.00016530791908159323,
"loss": 0.9072,
"step": 1680
},
{
"epoch": 0.34727947238252266,
"grad_norm": 0.28789636914166633,
"learning_rate": 0.0001650350398802751,
"loss": 1.2946,
"step": 1685
},
{
"epoch": 0.3483099752679307,
"grad_norm": 0.26335155835600166,
"learning_rate": 0.0001647613188589682,
"loss": 1.0242,
"step": 1690
},
{
"epoch": 0.34934047815333885,
"grad_norm": 0.3325623900144522,
"learning_rate": 0.00016448675956074444,
"loss": 1.0931,
"step": 1695
},
{
"epoch": 0.3503709810387469,
"grad_norm": 0.3773837775653675,
"learning_rate": 0.0001642113655395266,
"loss": 1.2456,
"step": 1700
},
{
"epoch": 0.351401483924155,
"grad_norm": 0.285753347787117,
"learning_rate": 0.00016393514036004204,
"loss": 0.9662,
"step": 1705
},
{
"epoch": 0.35243198680956306,
"grad_norm": 0.3046613746172406,
"learning_rate": 0.0001636580875977769,
"loss": 1.3162,
"step": 1710
},
{
"epoch": 0.3534624896949711,
"grad_norm": 0.25099480129499735,
"learning_rate": 0.0001633802108389295,
"loss": 1.0061,
"step": 1715
},
{
"epoch": 0.35449299258037925,
"grad_norm": 0.34150662422413947,
"learning_rate": 0.00016310151368036408,
"loss": 1.1411,
"step": 1720
},
{
"epoch": 0.3555234954657873,
"grad_norm": 0.33392898782461256,
"learning_rate": 0.00016282199972956425,
"loss": 1.2154,
"step": 1725
},
{
"epoch": 0.3565539983511954,
"grad_norm": 0.3016273908980527,
"learning_rate": 0.00016254167260458622,
"loss": 0.9269,
"step": 1730
},
{
"epoch": 0.35758450123660346,
"grad_norm": 0.3297238572031139,
"learning_rate": 0.000162260535934012,
"loss": 1.2298,
"step": 1735
},
{
"epoch": 0.3586150041220115,
"grad_norm": 0.26311212102325543,
"learning_rate": 0.00016197859335690247,
"loss": 1.0416,
"step": 1740
},
{
"epoch": 0.3596455070074196,
"grad_norm": 0.35401650365663206,
"learning_rate": 0.0001616958485227503,
"loss": 1.0915,
"step": 1745
},
{
"epoch": 0.3606760098928277,
"grad_norm": 0.3946861144577044,
"learning_rate": 0.0001614123050914325,
"loss": 1.2156,
"step": 1750
},
{
"epoch": 0.3617065127782358,
"grad_norm": 0.31649222443853475,
"learning_rate": 0.00016112796673316336,
"loss": 0.9126,
"step": 1755
},
{
"epoch": 0.36273701566364386,
"grad_norm": 0.3591291538700441,
"learning_rate": 0.00016084283712844666,
"loss": 1.2438,
"step": 1760
},
{
"epoch": 0.3637675185490519,
"grad_norm": 0.2567906575863548,
"learning_rate": 0.00016055691996802823,
"loss": 0.9908,
"step": 1765
},
{
"epoch": 0.36479802143446,
"grad_norm": 0.34039693260231785,
"learning_rate": 0.00016027021895284808,
"loss": 1.0576,
"step": 1770
},
{
"epoch": 0.3658285243198681,
"grad_norm": 0.45929799156275763,
"learning_rate": 0.0001599827377939925,
"loss": 1.3099,
"step": 1775
},
{
"epoch": 0.3668590272052762,
"grad_norm": 0.2736394916456097,
"learning_rate": 0.00015969448021264606,
"loss": 0.8622,
"step": 1780
},
{
"epoch": 0.36788953009068426,
"grad_norm": 0.2884997182361836,
"learning_rate": 0.00015940544994004334,
"loss": 1.2272,
"step": 1785
},
{
"epoch": 0.3689200329760923,
"grad_norm": 0.2392312698637367,
"learning_rate": 0.00015911565071742088,
"loss": 0.9626,
"step": 1790
},
{
"epoch": 0.3699505358615004,
"grad_norm": 0.339213144942226,
"learning_rate": 0.00015882508629596836,
"loss": 1.1233,
"step": 1795
},
{
"epoch": 0.37098103874690846,
"grad_norm": 0.40881279626532063,
"learning_rate": 0.00015853376043678053,
"loss": 1.2129,
"step": 1800
},
{
"epoch": 0.37098103874690846,
"eval_loss": 1.051790475845337,
"eval_runtime": 2468.3063,
"eval_samples_per_second": 3.241,
"eval_steps_per_second": 0.203,
"step": 1800
},
{
"epoch": 0.3720115416323166,
"grad_norm": 0.27225360643392454,
"learning_rate": 0.00015824167691080802,
"loss": 0.8599,
"step": 1805
},
{
"epoch": 0.37304204451772466,
"grad_norm": 0.3655053849913508,
"learning_rate": 0.00015794883949880894,
"loss": 1.2174,
"step": 1810
},
{
"epoch": 0.3740725474031327,
"grad_norm": 0.2810810355550597,
"learning_rate": 0.00015765525199129966,
"loss": 1.0166,
"step": 1815
},
{
"epoch": 0.3751030502885408,
"grad_norm": 0.3332282628999955,
"learning_rate": 0.000157360918188506,
"loss": 1.1182,
"step": 1820
},
{
"epoch": 0.37613355317394886,
"grad_norm": 0.3175764657865911,
"learning_rate": 0.0001570658419003137,
"loss": 1.2687,
"step": 1825
},
{
"epoch": 0.377164056059357,
"grad_norm": 0.3120363543994925,
"learning_rate": 0.00015677002694621948,
"loss": 0.9812,
"step": 1830
},
{
"epoch": 0.37819455894476506,
"grad_norm": 0.327602667442098,
"learning_rate": 0.00015647347715528137,
"loss": 1.1919,
"step": 1835
},
{
"epoch": 0.3792250618301731,
"grad_norm": 0.2668526402245522,
"learning_rate": 0.00015617619636606924,
"loss": 0.9995,
"step": 1840
},
{
"epoch": 0.3802555647155812,
"grad_norm": 0.31419704503426554,
"learning_rate": 0.00015587818842661494,
"loss": 1.1765,
"step": 1845
},
{
"epoch": 0.38128606760098926,
"grad_norm": 0.4282942251054684,
"learning_rate": 0.00015557945719436278,
"loss": 1.2522,
"step": 1850
},
{
"epoch": 0.3823165704863974,
"grad_norm": 0.308069434736153,
"learning_rate": 0.00015528000653611935,
"loss": 0.8995,
"step": 1855
},
{
"epoch": 0.38334707337180546,
"grad_norm": 0.272826512737121,
"learning_rate": 0.0001549798403280036,
"loss": 1.2827,
"step": 1860
},
{
"epoch": 0.3843775762572135,
"grad_norm": 0.2498018349436216,
"learning_rate": 0.0001546789624553966,
"loss": 1.0308,
"step": 1865
},
{
"epoch": 0.3854080791426216,
"grad_norm": 0.32988195624784494,
"learning_rate": 0.00015437737681289128,
"loss": 1.1726,
"step": 1870
},
{
"epoch": 0.38643858202802966,
"grad_norm": 0.38213734379034847,
"learning_rate": 0.00015407508730424206,
"loss": 1.2046,
"step": 1875
},
{
"epoch": 0.38746908491343773,
"grad_norm": 0.2839035313263252,
"learning_rate": 0.00015377209784231424,
"loss": 0.9713,
"step": 1880
},
{
"epoch": 0.38849958779884586,
"grad_norm": 0.26153622554750505,
"learning_rate": 0.00015346841234903337,
"loss": 1.2771,
"step": 1885
},
{
"epoch": 0.3895300906842539,
"grad_norm": 0.25504870927422546,
"learning_rate": 0.0001531640347553345,
"loss": 1.0141,
"step": 1890
},
{
"epoch": 0.390560593569662,
"grad_norm": 0.33076548230805997,
"learning_rate": 0.00015285896900111133,
"loss": 1.1153,
"step": 1895
},
{
"epoch": 0.39159109645507006,
"grad_norm": 0.3682940168037353,
"learning_rate": 0.00015255321903516516,
"loss": 1.2387,
"step": 1900
},
{
"epoch": 0.39262159934047813,
"grad_norm": 0.3030559581752516,
"learning_rate": 0.00015224678881515377,
"loss": 0.8932,
"step": 1905
},
{
"epoch": 0.39365210222588626,
"grad_norm": 0.3016417378121846,
"learning_rate": 0.00015193968230754024,
"loss": 1.2566,
"step": 1910
},
{
"epoch": 0.3946826051112943,
"grad_norm": 0.2562267990440701,
"learning_rate": 0.00015163190348754162,
"loss": 1.0584,
"step": 1915
},
{
"epoch": 0.3957131079967024,
"grad_norm": 0.3152397653362513,
"learning_rate": 0.00015132345633907734,
"loss": 1.1119,
"step": 1920
},
{
"epoch": 0.39674361088211046,
"grad_norm": 0.3550077716096075,
"learning_rate": 0.0001510143448547178,
"loss": 1.2196,
"step": 1925
},
{
"epoch": 0.39777411376751853,
"grad_norm": 0.2713673784146358,
"learning_rate": 0.00015070457303563268,
"loss": 0.8809,
"step": 1930
},
{
"epoch": 0.3988046166529266,
"grad_norm": 0.32482913652095136,
"learning_rate": 0.000150394144891539,
"loss": 1.2114,
"step": 1935
},
{
"epoch": 0.3998351195383347,
"grad_norm": 0.2681392320249827,
"learning_rate": 0.00015008306444064942,
"loss": 1.0092,
"step": 1940
},
{
"epoch": 0.4008656224237428,
"grad_norm": 0.30813511160423507,
"learning_rate": 0.00014977133570961997,
"loss": 1.0587,
"step": 1945
},
{
"epoch": 0.40189612530915086,
"grad_norm": 0.3045685029840853,
"learning_rate": 0.00014945896273349827,
"loss": 1.3074,
"step": 1950
},
{
"epoch": 0.40292662819455893,
"grad_norm": 0.3178744984824384,
"learning_rate": 0.00014914594955567099,
"loss": 0.9477,
"step": 1955
},
{
"epoch": 0.403957131079967,
"grad_norm": 0.36931867482484526,
"learning_rate": 0.00014883230022781163,
"loss": 1.2359,
"step": 1960
},
{
"epoch": 0.4049876339653751,
"grad_norm": 0.27801383031198734,
"learning_rate": 0.00014851801880982814,
"loss": 1.0366,
"step": 1965
},
{
"epoch": 0.4060181368507832,
"grad_norm": 0.31213479911432085,
"learning_rate": 0.00014820310936981026,
"loss": 1.1624,
"step": 1970
},
{
"epoch": 0.40704863973619126,
"grad_norm": 0.33227538300256093,
"learning_rate": 0.000147887575983977,
"loss": 1.2724,
"step": 1975
},
{
"epoch": 0.40807914262159933,
"grad_norm": 0.298687515913839,
"learning_rate": 0.00014757142273662358,
"loss": 0.8436,
"step": 1980
},
{
"epoch": 0.4091096455070074,
"grad_norm": 0.36351887792750426,
"learning_rate": 0.00014725465372006905,
"loss": 1.2507,
"step": 1985
},
{
"epoch": 0.41014014839241547,
"grad_norm": 0.2702448610300903,
"learning_rate": 0.0001469372730346028,
"loss": 1.0191,
"step": 1990
},
{
"epoch": 0.4111706512778236,
"grad_norm": 0.3224079902083336,
"learning_rate": 0.00014661928478843186,
"loss": 1.1598,
"step": 1995
},
{
"epoch": 0.41220115416323166,
"grad_norm": 0.41798586070796323,
"learning_rate": 0.00014630069309762753,
"loss": 1.1939,
"step": 2000
},
{
"epoch": 0.41220115416323166,
"eval_loss": 1.0405410528182983,
"eval_runtime": 2474.052,
"eval_samples_per_second": 3.234,
"eval_steps_per_second": 0.202,
"step": 2000
},
{
"epoch": 0.41323165704863973,
"grad_norm": 0.28365317528526945,
"learning_rate": 0.00014598150208607212,
"loss": 0.8639,
"step": 2005
},
{
"epoch": 0.4142621599340478,
"grad_norm": 0.3214047753618349,
"learning_rate": 0.00014566171588540572,
"loss": 1.2183,
"step": 2010
},
{
"epoch": 0.41529266281945587,
"grad_norm": 0.2644278825870432,
"learning_rate": 0.00014534133863497252,
"loss": 0.9564,
"step": 2015
},
{
"epoch": 0.416323165704864,
"grad_norm": 0.32226077969266037,
"learning_rate": 0.00014502037448176734,
"loss": 1.0942,
"step": 2020
},
{
"epoch": 0.41735366859027206,
"grad_norm": 0.4101159870787422,
"learning_rate": 0.00014469882758038193,
"loss": 1.244,
"step": 2025
},
{
"epoch": 0.41838417147568013,
"grad_norm": 0.28153576445753414,
"learning_rate": 0.00014437670209295112,
"loss": 0.8947,
"step": 2030
},
{
"epoch": 0.4194146743610882,
"grad_norm": 0.3785396934306716,
"learning_rate": 0.0001440540021890992,
"loss": 1.2432,
"step": 2035
},
{
"epoch": 0.42044517724649627,
"grad_norm": 0.2542390029018319,
"learning_rate": 0.00014373073204588556,
"loss": 1.0083,
"step": 2040
},
{
"epoch": 0.4214756801319044,
"grad_norm": 0.3123729353287892,
"learning_rate": 0.0001434068958477509,
"loss": 1.1195,
"step": 2045
},
{
"epoch": 0.42250618301731246,
"grad_norm": 0.39530507310232316,
"learning_rate": 0.00014308249778646306,
"loss": 1.2036,
"step": 2050
},
{
"epoch": 0.42353668590272053,
"grad_norm": 0.2818105109815067,
"learning_rate": 0.0001427575420610626,
"loss": 0.8847,
"step": 2055
},
{
"epoch": 0.4245671887881286,
"grad_norm": 0.3219044657747011,
"learning_rate": 0.00014243203287780856,
"loss": 1.2623,
"step": 2060
},
{
"epoch": 0.42559769167353667,
"grad_norm": 0.2611111361591686,
"learning_rate": 0.00014210597445012398,
"loss": 1.0083,
"step": 2065
},
{
"epoch": 0.42662819455894474,
"grad_norm": 0.33485314347370027,
"learning_rate": 0.0001417793709985415,
"loss": 1.1823,
"step": 2070
},
{
"epoch": 0.42765869744435286,
"grad_norm": 0.34925710089096873,
"learning_rate": 0.0001414522267506484,
"loss": 1.2728,
"step": 2075
},
{
"epoch": 0.42868920032976093,
"grad_norm": 0.25546314251507235,
"learning_rate": 0.0001411245459410322,
"loss": 0.8703,
"step": 2080
},
{
"epoch": 0.429719703215169,
"grad_norm": 0.38324554510188286,
"learning_rate": 0.00014079633281122573,
"loss": 1.2551,
"step": 2085
},
{
"epoch": 0.43075020610057707,
"grad_norm": 0.27810663295608457,
"learning_rate": 0.00014046759160965224,
"loss": 1.0267,
"step": 2090
},
{
"epoch": 0.43178070898598514,
"grad_norm": 0.35335784100847833,
"learning_rate": 0.0001401383265915703,
"loss": 1.1345,
"step": 2095
},
{
"epoch": 0.43281121187139326,
"grad_norm": 0.4118966005270665,
"learning_rate": 0.00013980854201901886,
"loss": 1.1936,
"step": 2100
},
{
"epoch": 0.43384171475680133,
"grad_norm": 0.3062362602687425,
"learning_rate": 0.00013947824216076207,
"loss": 0.9152,
"step": 2105
},
{
"epoch": 0.4348722176422094,
"grad_norm": 0.29620498290186464,
"learning_rate": 0.00013914743129223405,
"loss": 1.2309,
"step": 2110
},
{
"epoch": 0.43590272052761747,
"grad_norm": 0.2710224032495773,
"learning_rate": 0.00013881611369548325,
"loss": 1.0408,
"step": 2115
},
{
"epoch": 0.43693322341302554,
"grad_norm": 0.3056464151796937,
"learning_rate": 0.00013848429365911753,
"loss": 1.1743,
"step": 2120
},
{
"epoch": 0.4379637262984336,
"grad_norm": 0.3843468598801491,
"learning_rate": 0.00013815197547824824,
"loss": 1.2395,
"step": 2125
},
{
"epoch": 0.43899422918384173,
"grad_norm": 0.2798396925326106,
"learning_rate": 0.00013781916345443474,
"loss": 0.8612,
"step": 2130
},
{
"epoch": 0.4400247320692498,
"grad_norm": 0.3279250283208323,
"learning_rate": 0.00013748586189562878,
"loss": 1.2354,
"step": 2135
},
{
"epoch": 0.44105523495465787,
"grad_norm": 0.2735131812328209,
"learning_rate": 0.00013715207511611876,
"loss": 1.0026,
"step": 2140
},
{
"epoch": 0.44208573784006594,
"grad_norm": 0.37219144396512693,
"learning_rate": 0.0001368178074364737,
"loss": 1.0704,
"step": 2145
},
{
"epoch": 0.443116240725474,
"grad_norm": 0.46408445168045037,
"learning_rate": 0.00013648306318348762,
"loss": 1.2511,
"step": 2150
},
{
"epoch": 0.44414674361088213,
"grad_norm": 0.2720164174270696,
"learning_rate": 0.0001361478466901231,
"loss": 0.8989,
"step": 2155
},
{
"epoch": 0.4451772464962902,
"grad_norm": 0.3747167130586965,
"learning_rate": 0.0001358121622954557,
"loss": 1.3666,
"step": 2160
},
{
"epoch": 0.44620774938169827,
"grad_norm": 0.2962954738156199,
"learning_rate": 0.00013547601434461733,
"loss": 0.9376,
"step": 2165
},
{
"epoch": 0.44723825226710634,
"grad_norm": 0.3347575447270046,
"learning_rate": 0.0001351394071887404,
"loss": 1.0754,
"step": 2170
},
{
"epoch": 0.4482687551525144,
"grad_norm": 0.37805918523823334,
"learning_rate": 0.0001348023451849012,
"loss": 1.2106,
"step": 2175
},
{
"epoch": 0.44929925803792253,
"grad_norm": 0.302277885917108,
"learning_rate": 0.00013446483269606362,
"loss": 0.8481,
"step": 2180
},
{
"epoch": 0.4503297609233306,
"grad_norm": 0.35890494559800096,
"learning_rate": 0.00013412687409102277,
"loss": 1.2101,
"step": 2185
},
{
"epoch": 0.45136026380873867,
"grad_norm": 0.27832175193098935,
"learning_rate": 0.00013378847374434814,
"loss": 0.9563,
"step": 2190
},
{
"epoch": 0.45239076669414674,
"grad_norm": 0.35752890078545335,
"learning_rate": 0.0001334496360363274,
"loss": 1.1166,
"step": 2195
},
{
"epoch": 0.4534212695795548,
"grad_norm": 0.3208813709791692,
"learning_rate": 0.00013311036535290925,
"loss": 1.2658,
"step": 2200
},
{
"epoch": 0.4534212695795548,
"eval_loss": 1.0313003063201904,
"eval_runtime": 2462.8678,
"eval_samples_per_second": 3.248,
"eval_steps_per_second": 0.203,
"step": 2200
},
{
"epoch": 0.4544517724649629,
"grad_norm": 0.3164662130947771,
"learning_rate": 0.00013277066608564713,
"loss": 0.9046,
"step": 2205
},
{
"epoch": 0.455482275350371,
"grad_norm": 0.32760397119015733,
"learning_rate": 0.0001324305426316418,
"loss": 1.2596,
"step": 2210
},
{
"epoch": 0.45651277823577907,
"grad_norm": 0.26879676205264913,
"learning_rate": 0.000132089999393485,
"loss": 0.9842,
"step": 2215
},
{
"epoch": 0.45754328112118714,
"grad_norm": 0.35211465947054194,
"learning_rate": 0.000131749040779202,
"loss": 1.1149,
"step": 2220
},
{
"epoch": 0.4585737840065952,
"grad_norm": 0.3640438187361373,
"learning_rate": 0.0001314076712021949,
"loss": 1.2521,
"step": 2225
},
{
"epoch": 0.4596042868920033,
"grad_norm": 0.28859583596302096,
"learning_rate": 0.0001310658950811852,
"loss": 0.8689,
"step": 2230
},
{
"epoch": 0.4606347897774114,
"grad_norm": 0.3245571156879321,
"learning_rate": 0.00013072371684015688,
"loss": 1.2013,
"step": 2235
},
{
"epoch": 0.46166529266281947,
"grad_norm": 0.2630264452723511,
"learning_rate": 0.00013038114090829892,
"loss": 1.0054,
"step": 2240
},
{
"epoch": 0.46269579554822754,
"grad_norm": 0.3629689887711568,
"learning_rate": 0.00013003817171994807,
"loss": 1.1201,
"step": 2245
},
{
"epoch": 0.4637262984336356,
"grad_norm": 0.38051210660623236,
"learning_rate": 0.00012969481371453135,
"loss": 1.2291,
"step": 2250
},
{
"epoch": 0.4647568013190437,
"grad_norm": 0.3038049492726604,
"learning_rate": 0.00012935107133650885,
"loss": 0.8923,
"step": 2255
},
{
"epoch": 0.46578730420445175,
"grad_norm": 0.34319934420858916,
"learning_rate": 0.00012900694903531586,
"loss": 1.2325,
"step": 2260
},
{
"epoch": 0.46681780708985987,
"grad_norm": 0.2729387501421985,
"learning_rate": 0.0001286624512653055,
"loss": 0.9893,
"step": 2265
},
{
"epoch": 0.46784830997526794,
"grad_norm": 0.37018830848703915,
"learning_rate": 0.00012831758248569097,
"loss": 1.1021,
"step": 2270
},
{
"epoch": 0.468878812860676,
"grad_norm": 0.36278206444854155,
"learning_rate": 0.00012797234716048784,
"loss": 1.2005,
"step": 2275
},
{
"epoch": 0.4699093157460841,
"grad_norm": 0.2976065955292209,
"learning_rate": 0.00012762674975845637,
"loss": 0.8405,
"step": 2280
},
{
"epoch": 0.47093981863149215,
"grad_norm": 0.3358390029204746,
"learning_rate": 0.00012728079475304345,
"loss": 1.2722,
"step": 2285
},
{
"epoch": 0.47197032151690027,
"grad_norm": 0.24689134910878455,
"learning_rate": 0.0001269344866223249,
"loss": 0.9856,
"step": 2290
},
{
"epoch": 0.47300082440230834,
"grad_norm": 0.3183219956731036,
"learning_rate": 0.00012658782984894743,
"loss": 1.0672,
"step": 2295
},
{
"epoch": 0.4740313272877164,
"grad_norm": 0.3220468985387059,
"learning_rate": 0.00012624082892007064,
"loss": 1.2301,
"step": 2300
},
{
"epoch": 0.4750618301731245,
"grad_norm": 0.3012629746942751,
"learning_rate": 0.00012589348832730882,
"loss": 0.9345,
"step": 2305
},
{
"epoch": 0.47609233305853255,
"grad_norm": 0.34208223291291867,
"learning_rate": 0.00012554581256667296,
"loss": 1.1783,
"step": 2310
},
{
"epoch": 0.47712283594394067,
"grad_norm": 0.2633458752929658,
"learning_rate": 0.00012519780613851254,
"loss": 1.0127,
"step": 2315
},
{
"epoch": 0.47815333882934874,
"grad_norm": 0.3245639160611015,
"learning_rate": 0.00012484947354745714,
"loss": 1.074,
"step": 2320
},
{
"epoch": 0.4791838417147568,
"grad_norm": 0.4155838481655266,
"learning_rate": 0.0001245008193023583,
"loss": 1.2446,
"step": 2325
},
{
"epoch": 0.4802143446001649,
"grad_norm": 0.29474111212781395,
"learning_rate": 0.00012415184791623101,
"loss": 0.8756,
"step": 2330
},
{
"epoch": 0.48124484748557295,
"grad_norm": 0.38251199632272437,
"learning_rate": 0.00012380256390619548,
"loss": 1.2367,
"step": 2335
},
{
"epoch": 0.482275350370981,
"grad_norm": 0.2876263709986664,
"learning_rate": 0.00012345297179341844,
"loss": 0.942,
"step": 2340
},
{
"epoch": 0.48330585325638914,
"grad_norm": 0.3470650266011227,
"learning_rate": 0.00012310307610305477,
"loss": 1.1312,
"step": 2345
},
{
"epoch": 0.4843363561417972,
"grad_norm": 0.3437397458970076,
"learning_rate": 0.00012275288136418889,
"loss": 1.2013,
"step": 2350
},
{
"epoch": 0.4853668590272053,
"grad_norm": 0.2891400010122533,
"learning_rate": 0.0001224023921097762,
"loss": 0.875,
"step": 2355
},
{
"epoch": 0.48639736191261335,
"grad_norm": 0.3343377419438735,
"learning_rate": 0.0001220516128765842,
"loss": 1.2263,
"step": 2360
},
{
"epoch": 0.4874278647980214,
"grad_norm": 0.2815731947053776,
"learning_rate": 0.00012170054820513401,
"loss": 1.0279,
"step": 2365
},
{
"epoch": 0.48845836768342954,
"grad_norm": 0.34918143666071294,
"learning_rate": 0.00012134920263964147,
"loss": 1.1073,
"step": 2370
},
{
"epoch": 0.4894888705688376,
"grad_norm": 0.3378508040938572,
"learning_rate": 0.00012099758072795832,
"loss": 1.2131,
"step": 2375
},
{
"epoch": 0.4905193734542457,
"grad_norm": 0.3404097695213611,
"learning_rate": 0.00012064568702151335,
"loss": 0.9207,
"step": 2380
},
{
"epoch": 0.49154987633965375,
"grad_norm": 0.31871974527595615,
"learning_rate": 0.0001202935260752535,
"loss": 1.2815,
"step": 2385
},
{
"epoch": 0.4925803792250618,
"grad_norm": 0.27586523947951513,
"learning_rate": 0.00011994110244758496,
"loss": 0.9798,
"step": 2390
},
{
"epoch": 0.4936108821104699,
"grad_norm": 0.29258038219303606,
"learning_rate": 0.00011958842070031395,
"loss": 1.0612,
"step": 2395
},
{
"epoch": 0.494641384995878,
"grad_norm": 0.37118694749476105,
"learning_rate": 0.0001192354853985879,
"loss": 1.1718,
"step": 2400
},
{
"epoch": 0.494641384995878,
"eval_loss": 1.0186035633087158,
"eval_runtime": 2474.4655,
"eval_samples_per_second": 3.233,
"eval_steps_per_second": 0.202,
"step": 2400
},
{
"epoch": 0.4956718878812861,
"grad_norm": 0.3041203248300929,
"learning_rate": 0.00011888230111083627,
"loss": 0.8718,
"step": 2405
},
{
"epoch": 0.49670239076669415,
"grad_norm": 0.34626490332091997,
"learning_rate": 0.00011852887240871145,
"loss": 1.2384,
"step": 2410
},
{
"epoch": 0.4977328936521022,
"grad_norm": 0.2627628128026553,
"learning_rate": 0.00011817520386702947,
"loss": 0.9474,
"step": 2415
},
{
"epoch": 0.4987633965375103,
"grad_norm": 0.3243643599593401,
"learning_rate": 0.00011782130006371092,
"loss": 1.0514,
"step": 2420
},
{
"epoch": 0.4997938994229184,
"grad_norm": 0.38727382866372845,
"learning_rate": 0.00011746716557972167,
"loss": 1.2191,
"step": 2425
},
{
"epoch": 0.5008244023083265,
"grad_norm": 0.29252070214667514,
"learning_rate": 0.00011711280499901347,
"loss": 0.8687,
"step": 2430
},
{
"epoch": 0.5018549051937345,
"grad_norm": 0.32103295732093995,
"learning_rate": 0.00011675822290846474,
"loss": 1.2169,
"step": 2435
},
{
"epoch": 0.5028854080791426,
"grad_norm": 0.2619181848445574,
"learning_rate": 0.00011640342389782114,
"loss": 0.9916,
"step": 2440
},
{
"epoch": 0.5039159109645507,
"grad_norm": 0.36227995308769084,
"learning_rate": 0.00011604841255963616,
"loss": 1.0797,
"step": 2445
},
{
"epoch": 0.5049464138499588,
"grad_norm": 0.4472357927883422,
"learning_rate": 0.00011569319348921168,
"loss": 1.2004,
"step": 2450
},
{
"epoch": 0.5059769167353668,
"grad_norm": 0.2990090832488795,
"learning_rate": 0.00011533777128453844,
"loss": 0.905,
"step": 2455
},
{
"epoch": 0.5070074196207749,
"grad_norm": 0.3421802694540464,
"learning_rate": 0.00011498215054623664,
"loss": 1.2502,
"step": 2460
},
{
"epoch": 0.5080379225061831,
"grad_norm": 0.2619877856697061,
"learning_rate": 0.00011462633587749629,
"loss": 0.9878,
"step": 2465
},
{
"epoch": 0.5090684253915911,
"grad_norm": 0.3573733070627249,
"learning_rate": 0.00011427033188401768,
"loss": 1.0799,
"step": 2470
},
{
"epoch": 0.5100989282769992,
"grad_norm": 0.36110594973295185,
"learning_rate": 0.00011391414317395167,
"loss": 1.1831,
"step": 2475
},
{
"epoch": 0.5111294311624073,
"grad_norm": 0.29726968151818156,
"learning_rate": 0.00011355777435784016,
"loss": 0.9268,
"step": 2480
},
{
"epoch": 0.5121599340478153,
"grad_norm": 0.38992649702696064,
"learning_rate": 0.0001132012300485564,
"loss": 1.2403,
"step": 2485
},
{
"epoch": 0.5131904369332234,
"grad_norm": 0.31217559249274757,
"learning_rate": 0.00011284451486124514,
"loss": 0.9683,
"step": 2490
},
{
"epoch": 0.5142209398186315,
"grad_norm": 0.3309516167120589,
"learning_rate": 0.00011248763341326307,
"loss": 1.018,
"step": 2495
},
{
"epoch": 0.5152514427040396,
"grad_norm": 0.3669152523683029,
"learning_rate": 0.00011213059032411897,
"loss": 1.2018,
"step": 2500
},
{
"epoch": 0.5162819455894476,
"grad_norm": 0.26794057215849676,
"learning_rate": 0.00011177339021541387,
"loss": 0.8112,
"step": 2505
},
{
"epoch": 0.5173124484748557,
"grad_norm": 0.35964442987172796,
"learning_rate": 0.00011141603771078133,
"loss": 1.1878,
"step": 2510
},
{
"epoch": 0.5183429513602638,
"grad_norm": 0.26756565994743753,
"learning_rate": 0.00011105853743582751,
"loss": 1.0244,
"step": 2515
},
{
"epoch": 0.5193734542456719,
"grad_norm": 0.37262291618379195,
"learning_rate": 0.00011070089401807129,
"loss": 1.0956,
"step": 2520
},
{
"epoch": 0.52040395713108,
"grad_norm": 0.3862693430821137,
"learning_rate": 0.0001103431120868845,
"loss": 1.1777,
"step": 2525
},
{
"epoch": 0.5214344600164881,
"grad_norm": 0.34456489334354296,
"learning_rate": 0.00010998519627343182,
"loss": 0.9219,
"step": 2530
},
{
"epoch": 0.5224649629018961,
"grad_norm": 0.3363644750152242,
"learning_rate": 0.00010962715121061095,
"loss": 1.2219,
"step": 2535
},
{
"epoch": 0.5234954657873042,
"grad_norm": 0.2773020803302085,
"learning_rate": 0.00010926898153299259,
"loss": 0.9829,
"step": 2540
},
{
"epoch": 0.5245259686727123,
"grad_norm": 0.3387440320396881,
"learning_rate": 0.00010891069187676051,
"loss": 1.1028,
"step": 2545
},
{
"epoch": 0.5255564715581204,
"grad_norm": 0.4105637833575777,
"learning_rate": 0.00010855228687965138,
"loss": 1.1622,
"step": 2550
},
{
"epoch": 0.5265869744435284,
"grad_norm": 0.3203362105742117,
"learning_rate": 0.000108193771180895,
"loss": 0.8631,
"step": 2555
},
{
"epoch": 0.5276174773289365,
"grad_norm": 0.3176938770218858,
"learning_rate": 0.00010783514942115398,
"loss": 1.1615,
"step": 2560
},
{
"epoch": 0.5286479802143446,
"grad_norm": 0.27188600955614695,
"learning_rate": 0.0001074764262424639,
"loss": 1.0099,
"step": 2565
},
{
"epoch": 0.5296784830997526,
"grad_norm": 0.3225534680387049,
"learning_rate": 0.00010711760628817304,
"loss": 1.1351,
"step": 2570
},
{
"epoch": 0.5307089859851608,
"grad_norm": 0.39558907715108127,
"learning_rate": 0.0001067586942028824,
"loss": 1.1706,
"step": 2575
},
{
"epoch": 0.5317394888705689,
"grad_norm": 0.28753930251222504,
"learning_rate": 0.00010639969463238553,
"loss": 0.8745,
"step": 2580
},
{
"epoch": 0.532769991755977,
"grad_norm": 0.30734397473616565,
"learning_rate": 0.00010604061222360828,
"loss": 1.1448,
"step": 2585
},
{
"epoch": 0.533800494641385,
"grad_norm": 0.26804645458820414,
"learning_rate": 0.00010568145162454896,
"loss": 0.9725,
"step": 2590
},
{
"epoch": 0.5348309975267931,
"grad_norm": 0.3263053082797193,
"learning_rate": 0.00010532221748421787,
"loss": 1.0632,
"step": 2595
},
{
"epoch": 0.5358615004122012,
"grad_norm": 0.3650347794639528,
"learning_rate": 0.00010496291445257725,
"loss": 1.1795,
"step": 2600
},
{
"epoch": 0.5358615004122012,
"eval_loss": 1.0102049112319946,
"eval_runtime": 2464.7805,
"eval_samples_per_second": 3.246,
"eval_steps_per_second": 0.203,
"step": 2600
},
{
"epoch": 0.5368920032976092,
"grad_norm": 0.299206263893403,
"learning_rate": 0.00010460354718048109,
"loss": 0.8798,
"step": 2605
},
{
"epoch": 0.5379225061830173,
"grad_norm": 0.32212757306557593,
"learning_rate": 0.00010424412031961484,
"loss": 1.2768,
"step": 2610
},
{
"epoch": 0.5389530090684254,
"grad_norm": 0.2686223265118172,
"learning_rate": 0.0001038846385224354,
"loss": 0.9964,
"step": 2615
},
{
"epoch": 0.5399835119538334,
"grad_norm": 0.31440179887584824,
"learning_rate": 0.00010352510644211074,
"loss": 1.0418,
"step": 2620
},
{
"epoch": 0.5410140148392415,
"grad_norm": 0.3262137336119927,
"learning_rate": 0.0001031655287324596,
"loss": 1.2262,
"step": 2625
},
{
"epoch": 0.5420445177246497,
"grad_norm": 0.2977115868800609,
"learning_rate": 0.00010280591004789144,
"loss": 0.867,
"step": 2630
},
{
"epoch": 0.5430750206100577,
"grad_norm": 0.3481967708447838,
"learning_rate": 0.00010244625504334609,
"loss": 1.1666,
"step": 2635
},
{
"epoch": 0.5441055234954658,
"grad_norm": 0.27201591179231566,
"learning_rate": 0.00010208656837423351,
"loss": 0.9024,
"step": 2640
},
{
"epoch": 0.5451360263808739,
"grad_norm": 0.3829985152487614,
"learning_rate": 0.00010172685469637351,
"loss": 1.0713,
"step": 2645
},
{
"epoch": 0.546166529266282,
"grad_norm": 0.36090749295144076,
"learning_rate": 0.00010136711866593551,
"loss": 1.1479,
"step": 2650
},
{
"epoch": 0.54719703215169,
"grad_norm": 0.300126161900349,
"learning_rate": 0.00010100736493937828,
"loss": 0.8899,
"step": 2655
},
{
"epoch": 0.5482275350370981,
"grad_norm": 0.33344984792579047,
"learning_rate": 0.00010064759817338965,
"loss": 1.2477,
"step": 2660
},
{
"epoch": 0.5492580379225062,
"grad_norm": 0.2615632901768006,
"learning_rate": 0.00010028782302482617,
"loss": 0.9453,
"step": 2665
},
{
"epoch": 0.5502885408079142,
"grad_norm": 0.36602450939184455,
"learning_rate": 9.992804415065305e-05,
"loss": 1.0671,
"step": 2670
},
{
"epoch": 0.5513190436933223,
"grad_norm": 0.34898414138240497,
"learning_rate": 9.956826620788352e-05,
"loss": 1.1969,
"step": 2675
},
{
"epoch": 0.5523495465787304,
"grad_norm": 0.28926741748856266,
"learning_rate": 9.920849385351889e-05,
"loss": 0.8264,
"step": 2680
},
{
"epoch": 0.5533800494641385,
"grad_norm": 0.3470665476375764,
"learning_rate": 9.884873174448811e-05,
"loss": 1.2393,
"step": 2685
},
{
"epoch": 0.5544105523495466,
"grad_norm": 0.25248382277496384,
"learning_rate": 9.848898453758752e-05,
"loss": 0.9752,
"step": 2690
},
{
"epoch": 0.5554410552349547,
"grad_norm": 0.322941880174542,
"learning_rate": 9.812925688942054e-05,
"loss": 1.0431,
"step": 2695
},
{
"epoch": 0.5564715581203628,
"grad_norm": 0.5004506241518043,
"learning_rate": 9.776955345633739e-05,
"loss": 1.1417,
"step": 2700
},
{
"epoch": 0.5575020610057708,
"grad_norm": 0.2897297596123754,
"learning_rate": 9.740987889437492e-05,
"loss": 0.9032,
"step": 2705
},
{
"epoch": 0.5585325638911789,
"grad_norm": 0.3767677442786407,
"learning_rate": 9.705023785919623e-05,
"loss": 1.1532,
"step": 2710
},
{
"epoch": 0.559563066776587,
"grad_norm": 0.25503990950782773,
"learning_rate": 9.669063500603049e-05,
"loss": 0.9414,
"step": 2715
},
{
"epoch": 0.560593569661995,
"grad_norm": 0.34636076916747177,
"learning_rate": 9.633107498961251e-05,
"loss": 1.0689,
"step": 2720
},
{
"epoch": 0.5616240725474031,
"grad_norm": 0.39789228504761426,
"learning_rate": 9.597156246412277e-05,
"loss": 1.1464,
"step": 2725
},
{
"epoch": 0.5626545754328112,
"grad_norm": 0.289581783982976,
"learning_rate": 9.561210208312694e-05,
"loss": 0.818,
"step": 2730
},
{
"epoch": 0.5636850783182193,
"grad_norm": 0.31157714978145534,
"learning_rate": 9.525269849951578e-05,
"loss": 1.2706,
"step": 2735
},
{
"epoch": 0.5647155812036274,
"grad_norm": 0.28372271464928434,
"learning_rate": 9.489335636544476e-05,
"loss": 0.9805,
"step": 2740
},
{
"epoch": 0.5657460840890355,
"grad_norm": 0.33997031384483695,
"learning_rate": 9.453408033227406e-05,
"loss": 1.0937,
"step": 2745
},
{
"epoch": 0.5667765869744436,
"grad_norm": 0.33417374019033036,
"learning_rate": 9.417487505050816e-05,
"loss": 1.2517,
"step": 2750
},
{
"epoch": 0.5678070898598516,
"grad_norm": 0.3379621672677856,
"learning_rate": 9.38157451697358e-05,
"loss": 0.8737,
"step": 2755
},
{
"epoch": 0.5688375927452597,
"grad_norm": 0.3589988298317977,
"learning_rate": 9.345669533856961e-05,
"loss": 1.1273,
"step": 2760
},
{
"epoch": 0.5698680956306678,
"grad_norm": 0.26916205391576065,
"learning_rate": 9.309773020458616e-05,
"loss": 0.9603,
"step": 2765
},
{
"epoch": 0.5708985985160758,
"grad_norm": 0.3159986677198098,
"learning_rate": 9.273885441426562e-05,
"loss": 1.077,
"step": 2770
},
{
"epoch": 0.5719291014014839,
"grad_norm": 0.3412008285509142,
"learning_rate": 9.238007261293176e-05,
"loss": 1.1973,
"step": 2775
},
{
"epoch": 0.572959604286892,
"grad_norm": 0.28834466612116516,
"learning_rate": 9.202138944469168e-05,
"loss": 0.8306,
"step": 2780
},
{
"epoch": 0.5739901071723,
"grad_norm": 0.38911649440218604,
"learning_rate": 9.16628095523758e-05,
"loss": 1.141,
"step": 2785
},
{
"epoch": 0.5750206100577082,
"grad_norm": 0.29683818561664493,
"learning_rate": 9.130433757747772e-05,
"loss": 0.9729,
"step": 2790
},
{
"epoch": 0.5760511129431163,
"grad_norm": 0.3445235007246779,
"learning_rate": 9.094597816009409e-05,
"loss": 0.9819,
"step": 2795
},
{
"epoch": 0.5770816158285244,
"grad_norm": 0.3661330469833344,
"learning_rate": 9.058773593886469e-05,
"loss": 1.1984,
"step": 2800
},
{
"epoch": 0.5770816158285244,
"eval_loss": 1.0007954835891724,
"eval_runtime": 2470.6884,
"eval_samples_per_second": 3.238,
"eval_steps_per_second": 0.202,
"step": 2800
},
{
"epoch": 0.5781121187139324,
"grad_norm": 0.32153311729783396,
"learning_rate": 9.022961555091226e-05,
"loss": 0.8678,
"step": 2805
},
{
"epoch": 0.5791426215993405,
"grad_norm": 0.36605355908566417,
"learning_rate": 8.987162163178256e-05,
"loss": 1.2198,
"step": 2810
},
{
"epoch": 0.5801731244847486,
"grad_norm": 0.2691696483789003,
"learning_rate": 8.951375881538421e-05,
"loss": 0.9082,
"step": 2815
},
{
"epoch": 0.5812036273701566,
"grad_norm": 0.35640081108055943,
"learning_rate": 8.915603173392895e-05,
"loss": 1.0764,
"step": 2820
},
{
"epoch": 0.5822341302555647,
"grad_norm": 0.33493558837647336,
"learning_rate": 8.87984450178715e-05,
"loss": 1.1807,
"step": 2825
},
{
"epoch": 0.5832646331409728,
"grad_norm": 0.31748865735569,
"learning_rate": 8.84410032958497e-05,
"loss": 0.8788,
"step": 2830
},
{
"epoch": 0.5842951360263808,
"grad_norm": 0.36563160381436116,
"learning_rate": 8.808371119462452e-05,
"loss": 1.1127,
"step": 2835
},
{
"epoch": 0.5853256389117889,
"grad_norm": 0.26343660014410963,
"learning_rate": 8.772657333902027e-05,
"loss": 0.9977,
"step": 2840
},
{
"epoch": 0.5863561417971971,
"grad_norm": 0.3695249008293323,
"learning_rate": 8.736959435186466e-05,
"loss": 1.095,
"step": 2845
},
{
"epoch": 0.5873866446826052,
"grad_norm": 0.35773020515251225,
"learning_rate": 8.701277885392909e-05,
"loss": 1.2322,
"step": 2850
},
{
"epoch": 0.5884171475680132,
"grad_norm": 0.3269028767537368,
"learning_rate": 8.665613146386854e-05,
"loss": 0.8593,
"step": 2855
},
{
"epoch": 0.5894476504534213,
"grad_norm": 0.3544711574787662,
"learning_rate": 8.629965679816217e-05,
"loss": 1.1952,
"step": 2860
},
{
"epoch": 0.5904781533388294,
"grad_norm": 0.29639050674293943,
"learning_rate": 8.594335947105328e-05,
"loss": 0.9516,
"step": 2865
},
{
"epoch": 0.5915086562242374,
"grad_norm": 0.3426897718147058,
"learning_rate": 8.558724409448974e-05,
"loss": 1.0779,
"step": 2870
},
{
"epoch": 0.5925391591096455,
"grad_norm": 0.3740367235927869,
"learning_rate": 8.523131527806423e-05,
"loss": 1.1865,
"step": 2875
},
{
"epoch": 0.5935696619950536,
"grad_norm": 0.31189806294414407,
"learning_rate": 8.48755776289545e-05,
"loss": 0.8847,
"step": 2880
},
{
"epoch": 0.5946001648804616,
"grad_norm": 0.37495206238491713,
"learning_rate": 8.452003575186394e-05,
"loss": 1.1804,
"step": 2885
},
{
"epoch": 0.5956306677658697,
"grad_norm": 0.2797206274129371,
"learning_rate": 8.416469424896167e-05,
"loss": 0.9762,
"step": 2890
},
{
"epoch": 0.5966611706512778,
"grad_norm": 0.33348859645525075,
"learning_rate": 8.380955771982332e-05,
"loss": 1.0883,
"step": 2895
},
{
"epoch": 0.597691673536686,
"grad_norm": 0.44145796098661705,
"learning_rate": 8.345463076137125e-05,
"loss": 1.166,
"step": 2900
},
{
"epoch": 0.598722176422094,
"grad_norm": 0.286114945740939,
"learning_rate": 8.309991796781511e-05,
"loss": 0.8535,
"step": 2905
},
{
"epoch": 0.5997526793075021,
"grad_norm": 0.34343359665934886,
"learning_rate": 8.274542393059235e-05,
"loss": 1.1742,
"step": 2910
},
{
"epoch": 0.6007831821929102,
"grad_norm": 0.2624287823547147,
"learning_rate": 8.239115323830889e-05,
"loss": 0.955,
"step": 2915
},
{
"epoch": 0.6018136850783182,
"grad_norm": 0.35154822360336263,
"learning_rate": 8.203711047667958e-05,
"loss": 1.0426,
"step": 2920
},
{
"epoch": 0.6028441879637263,
"grad_norm": 0.3344934169585091,
"learning_rate": 8.1683300228469e-05,
"loss": 1.1785,
"step": 2925
},
{
"epoch": 0.6038746908491344,
"grad_norm": 0.3209999819274115,
"learning_rate": 8.132972707343192e-05,
"loss": 0.8842,
"step": 2930
},
{
"epoch": 0.6049051937345424,
"grad_norm": 0.39454283654575434,
"learning_rate": 8.097639558825427e-05,
"loss": 1.1693,
"step": 2935
},
{
"epoch": 0.6059356966199505,
"grad_norm": 0.2878286096920221,
"learning_rate": 8.062331034649376e-05,
"loss": 0.8965,
"step": 2940
},
{
"epoch": 0.6069661995053586,
"grad_norm": 0.33688522439048935,
"learning_rate": 8.027047591852069e-05,
"loss": 1.1381,
"step": 2945
},
{
"epoch": 0.6079967023907666,
"grad_norm": 0.38189870355144323,
"learning_rate": 7.991789687145873e-05,
"loss": 1.1257,
"step": 2950
},
{
"epoch": 0.6090272052761748,
"grad_norm": 0.3051536804305413,
"learning_rate": 7.956557776912596e-05,
"loss": 0.9044,
"step": 2955
},
{
"epoch": 0.6100577081615829,
"grad_norm": 0.3451327149435641,
"learning_rate": 7.921352317197574e-05,
"loss": 1.2264,
"step": 2960
},
{
"epoch": 0.611088211046991,
"grad_norm": 0.2827004152527063,
"learning_rate": 7.886173763703757e-05,
"loss": 0.9794,
"step": 2965
},
{
"epoch": 0.612118713932399,
"grad_norm": 0.40169478150821714,
"learning_rate": 7.851022571785819e-05,
"loss": 1.1179,
"step": 2970
},
{
"epoch": 0.6131492168178071,
"grad_norm": 0.4438701028140135,
"learning_rate": 7.815899196444267e-05,
"loss": 1.1662,
"step": 2975
},
{
"epoch": 0.6141797197032152,
"grad_norm": 0.3254217005122457,
"learning_rate": 7.780804092319547e-05,
"loss": 0.8434,
"step": 2980
},
{
"epoch": 0.6152102225886232,
"grad_norm": 0.3284145654312824,
"learning_rate": 7.745737713686152e-05,
"loss": 1.2109,
"step": 2985
},
{
"epoch": 0.6162407254740313,
"grad_norm": 0.28117376745819467,
"learning_rate": 7.710700514446762e-05,
"loss": 0.955,
"step": 2990
},
{
"epoch": 0.6172712283594394,
"grad_norm": 0.35764651325723473,
"learning_rate": 7.675692948126345e-05,
"loss": 1.032,
"step": 2995
},
{
"epoch": 0.6183017312448474,
"grad_norm": 0.4045276967582808,
"learning_rate": 7.640715467866307e-05,
"loss": 1.157,
"step": 3000
},
{
"epoch": 0.6183017312448474,
"eval_loss": 0.9929932951927185,
"eval_runtime": 2471.8115,
"eval_samples_per_second": 3.236,
"eval_steps_per_second": 0.202,
"step": 3000
},
{
"epoch": 0.6193322341302555,
"grad_norm": 0.3148405737434604,
"learning_rate": 7.605768526418605e-05,
"loss": 0.8743,
"step": 3005
},
{
"epoch": 0.6203627370156637,
"grad_norm": 0.35877325985884195,
"learning_rate": 7.57085257613991e-05,
"loss": 1.2573,
"step": 3010
},
{
"epoch": 0.6213932399010718,
"grad_norm": 0.2728312132512685,
"learning_rate": 7.535968068985737e-05,
"loss": 0.9663,
"step": 3015
},
{
"epoch": 0.6224237427864798,
"grad_norm": 0.3236792895371986,
"learning_rate": 7.501115456504595e-05,
"loss": 1.1109,
"step": 3020
},
{
"epoch": 0.6234542456718879,
"grad_norm": 0.40547856146336464,
"learning_rate": 7.466295189832148e-05,
"loss": 1.1882,
"step": 3025
},
{
"epoch": 0.624484748557296,
"grad_norm": 0.26523329910607873,
"learning_rate": 7.431507719685371e-05,
"loss": 0.8277,
"step": 3030
},
{
"epoch": 0.625515251442704,
"grad_norm": 0.3252197516332192,
"learning_rate": 7.396753496356718e-05,
"loss": 1.2303,
"step": 3035
},
{
"epoch": 0.6265457543281121,
"grad_norm": 0.2664136991956755,
"learning_rate": 7.362032969708297e-05,
"loss": 0.9909,
"step": 3040
},
{
"epoch": 0.6275762572135202,
"grad_norm": 0.3438706008672591,
"learning_rate": 7.327346589166035e-05,
"loss": 1.057,
"step": 3045
},
{
"epoch": 0.6286067600989282,
"grad_norm": 0.40393507740091983,
"learning_rate": 7.292694803713871e-05,
"loss": 1.1674,
"step": 3050
},
{
"epoch": 0.6296372629843363,
"grad_norm": 0.2966521374455327,
"learning_rate": 7.258078061887947e-05,
"loss": 0.8634,
"step": 3055
},
{
"epoch": 0.6306677658697445,
"grad_norm": 0.3313252298668368,
"learning_rate": 7.223496811770796e-05,
"loss": 1.1705,
"step": 3060
},
{
"epoch": 0.6316982687551526,
"grad_norm": 0.282517205910409,
"learning_rate": 7.188951500985533e-05,
"loss": 0.9568,
"step": 3065
},
{
"epoch": 0.6327287716405606,
"grad_norm": 0.3423739475910178,
"learning_rate": 7.154442576690083e-05,
"loss": 1.1215,
"step": 3070
},
{
"epoch": 0.6337592745259687,
"grad_norm": 0.3951313592694174,
"learning_rate": 7.119970485571375e-05,
"loss": 1.2196,
"step": 3075
},
{
"epoch": 0.6347897774113768,
"grad_norm": 0.27968067577198935,
"learning_rate": 7.08553567383956e-05,
"loss": 0.8517,
"step": 3080
},
{
"epoch": 0.6358202802967848,
"grad_norm": 0.3967375078625761,
"learning_rate": 7.051138587222255e-05,
"loss": 1.1345,
"step": 3085
},
{
"epoch": 0.6368507831821929,
"grad_norm": 0.275638071848918,
"learning_rate": 7.016779670958746e-05,
"loss": 0.9428,
"step": 3090
},
{
"epoch": 0.637881286067601,
"grad_norm": 0.3876563491000536,
"learning_rate": 6.982459369794247e-05,
"loss": 1.0172,
"step": 3095
},
{
"epoch": 0.638911788953009,
"grad_norm": 0.39932395010881844,
"learning_rate": 6.948178127974127e-05,
"loss": 1.136,
"step": 3100
},
{
"epoch": 0.6399422918384171,
"grad_norm": 0.344145718815036,
"learning_rate": 6.913936389238174e-05,
"loss": 0.9074,
"step": 3105
},
{
"epoch": 0.6409727947238252,
"grad_norm": 0.35702868127871346,
"learning_rate": 6.879734596814839e-05,
"loss": 1.2245,
"step": 3110
},
{
"epoch": 0.6420032976092334,
"grad_norm": 0.278189478468898,
"learning_rate": 6.845573193415508e-05,
"loss": 0.9795,
"step": 3115
},
{
"epoch": 0.6430338004946414,
"grad_norm": 0.3820554155556976,
"learning_rate": 6.811452621228766e-05,
"loss": 1.0815,
"step": 3120
},
{
"epoch": 0.6440643033800495,
"grad_norm": 0.42159702416774886,
"learning_rate": 6.777373321914671e-05,
"loss": 1.2058,
"step": 3125
},
{
"epoch": 0.6450948062654576,
"grad_norm": 0.30363813905867076,
"learning_rate": 6.743335736599045e-05,
"loss": 0.853,
"step": 3130
},
{
"epoch": 0.6461253091508656,
"grad_norm": 0.3626110839853577,
"learning_rate": 6.709340305867762e-05,
"loss": 1.1719,
"step": 3135
},
{
"epoch": 0.6471558120362737,
"grad_norm": 0.2741980722413254,
"learning_rate": 6.675387469761033e-05,
"loss": 0.9857,
"step": 3140
},
{
"epoch": 0.6481863149216818,
"grad_norm": 0.3458024372951965,
"learning_rate": 6.641477667767738e-05,
"loss": 1.0737,
"step": 3145
},
{
"epoch": 0.6492168178070898,
"grad_norm": 0.3659993842021884,
"learning_rate": 6.607611338819697e-05,
"loss": 1.2332,
"step": 3150
},
{
"epoch": 0.6502473206924979,
"grad_norm": 0.3134224056811165,
"learning_rate": 6.573788921286028e-05,
"loss": 0.833,
"step": 3155
},
{
"epoch": 0.651277823577906,
"grad_norm": 0.31557728275482577,
"learning_rate": 6.540010852967447e-05,
"loss": 1.169,
"step": 3160
},
{
"epoch": 0.652308326463314,
"grad_norm": 0.2745096877567603,
"learning_rate": 6.506277571090613e-05,
"loss": 0.9294,
"step": 3165
},
{
"epoch": 0.6533388293487222,
"grad_norm": 0.37277814205936943,
"learning_rate": 6.47258951230246e-05,
"loss": 1.081,
"step": 3170
},
{
"epoch": 0.6543693322341303,
"grad_norm": 0.38045711766200485,
"learning_rate": 6.438947112664555e-05,
"loss": 1.1493,
"step": 3175
},
{
"epoch": 0.6553998351195384,
"grad_norm": 0.2706340588373179,
"learning_rate": 6.405350807647444e-05,
"loss": 0.8498,
"step": 3180
},
{
"epoch": 0.6564303380049464,
"grad_norm": 0.3336018890725225,
"learning_rate": 6.371801032125026e-05,
"loss": 1.1778,
"step": 3185
},
{
"epoch": 0.6574608408903545,
"grad_norm": 0.2913475821459678,
"learning_rate": 6.338298220368912e-05,
"loss": 0.9571,
"step": 3190
},
{
"epoch": 0.6584913437757626,
"grad_norm": 0.36985494968453825,
"learning_rate": 6.304842806042812e-05,
"loss": 1.0778,
"step": 3195
},
{
"epoch": 0.6595218466611706,
"grad_norm": 0.40407830600531813,
"learning_rate": 6.271435222196916e-05,
"loss": 1.1542,
"step": 3200
},
{
"epoch": 0.6595218466611706,
"eval_loss": 0.9862346649169922,
"eval_runtime": 2465.9653,
"eval_samples_per_second": 3.244,
"eval_steps_per_second": 0.203,
"step": 3200
},
{
"epoch": 0.6605523495465787,
"grad_norm": 0.2524320710636017,
"learning_rate": 6.238075901262293e-05,
"loss": 0.8157,
"step": 3205
},
{
"epoch": 0.6615828524319868,
"grad_norm": 0.3278601577235125,
"learning_rate": 6.204765275045298e-05,
"loss": 1.1461,
"step": 3210
},
{
"epoch": 0.6626133553173948,
"grad_norm": 0.28938045624305053,
"learning_rate": 6.171503774721966e-05,
"loss": 0.9831,
"step": 3215
},
{
"epoch": 0.6636438582028029,
"grad_norm": 0.3531127055799286,
"learning_rate": 6.13829183083245e-05,
"loss": 1.0438,
"step": 3220
},
{
"epoch": 0.6646743610882111,
"grad_norm": 0.3337288007445976,
"learning_rate": 6.105129873275435e-05,
"loss": 1.2684,
"step": 3225
},
{
"epoch": 0.6657048639736192,
"grad_norm": 0.30432663000679183,
"learning_rate": 6.072018331302577e-05,
"loss": 0.8405,
"step": 3230
},
{
"epoch": 0.6667353668590272,
"grad_norm": 0.3481398036760903,
"learning_rate": 6.038957633512957e-05,
"loss": 1.1838,
"step": 3235
},
{
"epoch": 0.6677658697444353,
"grad_norm": 0.27336340413608334,
"learning_rate": 6.005948207847516e-05,
"loss": 0.9207,
"step": 3240
},
{
"epoch": 0.6687963726298434,
"grad_norm": 0.3440287779317674,
"learning_rate": 5.97299048158352e-05,
"loss": 1.0703,
"step": 3245
},
{
"epoch": 0.6698268755152514,
"grad_norm": 0.39581616124415014,
"learning_rate": 5.940084881329042e-05,
"loss": 1.2125,
"step": 3250
},
{
"epoch": 0.6708573784006595,
"grad_norm": 0.2935018842179493,
"learning_rate": 5.907231833017424e-05,
"loss": 0.8331,
"step": 3255
},
{
"epoch": 0.6718878812860676,
"grad_norm": 0.3836289298282413,
"learning_rate": 5.8744317619017755e-05,
"loss": 1.2524,
"step": 3260
},
{
"epoch": 0.6729183841714756,
"grad_norm": 0.2710041280727128,
"learning_rate": 5.841685092549456e-05,
"loss": 0.9405,
"step": 3265
},
{
"epoch": 0.6739488870568837,
"grad_norm": 0.32969149065243836,
"learning_rate": 5.8089922488365975e-05,
"loss": 1.0393,
"step": 3270
},
{
"epoch": 0.6749793899422918,
"grad_norm": 0.336443888291771,
"learning_rate": 5.776353653942602e-05,
"loss": 1.2373,
"step": 3275
},
{
"epoch": 0.6760098928277,
"grad_norm": 0.30665104617457667,
"learning_rate": 5.743769730344666e-05,
"loss": 0.8681,
"step": 3280
},
{
"epoch": 0.677040395713108,
"grad_norm": 0.3625802330226177,
"learning_rate": 5.7112408998123256e-05,
"loss": 1.2316,
"step": 3285
},
{
"epoch": 0.6780708985985161,
"grad_norm": 0.2953674696158516,
"learning_rate": 5.678767583401974e-05,
"loss": 0.9204,
"step": 3290
},
{
"epoch": 0.6791014014839242,
"grad_norm": 0.38225007623929924,
"learning_rate": 5.646350201451438e-05,
"loss": 1.066,
"step": 3295
},
{
"epoch": 0.6801319043693322,
"grad_norm": 0.4406303775855105,
"learning_rate": 5.613989173574512e-05,
"loss": 1.1569,
"step": 3300
},
{
"epoch": 0.6811624072547403,
"grad_norm": 0.2882802372851351,
"learning_rate": 5.5816849186555386e-05,
"loss": 0.86,
"step": 3305
},
{
"epoch": 0.6821929101401484,
"grad_norm": 0.3780252315323207,
"learning_rate": 5.549437854843995e-05,
"loss": 1.1646,
"step": 3310
},
{
"epoch": 0.6832234130255564,
"grad_norm": 0.2834828395630223,
"learning_rate": 5.517248399549063e-05,
"loss": 0.973,
"step": 3315
},
{
"epoch": 0.6842539159109645,
"grad_norm": 0.3721150712296605,
"learning_rate": 5.48511696943423e-05,
"loss": 1.0507,
"step": 3320
},
{
"epoch": 0.6852844187963726,
"grad_norm": 0.3748866522580312,
"learning_rate": 5.4530439804119096e-05,
"loss": 1.1834,
"step": 3325
},
{
"epoch": 0.6863149216817807,
"grad_norm": 0.2887534766618149,
"learning_rate": 5.4210298476380484e-05,
"loss": 0.888,
"step": 3330
},
{
"epoch": 0.6873454245671888,
"grad_norm": 0.3409152912595764,
"learning_rate": 5.38907498550674e-05,
"loss": 1.1606,
"step": 3335
},
{
"epoch": 0.6883759274525969,
"grad_norm": 0.29326388422606586,
"learning_rate": 5.357179807644887e-05,
"loss": 0.8845,
"step": 3340
},
{
"epoch": 0.689406430338005,
"grad_norm": 0.3589561770883435,
"learning_rate": 5.3253447269068245e-05,
"loss": 1.0743,
"step": 3345
},
{
"epoch": 0.690436933223413,
"grad_norm": 0.38383916804262924,
"learning_rate": 5.293570155368981e-05,
"loss": 1.183,
"step": 3350
},
{
"epoch": 0.6914674361088211,
"grad_norm": 0.3068092183719897,
"learning_rate": 5.261856504324563e-05,
"loss": 0.881,
"step": 3355
},
{
"epoch": 0.6924979389942292,
"grad_norm": 0.36385771675879847,
"learning_rate": 5.230204184278195e-05,
"loss": 1.2467,
"step": 3360
},
{
"epoch": 0.6935284418796372,
"grad_norm": 0.2907466758494882,
"learning_rate": 5.198613604940649e-05,
"loss": 0.9442,
"step": 3365
},
{
"epoch": 0.6945589447650453,
"grad_norm": 0.3566322978977828,
"learning_rate": 5.1670851752235025e-05,
"loss": 1.0632,
"step": 3370
},
{
"epoch": 0.6955894476504534,
"grad_norm": 0.36835384306712526,
"learning_rate": 5.135619303233867e-05,
"loss": 1.1819,
"step": 3375
},
{
"epoch": 0.6966199505358615,
"grad_norm": 0.30530163857595344,
"learning_rate": 5.104216396269109e-05,
"loss": 0.8697,
"step": 3380
},
{
"epoch": 0.6976504534212696,
"grad_norm": 0.3728398219157709,
"learning_rate": 5.072876860811553e-05,
"loss": 1.2065,
"step": 3385
},
{
"epoch": 0.6986809563066777,
"grad_norm": 0.2719186126381508,
"learning_rate": 5.0416011025232546e-05,
"loss": 0.9332,
"step": 3390
},
{
"epoch": 0.6997114591920858,
"grad_norm": 0.39560521280304406,
"learning_rate": 5.010389526240719e-05,
"loss": 1.0035,
"step": 3395
},
{
"epoch": 0.7007419620774938,
"grad_norm": 0.45096201363438326,
"learning_rate": 4.979242535969672e-05,
"loss": 1.1648,
"step": 3400
},
{
"epoch": 0.7007419620774938,
"eval_loss": 0.9801868796348572,
"eval_runtime": 2471.0518,
"eval_samples_per_second": 3.237,
"eval_steps_per_second": 0.202,
"step": 3400
},
{
"epoch": 0.7017724649629019,
"grad_norm": 0.3253388221391844,
"learning_rate": 4.9481605348798435e-05,
"loss": 0.8283,
"step": 3405
},
{
"epoch": 0.70280296784831,
"grad_norm": 0.3382283949543721,
"learning_rate": 4.917143925299728e-05,
"loss": 1.1642,
"step": 3410
},
{
"epoch": 0.703833470733718,
"grad_norm": 0.2650304501518648,
"learning_rate": 4.886193108711378e-05,
"loss": 0.9334,
"step": 3415
},
{
"epoch": 0.7048639736191261,
"grad_norm": 0.3163351222719128,
"learning_rate": 4.8553084857452426e-05,
"loss": 1.0316,
"step": 3420
},
{
"epoch": 0.7058944765045342,
"grad_norm": 0.4184991045727221,
"learning_rate": 4.824490456174926e-05,
"loss": 1.1313,
"step": 3425
},
{
"epoch": 0.7069249793899423,
"grad_norm": 0.31621686065106913,
"learning_rate": 4.7937394189120485e-05,
"loss": 0.8658,
"step": 3430
},
{
"epoch": 0.7079554822753503,
"grad_norm": 0.3252525212780396,
"learning_rate": 4.763055772001086e-05,
"loss": 1.1749,
"step": 3435
},
{
"epoch": 0.7089859851607585,
"grad_norm": 0.28077212260765644,
"learning_rate": 4.732439912614195e-05,
"loss": 0.9195,
"step": 3440
},
{
"epoch": 0.7100164880461666,
"grad_norm": 0.3404715984560569,
"learning_rate": 4.7018922370460835e-05,
"loss": 1.1004,
"step": 3445
},
{
"epoch": 0.7110469909315746,
"grad_norm": 0.3572114714960076,
"learning_rate": 4.671413140708893e-05,
"loss": 1.1555,
"step": 3450
},
{
"epoch": 0.7120774938169827,
"grad_norm": 0.31012926926152484,
"learning_rate": 4.6410030181270546e-05,
"loss": 0.8328,
"step": 3455
},
{
"epoch": 0.7131079967023908,
"grad_norm": 0.3814244438557788,
"learning_rate": 4.610662262932209e-05,
"loss": 1.2221,
"step": 3460
},
{
"epoch": 0.7141384995877988,
"grad_norm": 0.2587030449835817,
"learning_rate": 4.5803912678580906e-05,
"loss": 0.923,
"step": 3465
},
{
"epoch": 0.7151690024732069,
"grad_norm": 0.36964920266690043,
"learning_rate": 4.5501904247354474e-05,
"loss": 1.0765,
"step": 3470
},
{
"epoch": 0.716199505358615,
"grad_norm": 0.4007947343253882,
"learning_rate": 4.520060124486989e-05,
"loss": 1.0613,
"step": 3475
},
{
"epoch": 0.717230008244023,
"grad_norm": 0.31355619965035286,
"learning_rate": 4.4900007571222946e-05,
"loss": 0.8406,
"step": 3480
},
{
"epoch": 0.7182605111294311,
"grad_norm": 0.3709230553408382,
"learning_rate": 4.460012711732795e-05,
"loss": 1.202,
"step": 3485
},
{
"epoch": 0.7192910140148392,
"grad_norm": 0.2663391335521852,
"learning_rate": 4.430096376486713e-05,
"loss": 0.957,
"step": 3490
},
{
"epoch": 0.7203215169002474,
"grad_norm": 0.36149156644544367,
"learning_rate": 4.4002521386240466e-05,
"loss": 1.0885,
"step": 3495
},
{
"epoch": 0.7213520197856554,
"grad_norm": 0.4116390911425264,
"learning_rate": 4.3704803844515705e-05,
"loss": 1.1649,
"step": 3500
},
{
"epoch": 0.7223825226710635,
"grad_norm": 0.3337684750408035,
"learning_rate": 4.3407814993378095e-05,
"loss": 0.8993,
"step": 3505
},
{
"epoch": 0.7234130255564716,
"grad_norm": 0.3496736243866778,
"learning_rate": 4.311155867708071e-05,
"loss": 1.1648,
"step": 3510
},
{
"epoch": 0.7244435284418796,
"grad_norm": 0.28479774144881315,
"learning_rate": 4.2816038730394656e-05,
"loss": 0.9484,
"step": 3515
},
{
"epoch": 0.7254740313272877,
"grad_norm": 0.36918418554896276,
"learning_rate": 4.252125897855932e-05,
"loss": 1.0583,
"step": 3520
},
{
"epoch": 0.7265045342126958,
"grad_norm": 0.33704337919665334,
"learning_rate": 4.222722323723294e-05,
"loss": 1.1907,
"step": 3525
},
{
"epoch": 0.7275350370981039,
"grad_norm": 0.3148307470632366,
"learning_rate": 4.1933935312443286e-05,
"loss": 0.8783,
"step": 3530
},
{
"epoch": 0.7285655399835119,
"grad_norm": 0.3517810748272734,
"learning_rate": 4.164139900053824e-05,
"loss": 1.1717,
"step": 3535
},
{
"epoch": 0.72959604286892,
"grad_norm": 0.2855842847206538,
"learning_rate": 4.134961808813672e-05,
"loss": 0.9112,
"step": 3540
},
{
"epoch": 0.7306265457543281,
"grad_norm": 0.34108259740221386,
"learning_rate": 4.1058596352079805e-05,
"loss": 1.0142,
"step": 3545
},
{
"epoch": 0.7316570486397362,
"grad_norm": 0.45538302782398454,
"learning_rate": 4.076833755938153e-05,
"loss": 1.0974,
"step": 3550
},
{
"epoch": 0.7326875515251443,
"grad_norm": 0.31487990147891376,
"learning_rate": 4.0478845467180506e-05,
"loss": 0.8298,
"step": 3555
},
{
"epoch": 0.7337180544105524,
"grad_norm": 0.36935273464286145,
"learning_rate": 4.0190123822690965e-05,
"loss": 1.1436,
"step": 3560
},
{
"epoch": 0.7347485572959604,
"grad_norm": 0.27930316978222564,
"learning_rate": 3.990217636315441e-05,
"loss": 0.9989,
"step": 3565
},
{
"epoch": 0.7357790601813685,
"grad_norm": 0.3714521960753657,
"learning_rate": 3.961500681579129e-05,
"loss": 1.1123,
"step": 3570
},
{
"epoch": 0.7368095630667766,
"grad_norm": 0.43314908676280006,
"learning_rate": 3.9328618897752566e-05,
"loss": 1.1162,
"step": 3575
},
{
"epoch": 0.7378400659521847,
"grad_norm": 0.29947090161111584,
"learning_rate": 3.904301631607186e-05,
"loss": 0.8351,
"step": 3580
},
{
"epoch": 0.7388705688375927,
"grad_norm": 0.4143197770113969,
"learning_rate": 3.875820276761717e-05,
"loss": 1.2288,
"step": 3585
},
{
"epoch": 0.7399010717230008,
"grad_norm": 0.2821180575460458,
"learning_rate": 3.847418193904324e-05,
"loss": 0.9842,
"step": 3590
},
{
"epoch": 0.7409315746084089,
"grad_norm": 0.35222923714451126,
"learning_rate": 3.8190957506743806e-05,
"loss": 1.0377,
"step": 3595
},
{
"epoch": 0.7419620774938169,
"grad_norm": 0.36695400370029363,
"learning_rate": 3.790853313680389e-05,
"loss": 1.1403,
"step": 3600
},
{
"epoch": 0.7419620774938169,
"eval_loss": 0.9750496745109558,
"eval_runtime": 2464.5637,
"eval_samples_per_second": 3.246,
"eval_steps_per_second": 0.203,
"step": 3600
},
{
"epoch": 0.7429925803792251,
"grad_norm": 0.33168677598566876,
"learning_rate": 3.7626912484952495e-05,
"loss": 0.9223,
"step": 3605
},
{
"epoch": 0.7440230832646332,
"grad_norm": 0.3685099675858659,
"learning_rate": 3.734609919651523e-05,
"loss": 1.22,
"step": 3610
},
{
"epoch": 0.7450535861500412,
"grad_norm": 0.30519501707613905,
"learning_rate": 3.706609690636703e-05,
"loss": 0.9541,
"step": 3615
},
{
"epoch": 0.7460840890354493,
"grad_norm": 0.3227606827521623,
"learning_rate": 3.6786909238885215e-05,
"loss": 1.002,
"step": 3620
},
{
"epoch": 0.7471145919208574,
"grad_norm": 0.3907472086101187,
"learning_rate": 3.650853980790262e-05,
"loss": 1.1461,
"step": 3625
},
{
"epoch": 0.7481450948062655,
"grad_norm": 0.2919486775395443,
"learning_rate": 3.6230992216660664e-05,
"loss": 0.8063,
"step": 3630
},
{
"epoch": 0.7491755976916735,
"grad_norm": 0.32875886981694724,
"learning_rate": 3.595427005776281e-05,
"loss": 1.1514,
"step": 3635
},
{
"epoch": 0.7502061005770816,
"grad_norm": 0.29158449116331836,
"learning_rate": 3.5678376913128075e-05,
"loss": 0.9759,
"step": 3640
},
{
"epoch": 0.7512366034624897,
"grad_norm": 0.37631936130463955,
"learning_rate": 3.540331635394458e-05,
"loss": 0.9905,
"step": 3645
},
{
"epoch": 0.7522671063478977,
"grad_norm": 0.3439508621054155,
"learning_rate": 3.512909194062347e-05,
"loss": 1.1949,
"step": 3650
},
{
"epoch": 0.7532976092333058,
"grad_norm": 0.3146397920651041,
"learning_rate": 3.485570722275264e-05,
"loss": 0.8706,
"step": 3655
},
{
"epoch": 0.754328112118714,
"grad_norm": 0.38258052872478915,
"learning_rate": 3.458316573905087e-05,
"loss": 1.2172,
"step": 3660
},
{
"epoch": 0.755358615004122,
"grad_norm": 0.2793437890823054,
"learning_rate": 3.4311471017322175e-05,
"loss": 0.9391,
"step": 3665
},
{
"epoch": 0.7563891178895301,
"grad_norm": 0.3643128847802364,
"learning_rate": 3.4040626574409815e-05,
"loss": 0.9938,
"step": 3670
},
{
"epoch": 0.7574196207749382,
"grad_norm": 0.3709320339736749,
"learning_rate": 3.377063591615113e-05,
"loss": 1.1904,
"step": 3675
},
{
"epoch": 0.7584501236603463,
"grad_norm": 0.33437847713940755,
"learning_rate": 3.350150253733186e-05,
"loss": 0.9233,
"step": 3680
},
{
"epoch": 0.7594806265457543,
"grad_norm": 0.3607224445907033,
"learning_rate": 3.3233229921641064e-05,
"loss": 1.2125,
"step": 3685
},
{
"epoch": 0.7605111294311624,
"grad_norm": 0.27198443746283474,
"learning_rate": 3.296582154162604e-05,
"loss": 0.9196,
"step": 3690
},
{
"epoch": 0.7615416323165705,
"grad_norm": 0.3668276085455235,
"learning_rate": 3.2699280858647327e-05,
"loss": 1.0335,
"step": 3695
},
{
"epoch": 0.7625721352019785,
"grad_norm": 0.431636600953154,
"learning_rate": 3.2433611322833845e-05,
"loss": 1.1378,
"step": 3700
},
{
"epoch": 0.7636026380873866,
"grad_norm": 0.3284008231342369,
"learning_rate": 3.216881637303839e-05,
"loss": 0.8308,
"step": 3705
},
{
"epoch": 0.7646331409727948,
"grad_norm": 0.3443882838132438,
"learning_rate": 3.190489943679297e-05,
"loss": 1.1797,
"step": 3710
},
{
"epoch": 0.7656636438582028,
"grad_norm": 0.30472428853778355,
"learning_rate": 3.164186393026445e-05,
"loss": 0.8953,
"step": 3715
},
{
"epoch": 0.7666941467436109,
"grad_norm": 0.346855761423586,
"learning_rate": 3.137971325821054e-05,
"loss": 1.1678,
"step": 3720
},
{
"epoch": 0.767724649629019,
"grad_norm": 0.3586318210457773,
"learning_rate": 3.111845081393542e-05,
"loss": 1.1801,
"step": 3725
},
{
"epoch": 0.768755152514427,
"grad_norm": 0.3012111522622655,
"learning_rate": 3.0858079979245965e-05,
"loss": 0.8259,
"step": 3730
},
{
"epoch": 0.7697856553998351,
"grad_norm": 0.3675771233203778,
"learning_rate": 3.059860412440811e-05,
"loss": 1.1379,
"step": 3735
},
{
"epoch": 0.7708161582852432,
"grad_norm": 0.2997972891151249,
"learning_rate": 3.0340026608102902e-05,
"loss": 0.9012,
"step": 3740
},
{
"epoch": 0.7718466611706513,
"grad_norm": 0.3350924890911528,
"learning_rate": 3.008235077738334e-05,
"loss": 1.0476,
"step": 3745
},
{
"epoch": 0.7728771640560593,
"grad_norm": 0.42578144522976874,
"learning_rate": 2.9825579967630846e-05,
"loss": 1.2014,
"step": 3750
},
{
"epoch": 0.7739076669414674,
"grad_norm": 0.31905668249953223,
"learning_rate": 2.956971750251215e-05,
"loss": 0.9089,
"step": 3755
},
{
"epoch": 0.7749381698268755,
"grad_norm": 0.3320270260748608,
"learning_rate": 2.9314766693936356e-05,
"loss": 1.1712,
"step": 3760
},
{
"epoch": 0.7759686727122836,
"grad_norm": 0.26880506131338233,
"learning_rate": 2.906073084201191e-05,
"loss": 0.9532,
"step": 3765
},
{
"epoch": 0.7769991755976917,
"grad_norm": 0.360515318139803,
"learning_rate": 2.8807613235004037e-05,
"loss": 1.0235,
"step": 3770
},
{
"epoch": 0.7780296784830998,
"grad_norm": 0.40012910311865413,
"learning_rate": 2.855541714929206e-05,
"loss": 1.1292,
"step": 3775
},
{
"epoch": 0.7790601813685079,
"grad_norm": 0.3056042121876613,
"learning_rate": 2.8304145849327036e-05,
"loss": 0.7814,
"step": 3780
},
{
"epoch": 0.7800906842539159,
"grad_norm": 0.3475474680120366,
"learning_rate": 2.8053802587589538e-05,
"loss": 1.1668,
"step": 3785
},
{
"epoch": 0.781121187139324,
"grad_norm": 0.2843659913958508,
"learning_rate": 2.7804390604547557e-05,
"loss": 0.9237,
"step": 3790
},
{
"epoch": 0.7821516900247321,
"grad_norm": 0.36497922605390365,
"learning_rate": 2.7555913128614398e-05,
"loss": 1.0276,
"step": 3795
},
{
"epoch": 0.7831821929101401,
"grad_norm": 0.39023528771913,
"learning_rate": 2.7308373376107142e-05,
"loss": 1.1268,
"step": 3800
},
{
"epoch": 0.7831821929101401,
"eval_loss": 0.9705116748809814,
"eval_runtime": 2662.1546,
"eval_samples_per_second": 3.005,
"eval_steps_per_second": 0.188,
"step": 3800
},
{
"epoch": 0.7842126957955482,
"grad_norm": 0.3292637316481058,
"learning_rate": 2.706177455120482e-05,
"loss": 0.8586,
"step": 3805
},
{
"epoch": 0.7852431986809563,
"grad_norm": 0.3463786306953725,
"learning_rate": 2.681611984590696e-05,
"loss": 1.1979,
"step": 3810
},
{
"epoch": 0.7862737015663643,
"grad_norm": 0.3033390300653325,
"learning_rate": 2.6571412439992437e-05,
"loss": 0.9448,
"step": 3815
},
{
"epoch": 0.7873042044517725,
"grad_norm": 0.3744585100233523,
"learning_rate": 2.6327655500978076e-05,
"loss": 1.0825,
"step": 3820
},
{
"epoch": 0.7883347073371806,
"grad_norm": 0.3888535219196645,
"learning_rate": 2.608485218407779e-05,
"loss": 1.1602,
"step": 3825
},
{
"epoch": 0.7893652102225887,
"grad_norm": 0.3281836015354183,
"learning_rate": 2.5843005632161787e-05,
"loss": 0.8071,
"step": 3830
},
{
"epoch": 0.7903957131079967,
"grad_norm": 0.44764397719301374,
"learning_rate": 2.5602118975715683e-05,
"loss": 1.223,
"step": 3835
},
{
"epoch": 0.7914262159934048,
"grad_norm": 0.2605243101554053,
"learning_rate": 2.5362195332800253e-05,
"loss": 0.9269,
"step": 3840
},
{
"epoch": 0.7924567188788129,
"grad_norm": 0.34490759623743283,
"learning_rate": 2.5123237809010836e-05,
"loss": 0.9951,
"step": 3845
},
{
"epoch": 0.7934872217642209,
"grad_norm": 0.45219071339330563,
"learning_rate": 2.4885249497437223e-05,
"loss": 1.1591,
"step": 3850
},
{
"epoch": 0.794517724649629,
"grad_norm": 0.3701824230659728,
"learning_rate": 2.4648233478623705e-05,
"loss": 0.9237,
"step": 3855
},
{
"epoch": 0.7955482275350371,
"grad_norm": 0.3540574000256143,
"learning_rate": 2.4412192820529034e-05,
"loss": 1.1936,
"step": 3860
},
{
"epoch": 0.7965787304204451,
"grad_norm": 0.2979316113245556,
"learning_rate": 2.4177130578486885e-05,
"loss": 0.8818,
"step": 3865
},
{
"epoch": 0.7976092333058532,
"grad_norm": 0.3972288442858004,
"learning_rate": 2.3943049795166126e-05,
"loss": 1.0352,
"step": 3870
},
{
"epoch": 0.7986397361912614,
"grad_norm": 0.43218350490127366,
"learning_rate": 2.370995350053157e-05,
"loss": 1.141,
"step": 3875
},
{
"epoch": 0.7996702390766695,
"grad_norm": 0.33536259992407524,
"learning_rate": 2.3477844711804708e-05,
"loss": 0.8845,
"step": 3880
},
{
"epoch": 0.8007007419620775,
"grad_norm": 0.3715539599255406,
"learning_rate": 2.3246726433424716e-05,
"loss": 1.1807,
"step": 3885
},
{
"epoch": 0.8017312448474856,
"grad_norm": 0.27142620043338195,
"learning_rate": 2.301660165700936e-05,
"loss": 0.9382,
"step": 3890
},
{
"epoch": 0.8027617477328937,
"grad_norm": 0.3200117930180246,
"learning_rate": 2.2787473361316592e-05,
"loss": 1.1648,
"step": 3895
},
{
"epoch": 0.8037922506183017,
"grad_norm": 0.3832079088616207,
"learning_rate": 2.2559344512205705e-05,
"loss": 1.1815,
"step": 3900
},
{
"epoch": 0.8048227535037098,
"grad_norm": 0.28512724726672417,
"learning_rate": 2.233221806259903e-05,
"loss": 0.7237,
"step": 3905
},
{
"epoch": 0.8058532563891179,
"grad_norm": 0.347779304749182,
"learning_rate": 2.2106096952443888e-05,
"loss": 1.1468,
"step": 3910
},
{
"epoch": 0.8068837592745259,
"grad_norm": 0.27344000922062256,
"learning_rate": 2.188098410867424e-05,
"loss": 0.967,
"step": 3915
},
{
"epoch": 0.807914262159934,
"grad_norm": 0.35776849097052316,
"learning_rate": 2.165688244517299e-05,
"loss": 1.0351,
"step": 3920
},
{
"epoch": 0.8089447650453421,
"grad_norm": 0.6146898482449163,
"learning_rate": 2.143379486273428e-05,
"loss": 1.1856,
"step": 3925
},
{
"epoch": 0.8099752679307503,
"grad_norm": 0.37385831712675915,
"learning_rate": 2.1211724249025787e-05,
"loss": 0.9441,
"step": 3930
},
{
"epoch": 0.8110057708161583,
"grad_norm": 0.33568983400788555,
"learning_rate": 2.099067347855157e-05,
"loss": 1.2175,
"step": 3935
},
{
"epoch": 0.8120362737015664,
"grad_norm": 0.2905285106790405,
"learning_rate": 2.077064541261462e-05,
"loss": 0.9725,
"step": 3940
},
{
"epoch": 0.8130667765869745,
"grad_norm": 0.3292103830025177,
"learning_rate": 2.0551642899279975e-05,
"loss": 1.0363,
"step": 3945
},
{
"epoch": 0.8140972794723825,
"grad_norm": 0.4380964265869986,
"learning_rate": 2.0333668773337866e-05,
"loss": 1.1014,
"step": 3950
},
{
"epoch": 0.8151277823577906,
"grad_norm": 0.3297438808755647,
"learning_rate": 2.0116725856266926e-05,
"loss": 0.8749,
"step": 3955
},
{
"epoch": 0.8161582852431987,
"grad_norm": 0.3404142147430593,
"learning_rate": 1.9900816956197698e-05,
"loss": 1.2154,
"step": 3960
},
{
"epoch": 0.8171887881286067,
"grad_norm": 0.28990266176476354,
"learning_rate": 1.9685944867876373e-05,
"loss": 0.9402,
"step": 3965
},
{
"epoch": 0.8182192910140148,
"grad_norm": 0.3750595800822031,
"learning_rate": 1.9472112372628536e-05,
"loss": 1.0682,
"step": 3970
},
{
"epoch": 0.8192497938994229,
"grad_norm": 0.4415444668426194,
"learning_rate": 1.9259322238323095e-05,
"loss": 1.1201,
"step": 3975
},
{
"epoch": 0.8202802967848309,
"grad_norm": 0.31061384334227865,
"learning_rate": 1.9047577219336665e-05,
"loss": 0.8734,
"step": 3980
},
{
"epoch": 0.8213107996702391,
"grad_norm": 0.3577711991948567,
"learning_rate": 1.8836880056517658e-05,
"loss": 1.1896,
"step": 3985
},
{
"epoch": 0.8223413025556472,
"grad_norm": 0.2862925377080917,
"learning_rate": 1.862723347715103e-05,
"loss": 0.9206,
"step": 3990
},
{
"epoch": 0.8233718054410553,
"grad_norm": 0.36692123404130295,
"learning_rate": 1.841864019492282e-05,
"loss": 1.0504,
"step": 3995
},
{
"epoch": 0.8244023083264633,
"grad_norm": 0.3600663907207253,
"learning_rate": 1.821110290988509e-05,
"loss": 1.2122,
"step": 4000
},
{
"epoch": 0.8244023083264633,
"eval_loss": 0.9671783447265625,
"eval_runtime": 2672.6929,
"eval_samples_per_second": 2.993,
"eval_steps_per_second": 0.187,
"step": 4000
},
{
"epoch": 0.8254328112118714,
"grad_norm": 0.3081888529721894,
"learning_rate": 1.8004624308421026e-05,
"loss": 0.7821,
"step": 4005
},
{
"epoch": 0.8264633140972795,
"grad_norm": 0.36648973755342457,
"learning_rate": 1.7799207063210044e-05,
"loss": 1.2144,
"step": 4010
},
{
"epoch": 0.8274938169826875,
"grad_norm": 0.2898792198270504,
"learning_rate": 1.759485383319326e-05,
"loss": 0.9427,
"step": 4015
},
{
"epoch": 0.8285243198680956,
"grad_norm": 0.38174814126983564,
"learning_rate": 1.7391567263539144e-05,
"loss": 1.0154,
"step": 4020
},
{
"epoch": 0.8295548227535037,
"grad_norm": 0.4742619716693875,
"learning_rate": 1.7189349985609115e-05,
"loss": 1.0813,
"step": 4025
},
{
"epoch": 0.8305853256389117,
"grad_norm": 0.3664614977301218,
"learning_rate": 1.6988204616923666e-05,
"loss": 0.8368,
"step": 4030
},
{
"epoch": 0.8316158285243199,
"grad_norm": 0.43325455353643755,
"learning_rate": 1.6788133761128312e-05,
"loss": 1.1778,
"step": 4035
},
{
"epoch": 0.832646331409728,
"grad_norm": 0.29420614649546306,
"learning_rate": 1.658914000795999e-05,
"loss": 0.922,
"step": 4040
},
{
"epoch": 0.8336768342951361,
"grad_norm": 0.3437287490008131,
"learning_rate": 1.639122593321357e-05,
"loss": 1.0901,
"step": 4045
},
{
"epoch": 0.8347073371805441,
"grad_norm": 0.41394098029113957,
"learning_rate": 1.6194394098708377e-05,
"loss": 1.1766,
"step": 4050
},
{
"epoch": 0.8357378400659522,
"grad_norm": 0.3399407346712763,
"learning_rate": 1.59986470522551e-05,
"loss": 0.8491,
"step": 4055
},
{
"epoch": 0.8367683429513603,
"grad_norm": 0.3655997093950039,
"learning_rate": 1.580398732762297e-05,
"loss": 1.1981,
"step": 4060
},
{
"epoch": 0.8377988458367683,
"grad_norm": 0.30216864477326527,
"learning_rate": 1.5610417444506664e-05,
"loss": 0.9703,
"step": 4065
},
{
"epoch": 0.8388293487221764,
"grad_norm": 0.3993955860213197,
"learning_rate": 1.541793990849387e-05,
"loss": 1.0866,
"step": 4070
},
{
"epoch": 0.8398598516075845,
"grad_norm": 0.40765142147078065,
"learning_rate": 1.522655721103291e-05,
"loss": 1.1578,
"step": 4075
},
{
"epoch": 0.8408903544929925,
"grad_norm": 0.3132479277395683,
"learning_rate": 1.5036271829400294e-05,
"loss": 0.8317,
"step": 4080
},
{
"epoch": 0.8419208573784006,
"grad_norm": 0.353057034757265,
"learning_rate": 1.4847086226668872e-05,
"loss": 1.1387,
"step": 4085
},
{
"epoch": 0.8429513602638088,
"grad_norm": 0.28714788090256477,
"learning_rate": 1.4659002851675774e-05,
"loss": 0.9363,
"step": 4090
},
{
"epoch": 0.8439818631492169,
"grad_norm": 0.3558290938292747,
"learning_rate": 1.447202413899078e-05,
"loss": 0.9974,
"step": 4095
},
{
"epoch": 0.8450123660346249,
"grad_norm": 0.36577045780008416,
"learning_rate": 1.4286152508884898e-05,
"loss": 1.1466,
"step": 4100
},
{
"epoch": 0.846042868920033,
"grad_norm": 0.3010779731099561,
"learning_rate": 1.4101390367298861e-05,
"loss": 0.7895,
"step": 4105
},
{
"epoch": 0.8470733718054411,
"grad_norm": 0.3541525630578856,
"learning_rate": 1.3917740105812094e-05,
"loss": 1.1429,
"step": 4110
},
{
"epoch": 0.8481038746908491,
"grad_norm": 0.28819494987192507,
"learning_rate": 1.3735204101611776e-05,
"loss": 0.9335,
"step": 4115
},
{
"epoch": 0.8491343775762572,
"grad_norm": 0.34823589325876386,
"learning_rate": 1.355378471746196e-05,
"loss": 1.0577,
"step": 4120
},
{
"epoch": 0.8501648804616653,
"grad_norm": 0.34950447449734157,
"learning_rate": 1.3373484301673145e-05,
"loss": 1.1231,
"step": 4125
},
{
"epoch": 0.8511953833470733,
"grad_norm": 0.29975822569477556,
"learning_rate": 1.3194305188071732e-05,
"loss": 0.8424,
"step": 4130
},
{
"epoch": 0.8522258862324814,
"grad_norm": 0.390420435835868,
"learning_rate": 1.301624969596985e-05,
"loss": 1.1329,
"step": 4135
},
{
"epoch": 0.8532563891178895,
"grad_norm": 0.2911183426984063,
"learning_rate": 1.2839320130135468e-05,
"loss": 0.9841,
"step": 4140
},
{
"epoch": 0.8542868920032977,
"grad_norm": 0.3500369252129129,
"learning_rate": 1.266351878076234e-05,
"loss": 1.0513,
"step": 4145
},
{
"epoch": 0.8553173948887057,
"grad_norm": 0.4017098058926308,
"learning_rate": 1.2488847923440483e-05,
"loss": 1.1624,
"step": 4150
},
{
"epoch": 0.8563478977741138,
"grad_norm": 0.3441542578051488,
"learning_rate": 1.2315309819126852e-05,
"loss": 0.8611,
"step": 4155
},
{
"epoch": 0.8573784006595219,
"grad_norm": 0.36811687971995544,
"learning_rate": 1.2142906714115787e-05,
"loss": 1.2298,
"step": 4160
},
{
"epoch": 0.8584089035449299,
"grad_norm": 0.2934040538277352,
"learning_rate": 1.197164084001009e-05,
"loss": 0.9825,
"step": 4165
},
{
"epoch": 0.859439406430338,
"grad_norm": 0.3571442419781514,
"learning_rate": 1.1801514413692239e-05,
"loss": 1.0296,
"step": 4170
},
{
"epoch": 0.8604699093157461,
"grad_norm": 0.4388041060818487,
"learning_rate": 1.1632529637295475e-05,
"loss": 1.1665,
"step": 4175
},
{
"epoch": 0.8615004122011541,
"grad_norm": 0.3211936804672997,
"learning_rate": 1.1464688698175497e-05,
"loss": 0.8568,
"step": 4180
},
{
"epoch": 0.8625309150865622,
"grad_norm": 0.38975409701266794,
"learning_rate": 1.1297993768881998e-05,
"loss": 1.1854,
"step": 4185
},
{
"epoch": 0.8635614179719703,
"grad_norm": 0.27934297465783847,
"learning_rate": 1.113244700713063e-05,
"loss": 0.8697,
"step": 4190
},
{
"epoch": 0.8645919208573783,
"grad_norm": 0.39072549944659307,
"learning_rate": 1.0968050555775067e-05,
"loss": 1.0164,
"step": 4195
},
{
"epoch": 0.8656224237427865,
"grad_norm": 0.4998393287633088,
"learning_rate": 1.0804806542779223e-05,
"loss": 1.0571,
"step": 4200
},
{
"epoch": 0.8656224237427865,
"eval_loss": 0.9649366140365601,
"eval_runtime": 2477.8624,
"eval_samples_per_second": 3.229,
"eval_steps_per_second": 0.202,
"step": 4200
},
{
"epoch": 0.8666529266281946,
"grad_norm": 0.30161833311373715,
"learning_rate": 1.0642717081189735e-05,
"loss": 0.8134,
"step": 4205
},
{
"epoch": 0.8676834295136027,
"grad_norm": 0.3714099835276263,
"learning_rate": 1.0481784269108664e-05,
"loss": 1.2136,
"step": 4210
},
{
"epoch": 0.8687139323990107,
"grad_norm": 0.29858155300200195,
"learning_rate": 1.032201018966621e-05,
"loss": 0.9334,
"step": 4215
},
{
"epoch": 0.8697444352844188,
"grad_norm": 0.36365152527356986,
"learning_rate": 1.0163396910993883e-05,
"loss": 1.0167,
"step": 4220
},
{
"epoch": 0.8707749381698269,
"grad_norm": 0.3926164945656461,
"learning_rate": 1.0005946486197648e-05,
"loss": 1.1597,
"step": 4225
},
{
"epoch": 0.8718054410552349,
"grad_norm": 0.31631364428210557,
"learning_rate": 9.849660953331363e-06,
"loss": 0.8626,
"step": 4230
},
{
"epoch": 0.872835943940643,
"grad_norm": 0.397012339843304,
"learning_rate": 9.694542335370437e-06,
"loss": 1.1559,
"step": 4235
},
{
"epoch": 0.8738664468260511,
"grad_norm": 0.27478860653538173,
"learning_rate": 9.540592640185597e-06,
"loss": 0.9427,
"step": 4240
},
{
"epoch": 0.8748969497114591,
"grad_norm": 0.34552863820949387,
"learning_rate": 9.387813860516915e-06,
"loss": 1.0141,
"step": 4245
},
{
"epoch": 0.8759274525968672,
"grad_norm": 0.4001075016221677,
"learning_rate": 9.236207973948063e-06,
"loss": 1.1693,
"step": 4250
},
{
"epoch": 0.8769579554822754,
"grad_norm": 0.3089362146851165,
"learning_rate": 9.085776942880608e-06,
"loss": 0.8552,
"step": 4255
},
{
"epoch": 0.8779884583676835,
"grad_norm": 0.3341234310512727,
"learning_rate": 8.936522714508678e-06,
"loss": 1.1562,
"step": 4260
},
{
"epoch": 0.8790189612530915,
"grad_norm": 0.2798328581056808,
"learning_rate": 8.788447220793806e-06,
"loss": 0.99,
"step": 4265
},
{
"epoch": 0.8800494641384996,
"grad_norm": 0.39837992045978476,
"learning_rate": 8.641552378439776e-06,
"loss": 1.1073,
"step": 4270
},
{
"epoch": 0.8810799670239077,
"grad_norm": 0.4106276064285754,
"learning_rate": 8.495840088868024e-06,
"loss": 1.1004,
"step": 4275
},
{
"epoch": 0.8821104699093157,
"grad_norm": 0.37910153433606,
"learning_rate": 8.351312238192787e-06,
"loss": 0.9667,
"step": 4280
},
{
"epoch": 0.8831409727947238,
"grad_norm": 0.3835389510278448,
"learning_rate": 8.207970697196866e-06,
"loss": 1.1637,
"step": 4285
},
{
"epoch": 0.8841714756801319,
"grad_norm": 0.2996106330164154,
"learning_rate": 8.065817321307367e-06,
"loss": 0.8998,
"step": 4290
},
{
"epoch": 0.88520197856554,
"grad_norm": 0.37985003315229504,
"learning_rate": 7.924853950571642e-06,
"loss": 0.9981,
"step": 4295
},
{
"epoch": 0.886232481450948,
"grad_norm": 0.3734475653137268,
"learning_rate": 7.78508240963347e-06,
"loss": 1.1901,
"step": 4300
},
{
"epoch": 0.8872629843363561,
"grad_norm": 0.32917519272254847,
"learning_rate": 7.646504507709563e-06,
"loss": 0.8026,
"step": 4305
},
{
"epoch": 0.8882934872217643,
"grad_norm": 0.3184623209600695,
"learning_rate": 7.50912203856593e-06,
"loss": 1.1705,
"step": 4310
},
{
"epoch": 0.8893239901071723,
"grad_norm": 0.27564137851589965,
"learning_rate": 7.372936780494877e-06,
"loss": 0.8968,
"step": 4315
},
{
"epoch": 0.8903544929925804,
"grad_norm": 0.37744723978180933,
"learning_rate": 7.237950496291856e-06,
"loss": 0.9825,
"step": 4320
},
{
"epoch": 0.8913849958779885,
"grad_norm": 0.3497470810206515,
"learning_rate": 7.104164933232649e-06,
"loss": 1.1498,
"step": 4325
},
{
"epoch": 0.8924154987633965,
"grad_norm": 0.314664899690108,
"learning_rate": 6.971581823050832e-06,
"loss": 0.8124,
"step": 4330
},
{
"epoch": 0.8934460016488046,
"grad_norm": 0.3437097611977056,
"learning_rate": 6.840202881915325e-06,
"loss": 1.1555,
"step": 4335
},
{
"epoch": 0.8944765045342127,
"grad_norm": 0.28333067118500105,
"learning_rate": 6.710029810408092e-06,
"loss": 0.899,
"step": 4340
},
{
"epoch": 0.8955070074196207,
"grad_norm": 0.39572557163512656,
"learning_rate": 6.581064293502293e-06,
"loss": 1.0465,
"step": 4345
},
{
"epoch": 0.8965375103050288,
"grad_norm": 0.37753981586149277,
"learning_rate": 6.453308000540304e-06,
"loss": 1.1223,
"step": 4350
},
{
"epoch": 0.8975680131904369,
"grad_norm": 0.33675472826608077,
"learning_rate": 6.326762585212209e-06,
"loss": 0.8248,
"step": 4355
},
{
"epoch": 0.8985985160758451,
"grad_norm": 0.3825197592387351,
"learning_rate": 6.20142968553441e-06,
"loss": 1.1401,
"step": 4360
},
{
"epoch": 0.8996290189612531,
"grad_norm": 0.2838180681849475,
"learning_rate": 6.077310923828328e-06,
"loss": 0.909,
"step": 4365
},
{
"epoch": 0.9006595218466612,
"grad_norm": 0.37686082536565757,
"learning_rate": 5.954407906699511e-06,
"loss": 1.0217,
"step": 4370
},
{
"epoch": 0.9016900247320693,
"grad_norm": 0.3380783448392427,
"learning_rate": 5.8327222250167735e-06,
"loss": 1.1523,
"step": 4375
},
{
"epoch": 0.9027205276174773,
"grad_norm": 0.2883596670864561,
"learning_rate": 5.71225545389158e-06,
"loss": 0.8144,
"step": 4380
},
{
"epoch": 0.9037510305028854,
"grad_norm": 0.3483906272756692,
"learning_rate": 5.59300915265778e-06,
"loss": 1.1855,
"step": 4385
},
{
"epoch": 0.9047815333882935,
"grad_norm": 0.2639750686093266,
"learning_rate": 5.4749848648512624e-06,
"loss": 0.9381,
"step": 4390
},
{
"epoch": 0.9058120362737015,
"grad_norm": 0.3972431149683815,
"learning_rate": 5.358184118190068e-06,
"loss": 1.0683,
"step": 4395
},
{
"epoch": 0.9068425391591096,
"grad_norm": 0.35801168305170344,
"learning_rate": 5.242608424554651e-06,
"loss": 1.0903,
"step": 4400
},
{
"epoch": 0.9068425391591096,
"eval_loss": 0.9635400176048279,
"eval_runtime": 2476.6467,
"eval_samples_per_second": 3.23,
"eval_steps_per_second": 0.202,
"step": 4400
},
{
"epoch": 0.9078730420445177,
"grad_norm": 0.33090712708760794,
"learning_rate": 5.128259279968195e-06,
"loss": 0.8644,
"step": 4405
},
{
"epoch": 0.9089035449299258,
"grad_norm": 0.35066225756340424,
"learning_rate": 5.01513816457736e-06,
"loss": 1.1482,
"step": 4410
},
{
"epoch": 0.9099340478153339,
"grad_norm": 0.2777858386287618,
"learning_rate": 4.903246542633033e-06,
"loss": 0.9529,
"step": 4415
},
{
"epoch": 0.910964550700742,
"grad_norm": 0.3807172458090473,
"learning_rate": 4.7925858624714215e-06,
"loss": 1.0499,
"step": 4420
},
{
"epoch": 0.9119950535861501,
"grad_norm": 0.36543715138130545,
"learning_rate": 4.683157556495343e-06,
"loss": 1.1883,
"step": 4425
},
{
"epoch": 0.9130255564715581,
"grad_norm": 0.3339553261899812,
"learning_rate": 4.574963041155622e-06,
"loss": 0.8319,
"step": 4430
},
{
"epoch": 0.9140560593569662,
"grad_norm": 0.3202714496531148,
"learning_rate": 4.468003716932734e-06,
"loss": 1.1678,
"step": 4435
},
{
"epoch": 0.9150865622423743,
"grad_norm": 0.27715040842554794,
"learning_rate": 4.362280968318777e-06,
"loss": 0.9364,
"step": 4440
},
{
"epoch": 0.9161170651277823,
"grad_norm": 0.3628969851723974,
"learning_rate": 4.257796163799455e-06,
"loss": 1.0704,
"step": 4445
},
{
"epoch": 0.9171475680131904,
"grad_norm": 0.4255339930477384,
"learning_rate": 4.154550655836409e-06,
"loss": 1.1622,
"step": 4450
},
{
"epoch": 0.9181780708985985,
"grad_norm": 0.3098446765156209,
"learning_rate": 4.052545780849715e-06,
"loss": 0.8363,
"step": 4455
},
{
"epoch": 0.9192085737840066,
"grad_norm": 0.38626454996818416,
"learning_rate": 3.9517828592005475e-06,
"loss": 1.1148,
"step": 4460
},
{
"epoch": 0.9202390766694146,
"grad_norm": 0.2786681923266397,
"learning_rate": 3.852263195174155e-06,
"loss": 0.8743,
"step": 4465
},
{
"epoch": 0.9212695795548228,
"grad_norm": 0.3627972033581097,
"learning_rate": 3.7539880769628998e-06,
"loss": 1.0305,
"step": 4470
},
{
"epoch": 0.9223000824402309,
"grad_norm": 0.3545740104488335,
"learning_rate": 3.6569587766496216e-06,
"loss": 1.1188,
"step": 4475
},
{
"epoch": 0.9233305853256389,
"grad_norm": 0.30405956916616017,
"learning_rate": 3.561176550191203e-06,
"loss": 0.7902,
"step": 4480
},
{
"epoch": 0.924361088211047,
"grad_norm": 0.40805109311840626,
"learning_rate": 3.46664263740224e-06,
"loss": 1.0949,
"step": 4485
},
{
"epoch": 0.9253915910964551,
"grad_norm": 0.2937050314911077,
"learning_rate": 3.3733582619390523e-06,
"loss": 0.9339,
"step": 4490
},
{
"epoch": 0.9264220939818631,
"grad_norm": 0.37795976499150646,
"learning_rate": 3.281324631283833e-06,
"loss": 1.0108,
"step": 4495
},
{
"epoch": 0.9274525968672712,
"grad_norm": 0.4168578135164906,
"learning_rate": 3.1905429367289795e-06,
"loss": 1.1267,
"step": 4500
},
{
"epoch": 0.9284830997526793,
"grad_norm": 0.30324094735934326,
"learning_rate": 3.101014353361753e-06,
"loss": 0.8109,
"step": 4505
},
{
"epoch": 0.9295136026380874,
"grad_norm": 0.3929210483289662,
"learning_rate": 3.012740040048978e-06,
"loss": 1.1848,
"step": 4510
},
{
"epoch": 0.9305441055234954,
"grad_norm": 0.2889829253469363,
"learning_rate": 2.9257211394220773e-06,
"loss": 0.8944,
"step": 4515
},
{
"epoch": 0.9315746084089035,
"grad_norm": 0.39922043302261206,
"learning_rate": 2.8399587778623505e-06,
"loss": 1.0599,
"step": 4520
},
{
"epoch": 0.9326051112943117,
"grad_norm": 0.41401354380939953,
"learning_rate": 2.755454065486263e-06,
"loss": 1.1731,
"step": 4525
},
{
"epoch": 0.9336356141797197,
"grad_norm": 0.3816124740915107,
"learning_rate": 2.672208096131157e-06,
"loss": 0.9285,
"step": 4530
},
{
"epoch": 0.9346661170651278,
"grad_norm": 0.365839138684449,
"learning_rate": 2.5902219473411204e-06,
"loss": 1.1613,
"step": 4535
},
{
"epoch": 0.9356966199505359,
"grad_norm": 0.30150334355809666,
"learning_rate": 2.509496680352963e-06,
"loss": 0.9157,
"step": 4540
},
{
"epoch": 0.936727122835944,
"grad_norm": 0.36964220055921937,
"learning_rate": 2.430033340082516e-06,
"loss": 1.0569,
"step": 4545
},
{
"epoch": 0.937757625721352,
"grad_norm": 0.41058988578726896,
"learning_rate": 2.3518329551111217e-06,
"loss": 1.1936,
"step": 4550
},
{
"epoch": 0.9387881286067601,
"grad_norm": 0.32143724004908797,
"learning_rate": 2.2748965376723e-06,
"loss": 0.8617,
"step": 4555
},
{
"epoch": 0.9398186314921682,
"grad_norm": 0.3695703528509068,
"learning_rate": 2.199225083638656e-06,
"loss": 1.2086,
"step": 4560
},
{
"epoch": 0.9408491343775762,
"grad_norm": 0.28544602650039125,
"learning_rate": 2.1248195725089624e-06,
"loss": 0.9673,
"step": 4565
},
{
"epoch": 0.9418796372629843,
"grad_norm": 0.33475042033286156,
"learning_rate": 2.0516809673955083e-06,
"loss": 1.0888,
"step": 4570
},
{
"epoch": 0.9429101401483924,
"grad_norm": 0.40597870855708795,
"learning_rate": 1.9798102150116573e-06,
"loss": 1.1073,
"step": 4575
},
{
"epoch": 0.9439406430338005,
"grad_norm": 0.35068928153409934,
"learning_rate": 1.909208245659522e-06,
"loss": 0.8919,
"step": 4580
},
{
"epoch": 0.9449711459192086,
"grad_norm": 0.3520873665570541,
"learning_rate": 1.8398759732179637e-06,
"loss": 1.1762,
"step": 4585
},
{
"epoch": 0.9460016488046167,
"grad_norm": 0.31075650924614495,
"learning_rate": 1.7718142951307914e-06,
"loss": 0.8997,
"step": 4590
},
{
"epoch": 0.9470321516900247,
"grad_norm": 0.3520700228079484,
"learning_rate": 1.705024092395091e-06,
"loss": 1.0955,
"step": 4595
},
{
"epoch": 0.9480626545754328,
"grad_norm": 0.3680157698495691,
"learning_rate": 1.6395062295498698e-06,
"loss": 1.178,
"step": 4600
},
{
"epoch": 0.9480626545754328,
"eval_loss": 0.9628809690475464,
"eval_runtime": 2669.1038,
"eval_samples_per_second": 2.997,
"eval_steps_per_second": 0.187,
"step": 4600
},
{
"epoch": 0.9490931574608409,
"grad_norm": 0.31126006459380184,
"learning_rate": 1.5752615546647975e-06,
"loss": 0.8694,
"step": 4605
},
{
"epoch": 0.950123660346249,
"grad_norm": 0.34126438848006335,
"learning_rate": 1.5122908993293273e-06,
"loss": 1.1496,
"step": 4610
},
{
"epoch": 0.951154163231657,
"grad_norm": 0.28191122396530277,
"learning_rate": 1.4505950786418255e-06,
"loss": 0.9136,
"step": 4615
},
{
"epoch": 0.9521846661170651,
"grad_norm": 0.3757522319369373,
"learning_rate": 1.3901748911991253e-06,
"loss": 1.0589,
"step": 4620
},
{
"epoch": 0.9532151690024732,
"grad_norm": 0.3812139506827204,
"learning_rate": 1.331031119086079e-06,
"loss": 1.1577,
"step": 4625
},
{
"epoch": 0.9542456718878813,
"grad_norm": 0.3285095663923165,
"learning_rate": 1.2731645278655445e-06,
"loss": 0.9256,
"step": 4630
},
{
"epoch": 0.9552761747732894,
"grad_norm": 0.3434798919525598,
"learning_rate": 1.2165758665683924e-06,
"loss": 1.1373,
"step": 4635
},
{
"epoch": 0.9563066776586975,
"grad_norm": 0.3017676114908958,
"learning_rate": 1.1612658676838473e-06,
"loss": 0.8882,
"step": 4640
},
{
"epoch": 0.9573371805441055,
"grad_norm": 0.3795284009217967,
"learning_rate": 1.107235247150018e-06,
"loss": 1.0639,
"step": 4645
},
{
"epoch": 0.9583676834295136,
"grad_norm": 0.4845153985581749,
"learning_rate": 1.0544847043445938e-06,
"loss": 1.0753,
"step": 4650
},
{
"epoch": 0.9593981863149217,
"grad_norm": 0.32363946255340464,
"learning_rate": 1.0030149220758288e-06,
"loss": 0.8307,
"step": 4655
},
{
"epoch": 0.9604286892003298,
"grad_norm": 0.4184299714975136,
"learning_rate": 9.528265665736502e-07,
"loss": 1.1371,
"step": 4660
},
{
"epoch": 0.9614591920857378,
"grad_norm": 0.28512105148343453,
"learning_rate": 9.039202874811192e-07,
"loss": 0.9634,
"step": 4665
},
{
"epoch": 0.9624896949711459,
"grad_norm": 0.360971597617964,
"learning_rate": 8.562967178459391e-07,
"loss": 1.0805,
"step": 4670
},
{
"epoch": 0.963520197856554,
"grad_norm": 0.3822471953295701,
"learning_rate": 8.099564741123166e-07,
"loss": 1.2147,
"step": 4675
},
{
"epoch": 0.964550700741962,
"grad_norm": 0.32385746632379775,
"learning_rate": 7.649001561129354e-07,
"loss": 0.8677,
"step": 4680
},
{
"epoch": 0.9655812036273702,
"grad_norm": 0.33325536515100207,
"learning_rate": 7.211283470612395e-07,
"loss": 1.1736,
"step": 4685
},
{
"epoch": 0.9666117065127783,
"grad_norm": 0.2789936033919106,
"learning_rate": 6.786416135438512e-07,
"loss": 0.9563,
"step": 4690
},
{
"epoch": 0.9676422093981863,
"grad_norm": 0.3347741972532808,
"learning_rate": 6.374405055132537e-07,
"loss": 0.9439,
"step": 4695
},
{
"epoch": 0.9686727122835944,
"grad_norm": 0.3903192561742831,
"learning_rate": 5.975255562806647e-07,
"loss": 1.1797,
"step": 4700
},
{
"epoch": 0.9697032151690025,
"grad_norm": 0.32771315967905756,
"learning_rate": 5.58897282509141e-07,
"loss": 0.8953,
"step": 4705
},
{
"epoch": 0.9707337180544106,
"grad_norm": 0.34884009544181055,
"learning_rate": 5.215561842068728e-07,
"loss": 1.1761,
"step": 4710
},
{
"epoch": 0.9717642209398186,
"grad_norm": 0.2775010942276059,
"learning_rate": 4.855027447207338e-07,
"loss": 0.9457,
"step": 4715
},
{
"epoch": 0.9727947238252267,
"grad_norm": 0.382431683738394,
"learning_rate": 4.507374307299972e-07,
"loss": 1.0359,
"step": 4720
},
{
"epoch": 0.9738252267106348,
"grad_norm": 0.4253814777599987,
"learning_rate": 4.172606922403399e-07,
"loss": 1.1746,
"step": 4725
},
{
"epoch": 0.9748557295960428,
"grad_norm": 0.3119359666789732,
"learning_rate": 3.8507296257798145e-07,
"loss": 0.8048,
"step": 4730
},
{
"epoch": 0.9758862324814509,
"grad_norm": 0.35641657318401637,
"learning_rate": 3.541746583840655e-07,
"loss": 1.1508,
"step": 4735
},
{
"epoch": 0.9769167353668591,
"grad_norm": 0.28277759207142494,
"learning_rate": 3.24566179609298e-07,
"loss": 0.9406,
"step": 4740
},
{
"epoch": 0.9779472382522671,
"grad_norm": 0.3649978639842664,
"learning_rate": 2.9624790950875113e-07,
"loss": 1.0074,
"step": 4745
},
{
"epoch": 0.9789777411376752,
"grad_norm": 0.43930720949024876,
"learning_rate": 2.692202146369338e-07,
"loss": 1.1133,
"step": 4750
},
{
"epoch": 0.9800082440230833,
"grad_norm": 0.3568172083915493,
"learning_rate": 2.434834448429957e-07,
"loss": 0.8548,
"step": 4755
},
{
"epoch": 0.9810387469084914,
"grad_norm": 0.3512721884204812,
"learning_rate": 2.1903793326621957e-07,
"loss": 1.1658,
"step": 4760
},
{
"epoch": 0.9820692497938994,
"grad_norm": 0.28260427199429944,
"learning_rate": 1.9588399633173605e-07,
"loss": 0.8872,
"step": 4765
},
{
"epoch": 0.9830997526793075,
"grad_norm": 0.3342814004364708,
"learning_rate": 1.740219337463822e-07,
"loss": 1.0368,
"step": 4770
},
{
"epoch": 0.9841302555647156,
"grad_norm": 0.39410468491852235,
"learning_rate": 1.534520284948715e-07,
"loss": 1.1301,
"step": 4775
},
{
"epoch": 0.9851607584501236,
"grad_norm": 0.32300550472176315,
"learning_rate": 1.3417454683608554e-07,
"loss": 0.8525,
"step": 4780
},
{
"epoch": 0.9861912613355317,
"grad_norm": 0.38783208707768235,
"learning_rate": 1.1618973829966572e-07,
"loss": 1.2139,
"step": 4785
},
{
"epoch": 0.9872217642209398,
"grad_norm": 0.27657069522605154,
"learning_rate": 9.949783568272697e-08,
"loss": 0.9094,
"step": 4790
},
{
"epoch": 0.988252267106348,
"grad_norm": 0.37764224067778107,
"learning_rate": 8.409905504693782e-08,
"loss": 1.0543,
"step": 4795
},
{
"epoch": 0.989282769991756,
"grad_norm": 0.5230932784753239,
"learning_rate": 6.999359571561171e-08,
"loss": 1.1661,
"step": 4800
},
{
"epoch": 0.989282769991756,
"eval_loss": 0.9627586603164673,
"eval_runtime": 2560.4353,
"eval_samples_per_second": 3.124,
"eval_steps_per_second": 0.195,
"step": 4800
},
{
"epoch": 0.9903132728771641,
"grad_norm": 0.2751657982671328,
"learning_rate": 5.718164027121997e-08,
"loss": 0.7531,
"step": 4805
},
{
"epoch": 0.9913437757625722,
"grad_norm": 0.3764599692340869,
"learning_rate": 4.566335455299387e-08,
"loss": 1.1338,
"step": 4810
},
{
"epoch": 0.9923742786479802,
"grad_norm": 0.2992197664538524,
"learning_rate": 3.5438887654737355e-08,
"loss": 0.8897,
"step": 4815
},
{
"epoch": 0.9934047815333883,
"grad_norm": 0.3535257610205098,
"learning_rate": 2.6508371922984166e-08,
"loss": 1.0454,
"step": 4820
},
{
"epoch": 0.9944352844187964,
"grad_norm": 0.37026464651118146,
"learning_rate": 1.887192295521034e-08,
"loss": 1.1346,
"step": 4825
},
{
"epoch": 0.9954657873042044,
"grad_norm": 0.3271412553655858,
"learning_rate": 1.252963959834652e-08,
"loss": 0.8182,
"step": 4830
},
{
"epoch": 0.9964962901896125,
"grad_norm": 0.4040096861009905,
"learning_rate": 7.481603947556703e-09,
"loss": 1.0549,
"step": 4835
},
{
"epoch": 0.9975267930750206,
"grad_norm": 0.28077083214803567,
"learning_rate": 3.727881345105821e-09,
"loss": 0.918,
"step": 4840
},
{
"epoch": 0.9985572959604286,
"grad_norm": 0.38213498100295196,
"learning_rate": 1.2685203795492762e-09,
"loss": 1.0451,
"step": 4845
},
{
"epoch": 0.9995877988458368,
"grad_norm": 0.4052030227111039,
"learning_rate": 1.0355288510011107e-10,
"loss": 1.1878,
"step": 4850
},
{
"epoch": 1.0,
"step": 4852,
"total_flos": 3.101569705521971e+16,
"train_loss": 1.1041986294158312,
"train_runtime": 123226.4419,
"train_samples_per_second": 0.63,
"train_steps_per_second": 0.039
}
],
"logging_steps": 5,
"max_steps": 4852,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.101569705521971e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}