{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 200, "global_step": 4852, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00020610057708161583, "grad_norm": 0.7938007287198758, "learning_rate": 4.1152263374485604e-07, "loss": 1.2403, "step": 1 }, { "epoch": 0.001030502885408079, "grad_norm": 0.9829139995457005, "learning_rate": 2.05761316872428e-06, "loss": 1.5592, "step": 5 }, { "epoch": 0.002061005770816158, "grad_norm": 1.3104221596563024, "learning_rate": 4.11522633744856e-06, "loss": 1.9165, "step": 10 }, { "epoch": 0.0030915086562242375, "grad_norm": 0.8263152262533461, "learning_rate": 6.172839506172839e-06, "loss": 1.6148, "step": 15 }, { "epoch": 0.004122011541632316, "grad_norm": 1.2188669223330737, "learning_rate": 8.23045267489712e-06, "loss": 1.7383, "step": 20 }, { "epoch": 0.005152514427040396, "grad_norm": 1.3669452124425536, "learning_rate": 1.02880658436214e-05, "loss": 1.8777, "step": 25 }, { "epoch": 0.006183017312448475, "grad_norm": 0.6877557296913395, "learning_rate": 1.2345679012345678e-05, "loss": 1.3855, "step": 30 }, { "epoch": 0.007213520197856554, "grad_norm": 0.6292369266751937, "learning_rate": 1.440329218106996e-05, "loss": 1.7622, "step": 35 }, { "epoch": 0.008244023083264633, "grad_norm": 0.43117531258376596, "learning_rate": 1.646090534979424e-05, "loss": 1.4525, "step": 40 }, { "epoch": 0.009274525968672712, "grad_norm": 0.3651450490233128, "learning_rate": 1.8518518518518518e-05, "loss": 1.546, "step": 45 }, { "epoch": 0.010305028854080791, "grad_norm": 0.4844380918063832, "learning_rate": 2.05761316872428e-05, "loss": 1.6499, "step": 50 }, { "epoch": 0.01133553173948887, "grad_norm": 0.33393658192356523, "learning_rate": 2.2633744855967078e-05, "loss": 1.1817, "step": 55 }, { "epoch": 0.01236603462489695, "grad_norm": 0.34761501749088797, "learning_rate": 2.4691358024691357e-05, "loss": 1.5655, "step": 60 }, { "epoch": 0.01339653751030503, "grad_norm": 0.3221134242168732, "learning_rate": 2.6748971193415638e-05, "loss": 1.2608, "step": 65 }, { "epoch": 0.014427040395713108, "grad_norm": 0.31264722399213424, "learning_rate": 2.880658436213992e-05, "loss": 1.4104, "step": 70 }, { "epoch": 0.015457543281121188, "grad_norm": 0.3955664926406516, "learning_rate": 3.08641975308642e-05, "loss": 1.4622, "step": 75 }, { "epoch": 0.016488046166529265, "grad_norm": 0.3149436282257214, "learning_rate": 3.292181069958848e-05, "loss": 1.159, "step": 80 }, { "epoch": 0.017518549051937345, "grad_norm": 0.2854140231011414, "learning_rate": 3.497942386831276e-05, "loss": 1.4829, "step": 85 }, { "epoch": 0.018549051937345424, "grad_norm": 0.3326334208099821, "learning_rate": 3.7037037037037037e-05, "loss": 1.1516, "step": 90 }, { "epoch": 0.019579554822753503, "grad_norm": 0.27192427116328566, "learning_rate": 3.909465020576132e-05, "loss": 1.3785, "step": 95 }, { "epoch": 0.020610057708161583, "grad_norm": 0.3152208635246729, "learning_rate": 4.11522633744856e-05, "loss": 1.3886, "step": 100 }, { "epoch": 0.021640560593569662, "grad_norm": 0.27663290557612785, "learning_rate": 4.3209876543209875e-05, "loss": 1.1185, "step": 105 }, { "epoch": 0.02267106347897774, "grad_norm": 0.27185356829855356, "learning_rate": 4.5267489711934157e-05, "loss": 1.4773, "step": 110 }, { "epoch": 0.02370156636438582, "grad_norm": 0.2935081378811297, "learning_rate": 4.732510288065844e-05, "loss": 1.261, "step": 115 }, { "epoch": 0.0247320692497939, "grad_norm": 0.29236576081560334, "learning_rate": 4.938271604938271e-05, "loss": 1.3387, "step": 120 }, { "epoch": 0.02576257213520198, "grad_norm": 0.38637732181078555, "learning_rate": 5.1440329218106995e-05, "loss": 1.4305, "step": 125 }, { "epoch": 0.02679307502061006, "grad_norm": 0.2982040676995298, "learning_rate": 5.3497942386831277e-05, "loss": 1.0162, "step": 130 }, { "epoch": 0.027823577906018138, "grad_norm": 0.2699066491657892, "learning_rate": 5.555555555555556e-05, "loss": 1.4066, "step": 135 }, { "epoch": 0.028854080791426217, "grad_norm": 0.33203812944453787, "learning_rate": 5.761316872427984e-05, "loss": 1.1751, "step": 140 }, { "epoch": 0.029884583676834296, "grad_norm": 0.3312647784314266, "learning_rate": 5.9670781893004115e-05, "loss": 1.2269, "step": 145 }, { "epoch": 0.030915086562242376, "grad_norm": 0.541956195146165, "learning_rate": 6.17283950617284e-05, "loss": 1.354, "step": 150 }, { "epoch": 0.031945589447650455, "grad_norm": 0.3172355233573687, "learning_rate": 6.378600823045267e-05, "loss": 1.0666, "step": 155 }, { "epoch": 0.03297609233305853, "grad_norm": 0.32272337262102746, "learning_rate": 6.584362139917696e-05, "loss": 1.4928, "step": 160 }, { "epoch": 0.03400659521846661, "grad_norm": 0.4252299718796606, "learning_rate": 6.790123456790123e-05, "loss": 1.1979, "step": 165 }, { "epoch": 0.03503709810387469, "grad_norm": 0.2816942875302398, "learning_rate": 6.995884773662552e-05, "loss": 1.3223, "step": 170 }, { "epoch": 0.03606760098928277, "grad_norm": 0.34311165874954624, "learning_rate": 7.20164609053498e-05, "loss": 1.4296, "step": 175 }, { "epoch": 0.03709810387469085, "grad_norm": 0.297834971078222, "learning_rate": 7.407407407407407e-05, "loss": 1.087, "step": 180 }, { "epoch": 0.03812860676009893, "grad_norm": 0.2602802208757066, "learning_rate": 7.613168724279836e-05, "loss": 1.4754, "step": 185 }, { "epoch": 0.039159109645507006, "grad_norm": 0.40680201193562965, "learning_rate": 7.818930041152264e-05, "loss": 1.1889, "step": 190 }, { "epoch": 0.04018961253091509, "grad_norm": 0.45086574163739, "learning_rate": 8.024691358024692e-05, "loss": 1.2851, "step": 195 }, { "epoch": 0.041220115416323165, "grad_norm": 0.4308650820975275, "learning_rate": 8.23045267489712e-05, "loss": 1.3478, "step": 200 }, { "epoch": 0.041220115416323165, "eval_loss": 1.2309638261795044, "eval_runtime": 2469.2141, "eval_samples_per_second": 3.24, "eval_steps_per_second": 0.202, "step": 200 }, { "epoch": 0.04225061830173125, "grad_norm": 0.37845098126001203, "learning_rate": 8.436213991769549e-05, "loss": 1.0707, "step": 205 }, { "epoch": 0.043281121187139324, "grad_norm": 0.2851215824173796, "learning_rate": 8.641975308641975e-05, "loss": 1.4131, "step": 210 }, { "epoch": 0.044311624072547406, "grad_norm": 0.3309919261494515, "learning_rate": 8.847736625514404e-05, "loss": 1.1951, "step": 215 }, { "epoch": 0.04534212695795548, "grad_norm": 0.3065911094811643, "learning_rate": 9.053497942386831e-05, "loss": 1.2554, "step": 220 }, { "epoch": 0.04637262984336356, "grad_norm": 0.32586015367310234, "learning_rate": 9.25925925925926e-05, "loss": 1.4263, "step": 225 }, { "epoch": 0.04740313272877164, "grad_norm": 0.277075233563208, "learning_rate": 9.465020576131688e-05, "loss": 1.0216, "step": 230 }, { "epoch": 0.04843363561417972, "grad_norm": 0.2769072933842902, "learning_rate": 9.670781893004116e-05, "loss": 1.3689, "step": 235 }, { "epoch": 0.0494641384995878, "grad_norm": 0.27922329154610587, "learning_rate": 9.876543209876543e-05, "loss": 1.2181, "step": 240 }, { "epoch": 0.050494641384995875, "grad_norm": 0.30982282181292814, "learning_rate": 0.00010082304526748971, "loss": 1.2183, "step": 245 }, { "epoch": 0.05152514427040396, "grad_norm": 0.33257211954241994, "learning_rate": 0.00010288065843621399, "loss": 1.4305, "step": 250 }, { "epoch": 0.052555647155812034, "grad_norm": 0.2838288504939334, "learning_rate": 0.00010493827160493828, "loss": 1.0051, "step": 255 }, { "epoch": 0.05358615004122012, "grad_norm": 0.28623162042613864, "learning_rate": 0.00010699588477366255, "loss": 1.436, "step": 260 }, { "epoch": 0.05461665292662819, "grad_norm": 0.2720219498268959, "learning_rate": 0.00010905349794238684, "loss": 1.1381, "step": 265 }, { "epoch": 0.055647155812036275, "grad_norm": 0.30404224920143047, "learning_rate": 0.00011111111111111112, "loss": 1.1905, "step": 270 }, { "epoch": 0.05667765869744435, "grad_norm": 0.3469618797327675, "learning_rate": 0.0001131687242798354, "loss": 1.3797, "step": 275 }, { "epoch": 0.057708161582852434, "grad_norm": 0.30848130199190793, "learning_rate": 0.00011522633744855968, "loss": 1.0163, "step": 280 }, { "epoch": 0.05873866446826051, "grad_norm": 0.2587435053896794, "learning_rate": 0.00011728395061728397, "loss": 1.401, "step": 285 }, { "epoch": 0.05976916735366859, "grad_norm": 0.30665664697045764, "learning_rate": 0.00011934156378600823, "loss": 1.197, "step": 290 }, { "epoch": 0.06079967023907667, "grad_norm": 0.28688303187237163, "learning_rate": 0.0001213991769547325, "loss": 1.2817, "step": 295 }, { "epoch": 0.06183017312448475, "grad_norm": 0.32502880249151705, "learning_rate": 0.0001234567901234568, "loss": 1.388, "step": 300 }, { "epoch": 0.06286067600989283, "grad_norm": 0.30699533228385806, "learning_rate": 0.00012551440329218108, "loss": 1.0403, "step": 305 }, { "epoch": 0.06389117889530091, "grad_norm": 0.29902009304688737, "learning_rate": 0.00012757201646090534, "loss": 1.3604, "step": 310 }, { "epoch": 0.06492168178070899, "grad_norm": 0.26224004465269324, "learning_rate": 0.00012962962962962963, "loss": 1.12, "step": 315 }, { "epoch": 0.06595218466611706, "grad_norm": 0.2697856637362638, "learning_rate": 0.00013168724279835392, "loss": 1.295, "step": 320 }, { "epoch": 0.06698268755152514, "grad_norm": 0.3118309921774858, "learning_rate": 0.0001337448559670782, "loss": 1.3498, "step": 325 }, { "epoch": 0.06801319043693323, "grad_norm": 0.3774735054690596, "learning_rate": 0.00013580246913580247, "loss": 1.1412, "step": 330 }, { "epoch": 0.0690436933223413, "grad_norm": 0.27963249445347155, "learning_rate": 0.00013786008230452676, "loss": 1.4175, "step": 335 }, { "epoch": 0.07007419620774938, "grad_norm": 0.25343833022804746, "learning_rate": 0.00013991769547325105, "loss": 1.1228, "step": 340 }, { "epoch": 0.07110469909315746, "grad_norm": 0.2676785839608282, "learning_rate": 0.00014197530864197534, "loss": 1.2235, "step": 345 }, { "epoch": 0.07213520197856554, "grad_norm": 0.33130141183815665, "learning_rate": 0.0001440329218106996, "loss": 1.3623, "step": 350 }, { "epoch": 0.07316570486397361, "grad_norm": 0.2676095204791478, "learning_rate": 0.00014609053497942386, "loss": 1.0355, "step": 355 }, { "epoch": 0.0741962077493817, "grad_norm": 0.31354711887470305, "learning_rate": 0.00014814814814814815, "loss": 1.3388, "step": 360 }, { "epoch": 0.07522671063478978, "grad_norm": 0.26048475689786116, "learning_rate": 0.00015020576131687243, "loss": 1.1111, "step": 365 }, { "epoch": 0.07625721352019786, "grad_norm": 0.3339375055235073, "learning_rate": 0.00015226337448559672, "loss": 1.1734, "step": 370 }, { "epoch": 0.07728771640560593, "grad_norm": 0.327943444067688, "learning_rate": 0.00015432098765432098, "loss": 1.3208, "step": 375 }, { "epoch": 0.07831821929101401, "grad_norm": 0.2830129836407308, "learning_rate": 0.00015637860082304527, "loss": 1.0787, "step": 380 }, { "epoch": 0.0793487221764221, "grad_norm": 0.2751640321887791, "learning_rate": 0.00015843621399176956, "loss": 1.3772, "step": 385 }, { "epoch": 0.08037922506183018, "grad_norm": 0.27545532948015855, "learning_rate": 0.00016049382716049385, "loss": 1.1672, "step": 390 }, { "epoch": 0.08140972794723825, "grad_norm": 0.300224560861019, "learning_rate": 0.0001625514403292181, "loss": 1.3224, "step": 395 }, { "epoch": 0.08244023083264633, "grad_norm": 0.3227759504569782, "learning_rate": 0.0001646090534979424, "loss": 1.3495, "step": 400 }, { "epoch": 0.08244023083264633, "eval_loss": 1.1826071739196777, "eval_runtime": 2476.8142, "eval_samples_per_second": 3.23, "eval_steps_per_second": 0.202, "step": 400 }, { "epoch": 0.08347073371805441, "grad_norm": 0.27853687771921737, "learning_rate": 0.0001666666666666667, "loss": 1.1022, "step": 405 }, { "epoch": 0.0845012366034625, "grad_norm": 0.2685951286639512, "learning_rate": 0.00016872427983539098, "loss": 1.4061, "step": 410 }, { "epoch": 0.08553173948887056, "grad_norm": 0.2608899320776935, "learning_rate": 0.00017078189300411524, "loss": 1.1174, "step": 415 }, { "epoch": 0.08656224237427865, "grad_norm": 0.3397550472709015, "learning_rate": 0.0001728395061728395, "loss": 1.2198, "step": 420 }, { "epoch": 0.08759274525968673, "grad_norm": 0.25033631170694937, "learning_rate": 0.0001748971193415638, "loss": 1.4029, "step": 425 }, { "epoch": 0.08862324814509481, "grad_norm": 0.3049472388523693, "learning_rate": 0.00017695473251028808, "loss": 1.0265, "step": 430 }, { "epoch": 0.08965375103050288, "grad_norm": 0.29598713943285077, "learning_rate": 0.00017901234567901234, "loss": 1.3719, "step": 435 }, { "epoch": 0.09068425391591096, "grad_norm": 0.2656526738754943, "learning_rate": 0.00018106995884773663, "loss": 1.1238, "step": 440 }, { "epoch": 0.09171475680131905, "grad_norm": 0.2546218728984369, "learning_rate": 0.00018312757201646091, "loss": 1.1853, "step": 445 }, { "epoch": 0.09274525968672712, "grad_norm": 0.309485029099833, "learning_rate": 0.0001851851851851852, "loss": 1.3064, "step": 450 }, { "epoch": 0.0937757625721352, "grad_norm": 0.28184488141448555, "learning_rate": 0.00018724279835390946, "loss": 1.0605, "step": 455 }, { "epoch": 0.09480626545754328, "grad_norm": 0.2913075285039461, "learning_rate": 0.00018930041152263375, "loss": 1.4311, "step": 460 }, { "epoch": 0.09583676834295136, "grad_norm": 0.7186836779179142, "learning_rate": 0.00019135802469135804, "loss": 1.1263, "step": 465 }, { "epoch": 0.09686727122835943, "grad_norm": 0.29590695855715343, "learning_rate": 0.00019341563786008233, "loss": 1.2589, "step": 470 }, { "epoch": 0.09789777411376752, "grad_norm": 0.31832697133586063, "learning_rate": 0.0001954732510288066, "loss": 1.395, "step": 475 }, { "epoch": 0.0989282769991756, "grad_norm": 0.23627896192430478, "learning_rate": 0.00019753086419753085, "loss": 1.0703, "step": 480 }, { "epoch": 0.09995877988458368, "grad_norm": 0.27354848483986705, "learning_rate": 0.00019958847736625514, "loss": 1.3205, "step": 485 }, { "epoch": 0.10098928276999175, "grad_norm": 0.4895153568779187, "learning_rate": 0.00019999958578867407, "loss": 1.1488, "step": 490 }, { "epoch": 0.10201978565539983, "grad_norm": 0.2811706808521417, "learning_rate": 0.00019999790306104336, "loss": 1.2442, "step": 495 }, { "epoch": 0.10305028854080792, "grad_norm": 0.2572161259717821, "learning_rate": 0.0001999949259506647, "loss": 1.3355, "step": 500 }, { "epoch": 0.104080791426216, "grad_norm": 0.41643641721819225, "learning_rate": 0.00019999065449607402, "loss": 1.0238, "step": 505 }, { "epoch": 0.10511129431162407, "grad_norm": 0.2871689949893079, "learning_rate": 0.00019998508875256158, "loss": 1.3608, "step": 510 }, { "epoch": 0.10614179719703215, "grad_norm": 0.3219574868373838, "learning_rate": 0.0001999782287921708, "loss": 1.0774, "step": 515 }, { "epoch": 0.10717230008244023, "grad_norm": 0.23385146626580391, "learning_rate": 0.00019997007470369773, "loss": 1.184, "step": 520 }, { "epoch": 0.10820280296784832, "grad_norm": 0.3585994112684759, "learning_rate": 0.0001999606265926897, "loss": 1.3426, "step": 525 }, { "epoch": 0.10923330585325638, "grad_norm": 0.2894131584925935, "learning_rate": 0.000199949884581444, "loss": 1.099, "step": 530 }, { "epoch": 0.11026380873866447, "grad_norm": 0.2824735567473504, "learning_rate": 0.00019993784880900623, "loss": 1.3997, "step": 535 }, { "epoch": 0.11129431162407255, "grad_norm": 0.23322094258169807, "learning_rate": 0.0001999245194311687, "loss": 1.1019, "step": 540 }, { "epoch": 0.11232481450948063, "grad_norm": 0.30614162605617107, "learning_rate": 0.00019990989662046818, "loss": 1.2576, "step": 545 }, { "epoch": 0.1133553173948887, "grad_norm": 0.3389548425122287, "learning_rate": 0.0001998939805661837, "loss": 1.301, "step": 550 }, { "epoch": 0.11438582028029678, "grad_norm": 0.25179179489194975, "learning_rate": 0.00019987677147433432, "loss": 0.9871, "step": 555 }, { "epoch": 0.11541632316570487, "grad_norm": 0.24522921674032325, "learning_rate": 0.0001998582695676762, "loss": 1.4205, "step": 560 }, { "epoch": 0.11644682605111294, "grad_norm": 0.2528881733347297, "learning_rate": 0.00019983847508569987, "loss": 1.1283, "step": 565 }, { "epoch": 0.11747732893652102, "grad_norm": 0.28075131598249065, "learning_rate": 0.00019981738828462703, "loss": 1.2467, "step": 570 }, { "epoch": 0.1185078318219291, "grad_norm": 0.3223992578484773, "learning_rate": 0.00019979500943740735, "loss": 1.3673, "step": 575 }, { "epoch": 0.11953833470733718, "grad_norm": 0.2804260590851507, "learning_rate": 0.00019977133883371478, "loss": 1.1133, "step": 580 }, { "epoch": 0.12056883759274525, "grad_norm": 0.25183674337041767, "learning_rate": 0.00019974637677994404, "loss": 1.3493, "step": 585 }, { "epoch": 0.12159934047815334, "grad_norm": 0.23352143639236672, "learning_rate": 0.00019972012359920638, "loss": 1.1183, "step": 590 }, { "epoch": 0.12262984336356142, "grad_norm": 0.2597060461259452, "learning_rate": 0.0001996925796313256, "loss": 1.2021, "step": 595 }, { "epoch": 0.1236603462489695, "grad_norm": 0.31429525704141625, "learning_rate": 0.00019966374523283347, "loss": 1.3753, "step": 600 }, { "epoch": 0.1236603462489695, "eval_loss": 1.1556562185287476, "eval_runtime": 2465.6561, "eval_samples_per_second": 3.245, "eval_steps_per_second": 0.203, "step": 600 }, { "epoch": 0.12469084913437757, "grad_norm": 0.23970501230526484, "learning_rate": 0.00019963362077696537, "loss": 0.9978, "step": 605 }, { "epoch": 0.12572135201978565, "grad_norm": 0.24126751266896648, "learning_rate": 0.00019960220665365518, "loss": 1.3576, "step": 610 }, { "epoch": 0.12675185490519372, "grad_norm": 0.29806682020411596, "learning_rate": 0.0001995695032695305, "loss": 1.0721, "step": 615 }, { "epoch": 0.12778235779060182, "grad_norm": 0.3121199170110613, "learning_rate": 0.0001995355110479071, "loss": 1.2159, "step": 620 }, { "epoch": 0.1288128606760099, "grad_norm": 0.3534465377015386, "learning_rate": 0.00019950023042878366, "loss": 1.3058, "step": 625 }, { "epoch": 0.12984336356141798, "grad_norm": 0.23203831442914624, "learning_rate": 0.00019946366186883604, "loss": 0.9527, "step": 630 }, { "epoch": 0.13087386644682605, "grad_norm": 0.2720965062058241, "learning_rate": 0.00019942580584141127, "loss": 1.3558, "step": 635 }, { "epoch": 0.13190436933223412, "grad_norm": 0.2629592480398592, "learning_rate": 0.0001993866628365215, "loss": 1.1318, "step": 640 }, { "epoch": 0.13293487221764222, "grad_norm": 0.2920226997136243, "learning_rate": 0.00019934623336083772, "loss": 1.122, "step": 645 }, { "epoch": 0.1339653751030503, "grad_norm": 0.27168430341439026, "learning_rate": 0.00019930451793768298, "loss": 1.3241, "step": 650 }, { "epoch": 0.13499587798845836, "grad_norm": 0.2845184130718667, "learning_rate": 0.00019926151710702588, "loss": 0.9734, "step": 655 }, { "epoch": 0.13602638087386645, "grad_norm": 0.2888242420563682, "learning_rate": 0.00019921723142547347, "loss": 1.3656, "step": 660 }, { "epoch": 0.13705688375927452, "grad_norm": 0.27299177987397777, "learning_rate": 0.00019917166146626392, "loss": 1.0691, "step": 665 }, { "epoch": 0.1380873866446826, "grad_norm": 0.35236952647741737, "learning_rate": 0.0001991248078192593, "loss": 1.2263, "step": 670 }, { "epoch": 0.1391178895300907, "grad_norm": 0.2681417629469303, "learning_rate": 0.00019907667109093794, "loss": 1.354, "step": 675 }, { "epoch": 0.14014839241549876, "grad_norm": 0.310226199539697, "learning_rate": 0.00019902725190438627, "loss": 1.0365, "step": 680 }, { "epoch": 0.14117889530090685, "grad_norm": 0.3036115548984946, "learning_rate": 0.00019897655089929126, "loss": 1.3279, "step": 685 }, { "epoch": 0.14220939818631492, "grad_norm": 0.24182676057547745, "learning_rate": 0.00019892456873193165, "loss": 1.0803, "step": 690 }, { "epoch": 0.143239901071723, "grad_norm": 0.27410693799977515, "learning_rate": 0.00019887130607516978, "loss": 1.2236, "step": 695 }, { "epoch": 0.1442704039571311, "grad_norm": 0.3158493448762225, "learning_rate": 0.00019881676361844275, "loss": 1.3796, "step": 700 }, { "epoch": 0.14530090684253916, "grad_norm": 0.48562045991681607, "learning_rate": 0.0001987609420677535, "loss": 1.0014, "step": 705 }, { "epoch": 0.14633140972794723, "grad_norm": 0.24730942495184344, "learning_rate": 0.00019870384214566174, "loss": 1.3463, "step": 710 }, { "epoch": 0.14736191261335532, "grad_norm": 0.3153226513268582, "learning_rate": 0.00019864546459127448, "loss": 1.0977, "step": 715 }, { "epoch": 0.1483924154987634, "grad_norm": 0.249510122134111, "learning_rate": 0.0001985858101602366, "loss": 1.2687, "step": 720 }, { "epoch": 0.1494229183841715, "grad_norm": 0.30842159531538627, "learning_rate": 0.0001985248796247209, "loss": 1.2815, "step": 725 }, { "epoch": 0.15045342126957956, "grad_norm": 0.26379223642130784, "learning_rate": 0.00019846267377341827, "loss": 0.9403, "step": 730 }, { "epoch": 0.15148392415498763, "grad_norm": 0.23891641481355658, "learning_rate": 0.00019839919341152742, "loss": 1.3505, "step": 735 }, { "epoch": 0.15251442704039572, "grad_norm": 0.3238473098861904, "learning_rate": 0.00019833443936074442, "loss": 1.0409, "step": 740 }, { "epoch": 0.1535449299258038, "grad_norm": 0.2746647114109431, "learning_rate": 0.00019826841245925212, "loss": 1.0984, "step": 745 }, { "epoch": 0.15457543281121186, "grad_norm": 0.37674302298987083, "learning_rate": 0.00019820111356170923, "loss": 1.3147, "step": 750 }, { "epoch": 0.15560593569661996, "grad_norm": 0.25219340224453657, "learning_rate": 0.00019813254353923937, "loss": 1.0348, "step": 755 }, { "epoch": 0.15663643858202803, "grad_norm": 0.26735510810071406, "learning_rate": 0.00019806270327941971, "loss": 1.3468, "step": 760 }, { "epoch": 0.15766694146743612, "grad_norm": 0.25452976189367954, "learning_rate": 0.00019799159368626945, "loss": 1.0699, "step": 765 }, { "epoch": 0.1586974443528442, "grad_norm": 0.2863967455600455, "learning_rate": 0.00019791921568023822, "loss": 1.1045, "step": 770 }, { "epoch": 0.15972794723825226, "grad_norm": 0.32564915808035255, "learning_rate": 0.00019784557019819404, "loss": 1.301, "step": 775 }, { "epoch": 0.16075845012366036, "grad_norm": 0.2598862519671672, "learning_rate": 0.00019777065819341137, "loss": 0.9559, "step": 780 }, { "epoch": 0.16178895300906843, "grad_norm": 0.2694147600296894, "learning_rate": 0.00019769448063555856, "loss": 1.346, "step": 785 }, { "epoch": 0.1628194558944765, "grad_norm": 0.2464767798391986, "learning_rate": 0.00019761703851068553, "loss": 1.102, "step": 790 }, { "epoch": 0.1638499587798846, "grad_norm": 0.2753953016773011, "learning_rate": 0.0001975383328212107, "loss": 1.2188, "step": 795 }, { "epoch": 0.16488046166529266, "grad_norm": 0.38049304343681456, "learning_rate": 0.00019745836458590836, "loss": 1.3454, "step": 800 }, { "epoch": 0.16488046166529266, "eval_loss": 1.1297026872634888, "eval_runtime": 2464.1297, "eval_samples_per_second": 3.247, "eval_steps_per_second": 0.203, "step": 800 }, { "epoch": 0.16591096455070073, "grad_norm": 0.25205784909199425, "learning_rate": 0.0001973771348398953, "loss": 0.9905, "step": 805 }, { "epoch": 0.16694146743610883, "grad_norm": 0.2719665322536544, "learning_rate": 0.0001972946446346173, "loss": 1.3772, "step": 810 }, { "epoch": 0.1679719703215169, "grad_norm": 0.23804058554763297, "learning_rate": 0.00019721089503783577, "loss": 1.0834, "step": 815 }, { "epoch": 0.169002473206925, "grad_norm": 0.28963422898407165, "learning_rate": 0.00019712588713361378, "loss": 1.2064, "step": 820 }, { "epoch": 0.17003297609233306, "grad_norm": 0.3442324930908095, "learning_rate": 0.00019703962202230203, "loss": 1.2967, "step": 825 }, { "epoch": 0.17106347897774113, "grad_norm": 0.2547770116672794, "learning_rate": 0.00019695210082052472, "loss": 0.949, "step": 830 }, { "epoch": 0.17209398186314923, "grad_norm": 0.2832243503202021, "learning_rate": 0.00019686332466116487, "loss": 1.2852, "step": 835 }, { "epoch": 0.1731244847485573, "grad_norm": 0.2332582520614932, "learning_rate": 0.0001967732946933499, "loss": 0.979, "step": 840 }, { "epoch": 0.17415498763396536, "grad_norm": 0.29363618940675, "learning_rate": 0.00019668201208243658, "loss": 1.223, "step": 845 }, { "epoch": 0.17518549051937346, "grad_norm": 0.28680318445015207, "learning_rate": 0.0001965894780099961, "loss": 1.3119, "step": 850 }, { "epoch": 0.17621599340478153, "grad_norm": 0.2415531891983295, "learning_rate": 0.00019649569367379867, "loss": 0.9811, "step": 855 }, { "epoch": 0.17724649629018963, "grad_norm": 0.2766629334909396, "learning_rate": 0.00019640066028779794, "loss": 1.2464, "step": 860 }, { "epoch": 0.1782769991755977, "grad_norm": 0.24460494296561755, "learning_rate": 0.00019630437908211548, "loss": 1.0896, "step": 865 }, { "epoch": 0.17930750206100576, "grad_norm": 0.28580998835949145, "learning_rate": 0.00019620685130302478, "loss": 1.1636, "step": 870 }, { "epoch": 0.18033800494641386, "grad_norm": 0.29222255776500833, "learning_rate": 0.00019610807821293503, "loss": 1.3137, "step": 875 }, { "epoch": 0.18136850783182193, "grad_norm": 0.2684207256549492, "learning_rate": 0.00019600806109037485, "loss": 0.9964, "step": 880 }, { "epoch": 0.18239901071723, "grad_norm": 0.2559298498236225, "learning_rate": 0.00019590680122997582, "loss": 1.3378, "step": 885 }, { "epoch": 0.1834295136026381, "grad_norm": 0.2349865929838881, "learning_rate": 0.00019580429994245555, "loss": 1.0436, "step": 890 }, { "epoch": 0.18446001648804616, "grad_norm": 0.29736696922743316, "learning_rate": 0.0001957005585546009, "loss": 1.1778, "step": 895 }, { "epoch": 0.18549051937345423, "grad_norm": 0.34641352119598734, "learning_rate": 0.00019559557840925055, "loss": 1.289, "step": 900 }, { "epoch": 0.18652102225886233, "grad_norm": 0.2589416225643458, "learning_rate": 0.00019548936086527798, "loss": 0.9491, "step": 905 }, { "epoch": 0.1875515251442704, "grad_norm": 0.2778855204059982, "learning_rate": 0.00019538190729757356, "loss": 1.3701, "step": 910 }, { "epoch": 0.1885820280296785, "grad_norm": 0.2433946170800378, "learning_rate": 0.00019527321909702688, "loss": 1.0454, "step": 915 }, { "epoch": 0.18961253091508656, "grad_norm": 0.28491470830869087, "learning_rate": 0.00019516329767050878, "loss": 1.2001, "step": 920 }, { "epoch": 0.19064303380049463, "grad_norm": 0.30637698470530034, "learning_rate": 0.00019505214444085308, "loss": 1.3096, "step": 925 }, { "epoch": 0.19167353668590273, "grad_norm": 0.24736218088176834, "learning_rate": 0.00019493976084683813, "loss": 0.9663, "step": 930 }, { "epoch": 0.1927040395713108, "grad_norm": 0.24983643400010644, "learning_rate": 0.00019482614834316836, "loss": 1.3698, "step": 935 }, { "epoch": 0.19373454245671887, "grad_norm": 0.2420833122771942, "learning_rate": 0.00019471130840045518, "loss": 1.0335, "step": 940 }, { "epoch": 0.19476504534212696, "grad_norm": 0.2573293132603339, "learning_rate": 0.00019459524250519826, "loss": 1.1905, "step": 945 }, { "epoch": 0.19579554822753503, "grad_norm": 0.3710740261808263, "learning_rate": 0.00019447795215976594, "loss": 1.3102, "step": 950 }, { "epoch": 0.19682605111294313, "grad_norm": 0.2569919144815, "learning_rate": 0.0001943594388823761, "loss": 0.8868, "step": 955 }, { "epoch": 0.1978565539983512, "grad_norm": 0.3020196940157676, "learning_rate": 0.00019423970420707627, "loss": 1.3362, "step": 960 }, { "epoch": 0.19888705688375927, "grad_norm": 0.23638629002825132, "learning_rate": 0.00019411874968372402, "loss": 1.0545, "step": 965 }, { "epoch": 0.19991755976916736, "grad_norm": 0.2690903988670187, "learning_rate": 0.00019399657687796658, "loss": 1.1581, "step": 970 }, { "epoch": 0.20094806265457543, "grad_norm": 0.3219945588976219, "learning_rate": 0.00019387318737122092, "loss": 1.2816, "step": 975 }, { "epoch": 0.2019785655399835, "grad_norm": 0.2510582302230213, "learning_rate": 0.0001937485827606529, "loss": 0.9363, "step": 980 }, { "epoch": 0.2030090684253916, "grad_norm": 0.3075547394305651, "learning_rate": 0.00019362276465915702, "loss": 1.2897, "step": 985 }, { "epoch": 0.20403957131079967, "grad_norm": 0.24639351565359302, "learning_rate": 0.0001934957346953352, "loss": 1.0314, "step": 990 }, { "epoch": 0.20507007419620774, "grad_norm": 0.31059587986354126, "learning_rate": 0.00019336749451347586, "loss": 1.2041, "step": 995 }, { "epoch": 0.20610057708161583, "grad_norm": 0.2842021215906897, "learning_rate": 0.0001932380457735326, "loss": 1.2731, "step": 1000 }, { "epoch": 0.20610057708161583, "eval_loss": 1.1071134805679321, "eval_runtime": 2476.6697, "eval_samples_per_second": 3.23, "eval_steps_per_second": 0.202, "step": 1000 }, { "epoch": 0.2071310799670239, "grad_norm": 0.25497821921270564, "learning_rate": 0.00019310739015110267, "loss": 0.9285, "step": 1005 }, { "epoch": 0.208161582852432, "grad_norm": 0.2793650388701531, "learning_rate": 0.00019297552933740547, "loss": 1.3337, "step": 1010 }, { "epoch": 0.20919208573784007, "grad_norm": 0.2516697285323032, "learning_rate": 0.0001928424650392603, "loss": 1.0655, "step": 1015 }, { "epoch": 0.21022258862324814, "grad_norm": 0.30212273915620885, "learning_rate": 0.00019270819897906468, "loss": 1.2153, "step": 1020 }, { "epoch": 0.21125309150865623, "grad_norm": 0.33419010602150057, "learning_rate": 0.00019257273289477174, "loss": 1.3252, "step": 1025 }, { "epoch": 0.2122835943940643, "grad_norm": 0.27421309631436125, "learning_rate": 0.00019243606853986786, "loss": 1.0082, "step": 1030 }, { "epoch": 0.21331409727947237, "grad_norm": 0.34717245381281975, "learning_rate": 0.00019229820768335, "loss": 1.3901, "step": 1035 }, { "epoch": 0.21434460016488047, "grad_norm": 0.23427202649137227, "learning_rate": 0.00019215915210970267, "loss": 0.9838, "step": 1040 }, { "epoch": 0.21537510305028854, "grad_norm": 0.2923332885069089, "learning_rate": 0.00019201890361887506, "loss": 1.1746, "step": 1045 }, { "epoch": 0.21640560593569663, "grad_norm": 0.34553332711604556, "learning_rate": 0.0001918774640262574, "loss": 1.2803, "step": 1050 }, { "epoch": 0.2174361088211047, "grad_norm": 0.282937796924011, "learning_rate": 0.00019173483516265788, "loss": 1.0297, "step": 1055 }, { "epoch": 0.21846661170651277, "grad_norm": 0.3064500972204145, "learning_rate": 0.00019159101887427854, "loss": 1.3082, "step": 1060 }, { "epoch": 0.21949711459192087, "grad_norm": 0.24140887151281765, "learning_rate": 0.00019144601702269162, "loss": 1.0166, "step": 1065 }, { "epoch": 0.22052761747732894, "grad_norm": 0.2751668782987028, "learning_rate": 0.00019129983148481552, "loss": 1.149, "step": 1070 }, { "epoch": 0.221558120362737, "grad_norm": 0.32570210618301154, "learning_rate": 0.0001911524641528902, "loss": 1.287, "step": 1075 }, { "epoch": 0.2225886232481451, "grad_norm": 0.24698689309459554, "learning_rate": 0.00019100391693445306, "loss": 0.9291, "step": 1080 }, { "epoch": 0.22361912613355317, "grad_norm": 0.26397978494526464, "learning_rate": 0.00019085419175231394, "loss": 1.2573, "step": 1085 }, { "epoch": 0.22464962901896127, "grad_norm": 0.23809647134840975, "learning_rate": 0.00019070329054453046, "loss": 1.0226, "step": 1090 }, { "epoch": 0.22568013190436934, "grad_norm": 0.31058086961600134, "learning_rate": 0.00019055121526438272, "loss": 1.1561, "step": 1095 }, { "epoch": 0.2267106347897774, "grad_norm": 0.2778718312695914, "learning_rate": 0.00019039796788034822, "loss": 1.3034, "step": 1100 }, { "epoch": 0.2277411376751855, "grad_norm": 0.2417800787196308, "learning_rate": 0.00019024355037607622, "loss": 0.8994, "step": 1105 }, { "epoch": 0.22877164056059357, "grad_norm": 0.28904453048763723, "learning_rate": 0.00019008796475036222, "loss": 1.2439, "step": 1110 }, { "epoch": 0.22980214344600164, "grad_norm": 0.23624741886843423, "learning_rate": 0.00018993121301712193, "loss": 1.0735, "step": 1115 }, { "epoch": 0.23083264633140974, "grad_norm": 0.27985038336320134, "learning_rate": 0.00018977329720536529, "loss": 1.1098, "step": 1120 }, { "epoch": 0.2318631492168178, "grad_norm": 0.2891758603238283, "learning_rate": 0.00018961421935917016, "loss": 1.2739, "step": 1125 }, { "epoch": 0.23289365210222587, "grad_norm": 0.2635168830732569, "learning_rate": 0.00018945398153765597, "loss": 1.0068, "step": 1130 }, { "epoch": 0.23392415498763397, "grad_norm": 0.2773479010066933, "learning_rate": 0.00018929258581495685, "loss": 1.3011, "step": 1135 }, { "epoch": 0.23495465787304204, "grad_norm": 0.2560481495589164, "learning_rate": 0.00018913003428019506, "loss": 1.0617, "step": 1140 }, { "epoch": 0.23598516075845014, "grad_norm": 0.277847323563774, "learning_rate": 0.00018896632903745374, "loss": 1.0843, "step": 1145 }, { "epoch": 0.2370156636438582, "grad_norm": 0.3618488575426716, "learning_rate": 0.00018880147220574976, "loss": 1.2909, "step": 1150 }, { "epoch": 0.23804616652926627, "grad_norm": 0.2840442439659897, "learning_rate": 0.00018863546591900622, "loss": 0.9916, "step": 1155 }, { "epoch": 0.23907666941467437, "grad_norm": 0.2904562870162801, "learning_rate": 0.00018846831232602492, "loss": 1.3014, "step": 1160 }, { "epoch": 0.24010717230008244, "grad_norm": 0.24724715741223574, "learning_rate": 0.00018830001359045845, "loss": 0.9967, "step": 1165 }, { "epoch": 0.2411376751854905, "grad_norm": 0.26593630753000524, "learning_rate": 0.00018813057189078243, "loss": 1.0626, "step": 1170 }, { "epoch": 0.2421681780708986, "grad_norm": 0.3276447476001242, "learning_rate": 0.00018795998942026685, "loss": 1.2979, "step": 1175 }, { "epoch": 0.24319868095630667, "grad_norm": 0.26299318912598185, "learning_rate": 0.00018778826838694812, "loss": 0.9544, "step": 1180 }, { "epoch": 0.24422918384171477, "grad_norm": 0.281736858661113, "learning_rate": 0.0001876154110136003, "loss": 1.3208, "step": 1185 }, { "epoch": 0.24525968672712284, "grad_norm": 0.2855862979125554, "learning_rate": 0.0001874414195377063, "loss": 1.073, "step": 1190 }, { "epoch": 0.2462901896125309, "grad_norm": 0.29014069310260776, "learning_rate": 0.000187266296211429, "loss": 1.1681, "step": 1195 }, { "epoch": 0.247320692497939, "grad_norm": 0.3159103313554715, "learning_rate": 0.0001870900433015821, "loss": 1.3863, "step": 1200 }, { "epoch": 0.247320692497939, "eval_loss": 1.0878444910049438, "eval_runtime": 2463.5391, "eval_samples_per_second": 3.247, "eval_steps_per_second": 0.203, "step": 1200 }, { "epoch": 0.24835119538334707, "grad_norm": 0.2980434519288696, "learning_rate": 0.00018691266308960066, "loss": 1.0099, "step": 1205 }, { "epoch": 0.24938169826875514, "grad_norm": 0.2786508426961495, "learning_rate": 0.00018673415787151166, "loss": 1.3381, "step": 1210 }, { "epoch": 0.25041220115416324, "grad_norm": 0.2676203832313212, "learning_rate": 0.00018655452995790435, "loss": 1.0199, "step": 1215 }, { "epoch": 0.2514427040395713, "grad_norm": 0.28577827654986465, "learning_rate": 0.00018637378167390018, "loss": 1.2001, "step": 1220 }, { "epoch": 0.2524732069249794, "grad_norm": 0.3398905155731309, "learning_rate": 0.0001861919153591228, "loss": 1.2629, "step": 1225 }, { "epoch": 0.25350370981038745, "grad_norm": 0.26851195221770446, "learning_rate": 0.00018600893336766786, "loss": 0.9409, "step": 1230 }, { "epoch": 0.25453421269579557, "grad_norm": 0.26580580512721763, "learning_rate": 0.00018582483806807228, "loss": 1.3071, "step": 1235 }, { "epoch": 0.25556471558120364, "grad_norm": 0.2571726332870005, "learning_rate": 0.0001856396318432838, "loss": 1.0674, "step": 1240 }, { "epoch": 0.2565952184666117, "grad_norm": 0.28205467317170896, "learning_rate": 0.0001854533170906302, "loss": 1.1248, "step": 1245 }, { "epoch": 0.2576257213520198, "grad_norm": 0.3312983996287865, "learning_rate": 0.00018526589622178802, "loss": 1.2796, "step": 1250 }, { "epoch": 0.25865622423742785, "grad_norm": 0.2757071364626265, "learning_rate": 0.00018507737166275154, "loss": 0.9658, "step": 1255 }, { "epoch": 0.25968672712283597, "grad_norm": 0.3459316658489597, "learning_rate": 0.00018488774585380125, "loss": 1.3143, "step": 1260 }, { "epoch": 0.26071723000824404, "grad_norm": 0.2521379464850173, "learning_rate": 0.00018469702124947245, "loss": 0.9855, "step": 1265 }, { "epoch": 0.2617477328936521, "grad_norm": 0.28337200244179356, "learning_rate": 0.00018450520031852325, "loss": 1.2029, "step": 1270 }, { "epoch": 0.2627782357790602, "grad_norm": 0.3562233737662029, "learning_rate": 0.0001843122855439027, "loss": 1.2867, "step": 1275 }, { "epoch": 0.26380873866446825, "grad_norm": 0.2883121950988055, "learning_rate": 0.00018411827942271884, "loss": 0.978, "step": 1280 }, { "epoch": 0.2648392415498763, "grad_norm": 0.3157197550829857, "learning_rate": 0.000183923184466206, "loss": 1.2767, "step": 1285 }, { "epoch": 0.26586974443528444, "grad_norm": 0.24013448810718416, "learning_rate": 0.0001837270031996926, "loss": 1.0749, "step": 1290 }, { "epoch": 0.2669002473206925, "grad_norm": 0.32239639052379476, "learning_rate": 0.00018352973816256838, "loss": 1.1699, "step": 1295 }, { "epoch": 0.2679307502061006, "grad_norm": 0.29188168243683993, "learning_rate": 0.0001833313919082515, "loss": 1.3297, "step": 1300 }, { "epoch": 0.26896125309150865, "grad_norm": 0.2710883490462726, "learning_rate": 0.0001831319670041555, "loss": 0.9246, "step": 1305 }, { "epoch": 0.2699917559769167, "grad_norm": 0.2949025846368701, "learning_rate": 0.00018293146603165603, "loss": 1.339, "step": 1310 }, { "epoch": 0.27102225886232484, "grad_norm": 0.24960673732959246, "learning_rate": 0.00018272989158605752, "loss": 1.0767, "step": 1315 }, { "epoch": 0.2720527617477329, "grad_norm": 0.31392425706558585, "learning_rate": 0.00018252724627655954, "loss": 1.1564, "step": 1320 }, { "epoch": 0.273083264633141, "grad_norm": 0.3755895664237009, "learning_rate": 0.00018232353272622302, "loss": 1.2677, "step": 1325 }, { "epoch": 0.27411376751854905, "grad_norm": 0.2711084044553726, "learning_rate": 0.00018211875357193632, "loss": 0.9061, "step": 1330 }, { "epoch": 0.2751442704039571, "grad_norm": 0.33175445691564986, "learning_rate": 0.00018191291146438105, "loss": 1.273, "step": 1335 }, { "epoch": 0.2761747732893652, "grad_norm": 0.2537090735923007, "learning_rate": 0.0001817060090679978, "loss": 1.0512, "step": 1340 }, { "epoch": 0.2772052761747733, "grad_norm": 0.36552128539895057, "learning_rate": 0.00018149804906095163, "loss": 1.2273, "step": 1345 }, { "epoch": 0.2782357790601814, "grad_norm": 0.3485807735804365, "learning_rate": 0.00018128903413509756, "loss": 1.2532, "step": 1350 }, { "epoch": 0.27926628194558945, "grad_norm": 0.2737156223424568, "learning_rate": 0.0001810789669959453, "loss": 1.0208, "step": 1355 }, { "epoch": 0.2802967848309975, "grad_norm": 0.3091305751030036, "learning_rate": 0.0001808678503626248, "loss": 1.3024, "step": 1360 }, { "epoch": 0.2813272877164056, "grad_norm": 0.2448701581737246, "learning_rate": 0.00018065568696785058, "loss": 1.0065, "step": 1365 }, { "epoch": 0.2823577906018137, "grad_norm": 0.33520109656163827, "learning_rate": 0.00018044247955788662, "loss": 1.1409, "step": 1370 }, { "epoch": 0.2833882934872218, "grad_norm": 0.36094563501069604, "learning_rate": 0.00018022823089251073, "loss": 1.2687, "step": 1375 }, { "epoch": 0.28441879637262985, "grad_norm": 0.29785608686411563, "learning_rate": 0.00018001294374497882, "loss": 0.9357, "step": 1380 }, { "epoch": 0.2854492992580379, "grad_norm": 0.3875323214985989, "learning_rate": 0.00017979662090198906, "loss": 1.2922, "step": 1385 }, { "epoch": 0.286479802143446, "grad_norm": 0.3058769686259136, "learning_rate": 0.00017957926516364565, "loss": 1.0047, "step": 1390 }, { "epoch": 0.2875103050288541, "grad_norm": 0.3030215125843954, "learning_rate": 0.00017936087934342283, "loss": 1.0829, "step": 1395 }, { "epoch": 0.2885408079142622, "grad_norm": 0.32726489328282893, "learning_rate": 0.00017914146626812823, "loss": 1.2567, "step": 1400 }, { "epoch": 0.2885408079142622, "eval_loss": 1.077731728553772, "eval_runtime": 2463.5195, "eval_samples_per_second": 3.247, "eval_steps_per_second": 0.203, "step": 1400 }, { "epoch": 0.28957131079967025, "grad_norm": 0.2992981015754632, "learning_rate": 0.0001789210287778664, "loss": 0.9792, "step": 1405 }, { "epoch": 0.2906018136850783, "grad_norm": 0.3418337934197468, "learning_rate": 0.00017869956972600202, "loss": 1.3454, "step": 1410 }, { "epoch": 0.2916323165704864, "grad_norm": 0.2526294763625045, "learning_rate": 0.00017847709197912296, "loss": 0.9715, "step": 1415 }, { "epoch": 0.29266281945589445, "grad_norm": 0.3445050915146703, "learning_rate": 0.0001782535984170032, "loss": 1.113, "step": 1420 }, { "epoch": 0.2936933223413026, "grad_norm": 0.32389102430563055, "learning_rate": 0.00017802909193256547, "loss": 1.2316, "step": 1425 }, { "epoch": 0.29472382522671065, "grad_norm": 0.2613877323741928, "learning_rate": 0.00017780357543184397, "loss": 0.9269, "step": 1430 }, { "epoch": 0.2957543281121187, "grad_norm": 0.2752040315178045, "learning_rate": 0.00017757705183394653, "loss": 1.2481, "step": 1435 }, { "epoch": 0.2967848309975268, "grad_norm": 0.2535871144663709, "learning_rate": 0.00017734952407101706, "loss": 1.0301, "step": 1440 }, { "epoch": 0.29781533388293485, "grad_norm": 0.28927983501411936, "learning_rate": 0.0001771209950881974, "loss": 1.1218, "step": 1445 }, { "epoch": 0.298845836768343, "grad_norm": 0.37005769485403384, "learning_rate": 0.00017689146784358927, "loss": 1.2699, "step": 1450 }, { "epoch": 0.29987633965375105, "grad_norm": 0.257690573166382, "learning_rate": 0.0001766609453082161, "loss": 0.9097, "step": 1455 }, { "epoch": 0.3009068425391591, "grad_norm": 0.2708622421638633, "learning_rate": 0.00017642943046598436, "loss": 1.2685, "step": 1460 }, { "epoch": 0.3019373454245672, "grad_norm": 0.2553661493662262, "learning_rate": 0.00017619692631364506, "loss": 1.062, "step": 1465 }, { "epoch": 0.30296784830997525, "grad_norm": 0.3268803054784382, "learning_rate": 0.00017596343586075497, "loss": 1.1081, "step": 1470 }, { "epoch": 0.3039983511953833, "grad_norm": 0.36412223215382217, "learning_rate": 0.00017572896212963754, "loss": 1.3221, "step": 1475 }, { "epoch": 0.30502885408079145, "grad_norm": 0.2624897168675397, "learning_rate": 0.000175493508155344, "loss": 0.854, "step": 1480 }, { "epoch": 0.3060593569661995, "grad_norm": 0.2910755462081272, "learning_rate": 0.00017525707698561385, "loss": 1.2961, "step": 1485 }, { "epoch": 0.3070898598516076, "grad_norm": 0.26020570413888594, "learning_rate": 0.00017501967168083557, "loss": 1.0475, "step": 1490 }, { "epoch": 0.30812036273701565, "grad_norm": 0.31355952905674644, "learning_rate": 0.00017478129531400688, "loss": 1.1383, "step": 1495 }, { "epoch": 0.3091508656224237, "grad_norm": 0.3394529288865745, "learning_rate": 0.00017454195097069505, "loss": 1.3237, "step": 1500 }, { "epoch": 0.31018136850783184, "grad_norm": 0.2623708218727371, "learning_rate": 0.00017430164174899696, "loss": 0.8949, "step": 1505 }, { "epoch": 0.3112118713932399, "grad_norm": 0.3129719424005796, "learning_rate": 0.0001740603707594989, "loss": 1.2513, "step": 1510 }, { "epoch": 0.312242374278648, "grad_norm": 0.2457430728630137, "learning_rate": 0.00017381814112523648, "loss": 1.0458, "step": 1515 }, { "epoch": 0.31327287716405605, "grad_norm": 0.3235157301586437, "learning_rate": 0.000173574955981654, "loss": 1.1418, "step": 1520 }, { "epoch": 0.3143033800494641, "grad_norm": 0.38068266969796094, "learning_rate": 0.00017333081847656397, "loss": 1.2703, "step": 1525 }, { "epoch": 0.31533388293487224, "grad_norm": 0.2845050833721142, "learning_rate": 0.00017308573177010652, "loss": 0.8513, "step": 1530 }, { "epoch": 0.3163643858202803, "grad_norm": 0.3011956997232069, "learning_rate": 0.00017283969903470815, "loss": 1.2671, "step": 1535 }, { "epoch": 0.3173948887056884, "grad_norm": 0.2548678800376021, "learning_rate": 0.0001725927234550409, "loss": 0.9834, "step": 1540 }, { "epoch": 0.31842539159109645, "grad_norm": 0.33070817321371104, "learning_rate": 0.00017234480822798113, "loss": 1.204, "step": 1545 }, { "epoch": 0.3194558944765045, "grad_norm": 0.3739990601669961, "learning_rate": 0.00017209595656256807, "loss": 1.2874, "step": 1550 }, { "epoch": 0.3204863973619126, "grad_norm": 0.28630568916559335, "learning_rate": 0.0001718461716799623, "loss": 0.9404, "step": 1555 }, { "epoch": 0.3215169002473207, "grad_norm": 0.2984332733087229, "learning_rate": 0.000171595456813404, "loss": 1.3065, "step": 1560 }, { "epoch": 0.3225474031327288, "grad_norm": 0.25297494310840185, "learning_rate": 0.00017134381520817127, "loss": 0.9533, "step": 1565 }, { "epoch": 0.32357790601813685, "grad_norm": 0.3035622251707662, "learning_rate": 0.00017109125012153783, "loss": 1.1281, "step": 1570 }, { "epoch": 0.3246084089035449, "grad_norm": 0.3401968875424911, "learning_rate": 0.00017083776482273126, "loss": 1.2439, "step": 1575 }, { "epoch": 0.325638911788953, "grad_norm": 0.2742633328984466, "learning_rate": 0.00017058336259289026, "loss": 0.9067, "step": 1580 }, { "epoch": 0.3266694146743611, "grad_norm": 0.33201726031181766, "learning_rate": 0.0001703280467250225, "loss": 1.2643, "step": 1585 }, { "epoch": 0.3276999175597692, "grad_norm": 0.2740668506157443, "learning_rate": 0.0001700718205239618, "loss": 0.9757, "step": 1590 }, { "epoch": 0.32873042044517725, "grad_norm": 0.28447062660977135, "learning_rate": 0.0001698146873063255, "loss": 1.0629, "step": 1595 }, { "epoch": 0.3297609233305853, "grad_norm": 0.3855059276442332, "learning_rate": 0.00016955665040047134, "loss": 1.257, "step": 1600 }, { "epoch": 0.3297609233305853, "eval_loss": 1.0629926919937134, "eval_runtime": 2467.5248, "eval_samples_per_second": 3.242, "eval_steps_per_second": 0.203, "step": 1600 }, { "epoch": 0.3307914262159934, "grad_norm": 0.30766662381135884, "learning_rate": 0.00016929771314645454, "loss": 0.9716, "step": 1605 }, { "epoch": 0.33182192910140146, "grad_norm": 0.3070232969296392, "learning_rate": 0.00016903787889598458, "loss": 1.3027, "step": 1610 }, { "epoch": 0.3328524319868096, "grad_norm": 0.29718017467704144, "learning_rate": 0.00016877715101238172, "loss": 1.053, "step": 1615 }, { "epoch": 0.33388293487221765, "grad_norm": 0.29124927902233755, "learning_rate": 0.00016851553287053342, "loss": 1.1027, "step": 1620 }, { "epoch": 0.3349134377576257, "grad_norm": 0.37490768596522844, "learning_rate": 0.00016825302785685077, "loss": 1.2816, "step": 1625 }, { "epoch": 0.3359439406430338, "grad_norm": 0.2909222245308242, "learning_rate": 0.00016798963936922467, "loss": 0.9477, "step": 1630 }, { "epoch": 0.33697444352844186, "grad_norm": 0.369759262778725, "learning_rate": 0.00016772537081698175, "loss": 1.2659, "step": 1635 }, { "epoch": 0.33800494641385, "grad_norm": 0.25527189055893823, "learning_rate": 0.00016746022562084026, "loss": 1.0374, "step": 1640 }, { "epoch": 0.33903544929925805, "grad_norm": 0.332617684071012, "learning_rate": 0.0001671942072128659, "loss": 1.1066, "step": 1645 }, { "epoch": 0.3400659521846661, "grad_norm": 0.3184886656626706, "learning_rate": 0.00016692731903642725, "loss": 1.2649, "step": 1650 }, { "epoch": 0.3410964550700742, "grad_norm": 0.2799293838329678, "learning_rate": 0.0001666595645461512, "loss": 0.9379, "step": 1655 }, { "epoch": 0.34212695795548226, "grad_norm": 0.3517762307499943, "learning_rate": 0.0001663909472078784, "loss": 1.252, "step": 1660 }, { "epoch": 0.3431574608408903, "grad_norm": 0.25384775837778967, "learning_rate": 0.0001661214704986182, "loss": 0.9847, "step": 1665 }, { "epoch": 0.34418796372629845, "grad_norm": 0.29743299437921455, "learning_rate": 0.00016585113790650388, "loss": 1.1201, "step": 1670 }, { "epoch": 0.3452184666117065, "grad_norm": 0.35589431095248486, "learning_rate": 0.00016557995293074715, "loss": 1.2206, "step": 1675 }, { "epoch": 0.3462489694971146, "grad_norm": 0.2911495340019745, "learning_rate": 0.00016530791908159323, "loss": 0.9072, "step": 1680 }, { "epoch": 0.34727947238252266, "grad_norm": 0.28789636914166633, "learning_rate": 0.0001650350398802751, "loss": 1.2946, "step": 1685 }, { "epoch": 0.3483099752679307, "grad_norm": 0.26335155835600166, "learning_rate": 0.0001647613188589682, "loss": 1.0242, "step": 1690 }, { "epoch": 0.34934047815333885, "grad_norm": 0.3325623900144522, "learning_rate": 0.00016448675956074444, "loss": 1.0931, "step": 1695 }, { "epoch": 0.3503709810387469, "grad_norm": 0.3773837775653675, "learning_rate": 0.0001642113655395266, "loss": 1.2456, "step": 1700 }, { "epoch": 0.351401483924155, "grad_norm": 0.285753347787117, "learning_rate": 0.00016393514036004204, "loss": 0.9662, "step": 1705 }, { "epoch": 0.35243198680956306, "grad_norm": 0.3046613746172406, "learning_rate": 0.0001636580875977769, "loss": 1.3162, "step": 1710 }, { "epoch": 0.3534624896949711, "grad_norm": 0.25099480129499735, "learning_rate": 0.0001633802108389295, "loss": 1.0061, "step": 1715 }, { "epoch": 0.35449299258037925, "grad_norm": 0.34150662422413947, "learning_rate": 0.00016310151368036408, "loss": 1.1411, "step": 1720 }, { "epoch": 0.3555234954657873, "grad_norm": 0.33392898782461256, "learning_rate": 0.00016282199972956425, "loss": 1.2154, "step": 1725 }, { "epoch": 0.3565539983511954, "grad_norm": 0.3016273908980527, "learning_rate": 0.00016254167260458622, "loss": 0.9269, "step": 1730 }, { "epoch": 0.35758450123660346, "grad_norm": 0.3297238572031139, "learning_rate": 0.000162260535934012, "loss": 1.2298, "step": 1735 }, { "epoch": 0.3586150041220115, "grad_norm": 0.26311212102325543, "learning_rate": 0.00016197859335690247, "loss": 1.0416, "step": 1740 }, { "epoch": 0.3596455070074196, "grad_norm": 0.35401650365663206, "learning_rate": 0.0001616958485227503, "loss": 1.0915, "step": 1745 }, { "epoch": 0.3606760098928277, "grad_norm": 0.3946861144577044, "learning_rate": 0.0001614123050914325, "loss": 1.2156, "step": 1750 }, { "epoch": 0.3617065127782358, "grad_norm": 0.31649222443853475, "learning_rate": 0.00016112796673316336, "loss": 0.9126, "step": 1755 }, { "epoch": 0.36273701566364386, "grad_norm": 0.3591291538700441, "learning_rate": 0.00016084283712844666, "loss": 1.2438, "step": 1760 }, { "epoch": 0.3637675185490519, "grad_norm": 0.2567906575863548, "learning_rate": 0.00016055691996802823, "loss": 0.9908, "step": 1765 }, { "epoch": 0.36479802143446, "grad_norm": 0.34039693260231785, "learning_rate": 0.00016027021895284808, "loss": 1.0576, "step": 1770 }, { "epoch": 0.3658285243198681, "grad_norm": 0.45929799156275763, "learning_rate": 0.0001599827377939925, "loss": 1.3099, "step": 1775 }, { "epoch": 0.3668590272052762, "grad_norm": 0.2736394916456097, "learning_rate": 0.00015969448021264606, "loss": 0.8622, "step": 1780 }, { "epoch": 0.36788953009068426, "grad_norm": 0.2884997182361836, "learning_rate": 0.00015940544994004334, "loss": 1.2272, "step": 1785 }, { "epoch": 0.3689200329760923, "grad_norm": 0.2392312698637367, "learning_rate": 0.00015911565071742088, "loss": 0.9626, "step": 1790 }, { "epoch": 0.3699505358615004, "grad_norm": 0.339213144942226, "learning_rate": 0.00015882508629596836, "loss": 1.1233, "step": 1795 }, { "epoch": 0.37098103874690846, "grad_norm": 0.40881279626532063, "learning_rate": 0.00015853376043678053, "loss": 1.2129, "step": 1800 }, { "epoch": 0.37098103874690846, "eval_loss": 1.051790475845337, "eval_runtime": 2468.3063, "eval_samples_per_second": 3.241, "eval_steps_per_second": 0.203, "step": 1800 }, { "epoch": 0.3720115416323166, "grad_norm": 0.27225360643392454, "learning_rate": 0.00015824167691080802, "loss": 0.8599, "step": 1805 }, { "epoch": 0.37304204451772466, "grad_norm": 0.3655053849913508, "learning_rate": 0.00015794883949880894, "loss": 1.2174, "step": 1810 }, { "epoch": 0.3740725474031327, "grad_norm": 0.2810810355550597, "learning_rate": 0.00015765525199129966, "loss": 1.0166, "step": 1815 }, { "epoch": 0.3751030502885408, "grad_norm": 0.3332282628999955, "learning_rate": 0.000157360918188506, "loss": 1.1182, "step": 1820 }, { "epoch": 0.37613355317394886, "grad_norm": 0.3175764657865911, "learning_rate": 0.0001570658419003137, "loss": 1.2687, "step": 1825 }, { "epoch": 0.377164056059357, "grad_norm": 0.3120363543994925, "learning_rate": 0.00015677002694621948, "loss": 0.9812, "step": 1830 }, { "epoch": 0.37819455894476506, "grad_norm": 0.327602667442098, "learning_rate": 0.00015647347715528137, "loss": 1.1919, "step": 1835 }, { "epoch": 0.3792250618301731, "grad_norm": 0.2668526402245522, "learning_rate": 0.00015617619636606924, "loss": 0.9995, "step": 1840 }, { "epoch": 0.3802555647155812, "grad_norm": 0.31419704503426554, "learning_rate": 0.00015587818842661494, "loss": 1.1765, "step": 1845 }, { "epoch": 0.38128606760098926, "grad_norm": 0.4282942251054684, "learning_rate": 0.00015557945719436278, "loss": 1.2522, "step": 1850 }, { "epoch": 0.3823165704863974, "grad_norm": 0.308069434736153, "learning_rate": 0.00015528000653611935, "loss": 0.8995, "step": 1855 }, { "epoch": 0.38334707337180546, "grad_norm": 0.272826512737121, "learning_rate": 0.0001549798403280036, "loss": 1.2827, "step": 1860 }, { "epoch": 0.3843775762572135, "grad_norm": 0.2498018349436216, "learning_rate": 0.0001546789624553966, "loss": 1.0308, "step": 1865 }, { "epoch": 0.3854080791426216, "grad_norm": 0.32988195624784494, "learning_rate": 0.00015437737681289128, "loss": 1.1726, "step": 1870 }, { "epoch": 0.38643858202802966, "grad_norm": 0.38213734379034847, "learning_rate": 0.00015407508730424206, "loss": 1.2046, "step": 1875 }, { "epoch": 0.38746908491343773, "grad_norm": 0.2839035313263252, "learning_rate": 0.00015377209784231424, "loss": 0.9713, "step": 1880 }, { "epoch": 0.38849958779884586, "grad_norm": 0.26153622554750505, "learning_rate": 0.00015346841234903337, "loss": 1.2771, "step": 1885 }, { "epoch": 0.3895300906842539, "grad_norm": 0.25504870927422546, "learning_rate": 0.0001531640347553345, "loss": 1.0141, "step": 1890 }, { "epoch": 0.390560593569662, "grad_norm": 0.33076548230805997, "learning_rate": 0.00015285896900111133, "loss": 1.1153, "step": 1895 }, { "epoch": 0.39159109645507006, "grad_norm": 0.3682940168037353, "learning_rate": 0.00015255321903516516, "loss": 1.2387, "step": 1900 }, { "epoch": 0.39262159934047813, "grad_norm": 0.3030559581752516, "learning_rate": 0.00015224678881515377, "loss": 0.8932, "step": 1905 }, { "epoch": 0.39365210222588626, "grad_norm": 0.3016417378121846, "learning_rate": 0.00015193968230754024, "loss": 1.2566, "step": 1910 }, { "epoch": 0.3946826051112943, "grad_norm": 0.2562267990440701, "learning_rate": 0.00015163190348754162, "loss": 1.0584, "step": 1915 }, { "epoch": 0.3957131079967024, "grad_norm": 0.3152397653362513, "learning_rate": 0.00015132345633907734, "loss": 1.1119, "step": 1920 }, { "epoch": 0.39674361088211046, "grad_norm": 0.3550077716096075, "learning_rate": 0.0001510143448547178, "loss": 1.2196, "step": 1925 }, { "epoch": 0.39777411376751853, "grad_norm": 0.2713673784146358, "learning_rate": 0.00015070457303563268, "loss": 0.8809, "step": 1930 }, { "epoch": 0.3988046166529266, "grad_norm": 0.32482913652095136, "learning_rate": 0.000150394144891539, "loss": 1.2114, "step": 1935 }, { "epoch": 0.3998351195383347, "grad_norm": 0.2681392320249827, "learning_rate": 0.00015008306444064942, "loss": 1.0092, "step": 1940 }, { "epoch": 0.4008656224237428, "grad_norm": 0.30813511160423507, "learning_rate": 0.00014977133570961997, "loss": 1.0587, "step": 1945 }, { "epoch": 0.40189612530915086, "grad_norm": 0.3045685029840853, "learning_rate": 0.00014945896273349827, "loss": 1.3074, "step": 1950 }, { "epoch": 0.40292662819455893, "grad_norm": 0.3178744984824384, "learning_rate": 0.00014914594955567099, "loss": 0.9477, "step": 1955 }, { "epoch": 0.403957131079967, "grad_norm": 0.36931867482484526, "learning_rate": 0.00014883230022781163, "loss": 1.2359, "step": 1960 }, { "epoch": 0.4049876339653751, "grad_norm": 0.27801383031198734, "learning_rate": 0.00014851801880982814, "loss": 1.0366, "step": 1965 }, { "epoch": 0.4060181368507832, "grad_norm": 0.31213479911432085, "learning_rate": 0.00014820310936981026, "loss": 1.1624, "step": 1970 }, { "epoch": 0.40704863973619126, "grad_norm": 0.33227538300256093, "learning_rate": 0.000147887575983977, "loss": 1.2724, "step": 1975 }, { "epoch": 0.40807914262159933, "grad_norm": 0.298687515913839, "learning_rate": 0.00014757142273662358, "loss": 0.8436, "step": 1980 }, { "epoch": 0.4091096455070074, "grad_norm": 0.36351887792750426, "learning_rate": 0.00014725465372006905, "loss": 1.2507, "step": 1985 }, { "epoch": 0.41014014839241547, "grad_norm": 0.2702448610300903, "learning_rate": 0.0001469372730346028, "loss": 1.0191, "step": 1990 }, { "epoch": 0.4111706512778236, "grad_norm": 0.3224079902083336, "learning_rate": 0.00014661928478843186, "loss": 1.1598, "step": 1995 }, { "epoch": 0.41220115416323166, "grad_norm": 0.41798586070796323, "learning_rate": 0.00014630069309762753, "loss": 1.1939, "step": 2000 }, { "epoch": 0.41220115416323166, "eval_loss": 1.0405410528182983, "eval_runtime": 2474.052, "eval_samples_per_second": 3.234, "eval_steps_per_second": 0.202, "step": 2000 }, { "epoch": 0.41323165704863973, "grad_norm": 0.28365317528526945, "learning_rate": 0.00014598150208607212, "loss": 0.8639, "step": 2005 }, { "epoch": 0.4142621599340478, "grad_norm": 0.3214047753618349, "learning_rate": 0.00014566171588540572, "loss": 1.2183, "step": 2010 }, { "epoch": 0.41529266281945587, "grad_norm": 0.2644278825870432, "learning_rate": 0.00014534133863497252, "loss": 0.9564, "step": 2015 }, { "epoch": 0.416323165704864, "grad_norm": 0.32226077969266037, "learning_rate": 0.00014502037448176734, "loss": 1.0942, "step": 2020 }, { "epoch": 0.41735366859027206, "grad_norm": 0.4101159870787422, "learning_rate": 0.00014469882758038193, "loss": 1.244, "step": 2025 }, { "epoch": 0.41838417147568013, "grad_norm": 0.28153576445753414, "learning_rate": 0.00014437670209295112, "loss": 0.8947, "step": 2030 }, { "epoch": 0.4194146743610882, "grad_norm": 0.3785396934306716, "learning_rate": 0.0001440540021890992, "loss": 1.2432, "step": 2035 }, { "epoch": 0.42044517724649627, "grad_norm": 0.2542390029018319, "learning_rate": 0.00014373073204588556, "loss": 1.0083, "step": 2040 }, { "epoch": 0.4214756801319044, "grad_norm": 0.3123729353287892, "learning_rate": 0.0001434068958477509, "loss": 1.1195, "step": 2045 }, { "epoch": 0.42250618301731246, "grad_norm": 0.39530507310232316, "learning_rate": 0.00014308249778646306, "loss": 1.2036, "step": 2050 }, { "epoch": 0.42353668590272053, "grad_norm": 0.2818105109815067, "learning_rate": 0.0001427575420610626, "loss": 0.8847, "step": 2055 }, { "epoch": 0.4245671887881286, "grad_norm": 0.3219044657747011, "learning_rate": 0.00014243203287780856, "loss": 1.2623, "step": 2060 }, { "epoch": 0.42559769167353667, "grad_norm": 0.2611111361591686, "learning_rate": 0.00014210597445012398, "loss": 1.0083, "step": 2065 }, { "epoch": 0.42662819455894474, "grad_norm": 0.33485314347370027, "learning_rate": 0.0001417793709985415, "loss": 1.1823, "step": 2070 }, { "epoch": 0.42765869744435286, "grad_norm": 0.34925710089096873, "learning_rate": 0.0001414522267506484, "loss": 1.2728, "step": 2075 }, { "epoch": 0.42868920032976093, "grad_norm": 0.25546314251507235, "learning_rate": 0.0001411245459410322, "loss": 0.8703, "step": 2080 }, { "epoch": 0.429719703215169, "grad_norm": 0.38324554510188286, "learning_rate": 0.00014079633281122573, "loss": 1.2551, "step": 2085 }, { "epoch": 0.43075020610057707, "grad_norm": 0.27810663295608457, "learning_rate": 0.00014046759160965224, "loss": 1.0267, "step": 2090 }, { "epoch": 0.43178070898598514, "grad_norm": 0.35335784100847833, "learning_rate": 0.0001401383265915703, "loss": 1.1345, "step": 2095 }, { "epoch": 0.43281121187139326, "grad_norm": 0.4118966005270665, "learning_rate": 0.00013980854201901886, "loss": 1.1936, "step": 2100 }, { "epoch": 0.43384171475680133, "grad_norm": 0.3062362602687425, "learning_rate": 0.00013947824216076207, "loss": 0.9152, "step": 2105 }, { "epoch": 0.4348722176422094, "grad_norm": 0.29620498290186464, "learning_rate": 0.00013914743129223405, "loss": 1.2309, "step": 2110 }, { "epoch": 0.43590272052761747, "grad_norm": 0.2710224032495773, "learning_rate": 0.00013881611369548325, "loss": 1.0408, "step": 2115 }, { "epoch": 0.43693322341302554, "grad_norm": 0.3056464151796937, "learning_rate": 0.00013848429365911753, "loss": 1.1743, "step": 2120 }, { "epoch": 0.4379637262984336, "grad_norm": 0.3843468598801491, "learning_rate": 0.00013815197547824824, "loss": 1.2395, "step": 2125 }, { "epoch": 0.43899422918384173, "grad_norm": 0.2798396925326106, "learning_rate": 0.00013781916345443474, "loss": 0.8612, "step": 2130 }, { "epoch": 0.4400247320692498, "grad_norm": 0.3279250283208323, "learning_rate": 0.00013748586189562878, "loss": 1.2354, "step": 2135 }, { "epoch": 0.44105523495465787, "grad_norm": 0.2735131812328209, "learning_rate": 0.00013715207511611876, "loss": 1.0026, "step": 2140 }, { "epoch": 0.44208573784006594, "grad_norm": 0.37219144396512693, "learning_rate": 0.0001368178074364737, "loss": 1.0704, "step": 2145 }, { "epoch": 0.443116240725474, "grad_norm": 0.46408445168045037, "learning_rate": 0.00013648306318348762, "loss": 1.2511, "step": 2150 }, { "epoch": 0.44414674361088213, "grad_norm": 0.2720164174270696, "learning_rate": 0.0001361478466901231, "loss": 0.8989, "step": 2155 }, { "epoch": 0.4451772464962902, "grad_norm": 0.3747167130586965, "learning_rate": 0.0001358121622954557, "loss": 1.3666, "step": 2160 }, { "epoch": 0.44620774938169827, "grad_norm": 0.2962954738156199, "learning_rate": 0.00013547601434461733, "loss": 0.9376, "step": 2165 }, { "epoch": 0.44723825226710634, "grad_norm": 0.3347575447270046, "learning_rate": 0.0001351394071887404, "loss": 1.0754, "step": 2170 }, { "epoch": 0.4482687551525144, "grad_norm": 0.37805918523823334, "learning_rate": 0.0001348023451849012, "loss": 1.2106, "step": 2175 }, { "epoch": 0.44929925803792253, "grad_norm": 0.302277885917108, "learning_rate": 0.00013446483269606362, "loss": 0.8481, "step": 2180 }, { "epoch": 0.4503297609233306, "grad_norm": 0.35890494559800096, "learning_rate": 0.00013412687409102277, "loss": 1.2101, "step": 2185 }, { "epoch": 0.45136026380873867, "grad_norm": 0.27832175193098935, "learning_rate": 0.00013378847374434814, "loss": 0.9563, "step": 2190 }, { "epoch": 0.45239076669414674, "grad_norm": 0.35752890078545335, "learning_rate": 0.0001334496360363274, "loss": 1.1166, "step": 2195 }, { "epoch": 0.4534212695795548, "grad_norm": 0.3208813709791692, "learning_rate": 0.00013311036535290925, "loss": 1.2658, "step": 2200 }, { "epoch": 0.4534212695795548, "eval_loss": 1.0313003063201904, "eval_runtime": 2462.8678, "eval_samples_per_second": 3.248, "eval_steps_per_second": 0.203, "step": 2200 }, { "epoch": 0.4544517724649629, "grad_norm": 0.3164662130947771, "learning_rate": 0.00013277066608564713, "loss": 0.9046, "step": 2205 }, { "epoch": 0.455482275350371, "grad_norm": 0.32760397119015733, "learning_rate": 0.0001324305426316418, "loss": 1.2596, "step": 2210 }, { "epoch": 0.45651277823577907, "grad_norm": 0.26879676205264913, "learning_rate": 0.000132089999393485, "loss": 0.9842, "step": 2215 }, { "epoch": 0.45754328112118714, "grad_norm": 0.35211465947054194, "learning_rate": 0.000131749040779202, "loss": 1.1149, "step": 2220 }, { "epoch": 0.4585737840065952, "grad_norm": 0.3640438187361373, "learning_rate": 0.0001314076712021949, "loss": 1.2521, "step": 2225 }, { "epoch": 0.4596042868920033, "grad_norm": 0.28859583596302096, "learning_rate": 0.0001310658950811852, "loss": 0.8689, "step": 2230 }, { "epoch": 0.4606347897774114, "grad_norm": 0.3245571156879321, "learning_rate": 0.00013072371684015688, "loss": 1.2013, "step": 2235 }, { "epoch": 0.46166529266281947, "grad_norm": 0.2630264452723511, "learning_rate": 0.00013038114090829892, "loss": 1.0054, "step": 2240 }, { "epoch": 0.46269579554822754, "grad_norm": 0.3629689887711568, "learning_rate": 0.00013003817171994807, "loss": 1.1201, "step": 2245 }, { "epoch": 0.4637262984336356, "grad_norm": 0.38051210660623236, "learning_rate": 0.00012969481371453135, "loss": 1.2291, "step": 2250 }, { "epoch": 0.4647568013190437, "grad_norm": 0.3038049492726604, "learning_rate": 0.00012935107133650885, "loss": 0.8923, "step": 2255 }, { "epoch": 0.46578730420445175, "grad_norm": 0.34319934420858916, "learning_rate": 0.00012900694903531586, "loss": 1.2325, "step": 2260 }, { "epoch": 0.46681780708985987, "grad_norm": 0.2729387501421985, "learning_rate": 0.0001286624512653055, "loss": 0.9893, "step": 2265 }, { "epoch": 0.46784830997526794, "grad_norm": 0.37018830848703915, "learning_rate": 0.00012831758248569097, "loss": 1.1021, "step": 2270 }, { "epoch": 0.468878812860676, "grad_norm": 0.36278206444854155, "learning_rate": 0.00012797234716048784, "loss": 1.2005, "step": 2275 }, { "epoch": 0.4699093157460841, "grad_norm": 0.2976065955292209, "learning_rate": 0.00012762674975845637, "loss": 0.8405, "step": 2280 }, { "epoch": 0.47093981863149215, "grad_norm": 0.3358390029204746, "learning_rate": 0.00012728079475304345, "loss": 1.2722, "step": 2285 }, { "epoch": 0.47197032151690027, "grad_norm": 0.24689134910878455, "learning_rate": 0.0001269344866223249, "loss": 0.9856, "step": 2290 }, { "epoch": 0.47300082440230834, "grad_norm": 0.3183219956731036, "learning_rate": 0.00012658782984894743, "loss": 1.0672, "step": 2295 }, { "epoch": 0.4740313272877164, "grad_norm": 0.3220468985387059, "learning_rate": 0.00012624082892007064, "loss": 1.2301, "step": 2300 }, { "epoch": 0.4750618301731245, "grad_norm": 0.3012629746942751, "learning_rate": 0.00012589348832730882, "loss": 0.9345, "step": 2305 }, { "epoch": 0.47609233305853255, "grad_norm": 0.34208223291291867, "learning_rate": 0.00012554581256667296, "loss": 1.1783, "step": 2310 }, { "epoch": 0.47712283594394067, "grad_norm": 0.2633458752929658, "learning_rate": 0.00012519780613851254, "loss": 1.0127, "step": 2315 }, { "epoch": 0.47815333882934874, "grad_norm": 0.3245639160611015, "learning_rate": 0.00012484947354745714, "loss": 1.074, "step": 2320 }, { "epoch": 0.4791838417147568, "grad_norm": 0.4155838481655266, "learning_rate": 0.0001245008193023583, "loss": 1.2446, "step": 2325 }, { "epoch": 0.4802143446001649, "grad_norm": 0.29474111212781395, "learning_rate": 0.00012415184791623101, "loss": 0.8756, "step": 2330 }, { "epoch": 0.48124484748557295, "grad_norm": 0.38251199632272437, "learning_rate": 0.00012380256390619548, "loss": 1.2367, "step": 2335 }, { "epoch": 0.482275350370981, "grad_norm": 0.2876263709986664, "learning_rate": 0.00012345297179341844, "loss": 0.942, "step": 2340 }, { "epoch": 0.48330585325638914, "grad_norm": 0.3470650266011227, "learning_rate": 0.00012310307610305477, "loss": 1.1312, "step": 2345 }, { "epoch": 0.4843363561417972, "grad_norm": 0.3437397458970076, "learning_rate": 0.00012275288136418889, "loss": 1.2013, "step": 2350 }, { "epoch": 0.4853668590272053, "grad_norm": 0.2891400010122533, "learning_rate": 0.0001224023921097762, "loss": 0.875, "step": 2355 }, { "epoch": 0.48639736191261335, "grad_norm": 0.3343377419438735, "learning_rate": 0.0001220516128765842, "loss": 1.2263, "step": 2360 }, { "epoch": 0.4874278647980214, "grad_norm": 0.2815731947053776, "learning_rate": 0.00012170054820513401, "loss": 1.0279, "step": 2365 }, { "epoch": 0.48845836768342954, "grad_norm": 0.34918143666071294, "learning_rate": 0.00012134920263964147, "loss": 1.1073, "step": 2370 }, { "epoch": 0.4894888705688376, "grad_norm": 0.3378508040938572, "learning_rate": 0.00012099758072795832, "loss": 1.2131, "step": 2375 }, { "epoch": 0.4905193734542457, "grad_norm": 0.3404097695213611, "learning_rate": 0.00012064568702151335, "loss": 0.9207, "step": 2380 }, { "epoch": 0.49154987633965375, "grad_norm": 0.31871974527595615, "learning_rate": 0.0001202935260752535, "loss": 1.2815, "step": 2385 }, { "epoch": 0.4925803792250618, "grad_norm": 0.27586523947951513, "learning_rate": 0.00011994110244758496, "loss": 0.9798, "step": 2390 }, { "epoch": 0.4936108821104699, "grad_norm": 0.29258038219303606, "learning_rate": 0.00011958842070031395, "loss": 1.0612, "step": 2395 }, { "epoch": 0.494641384995878, "grad_norm": 0.37118694749476105, "learning_rate": 0.0001192354853985879, "loss": 1.1718, "step": 2400 }, { "epoch": 0.494641384995878, "eval_loss": 1.0186035633087158, "eval_runtime": 2474.4655, "eval_samples_per_second": 3.233, "eval_steps_per_second": 0.202, "step": 2400 }, { "epoch": 0.4956718878812861, "grad_norm": 0.3041203248300929, "learning_rate": 0.00011888230111083627, "loss": 0.8718, "step": 2405 }, { "epoch": 0.49670239076669415, "grad_norm": 0.34626490332091997, "learning_rate": 0.00011852887240871145, "loss": 1.2384, "step": 2410 }, { "epoch": 0.4977328936521022, "grad_norm": 0.2627628128026553, "learning_rate": 0.00011817520386702947, "loss": 0.9474, "step": 2415 }, { "epoch": 0.4987633965375103, "grad_norm": 0.3243643599593401, "learning_rate": 0.00011782130006371092, "loss": 1.0514, "step": 2420 }, { "epoch": 0.4997938994229184, "grad_norm": 0.38727382866372845, "learning_rate": 0.00011746716557972167, "loss": 1.2191, "step": 2425 }, { "epoch": 0.5008244023083265, "grad_norm": 0.29252070214667514, "learning_rate": 0.00011711280499901347, "loss": 0.8687, "step": 2430 }, { "epoch": 0.5018549051937345, "grad_norm": 0.32103295732093995, "learning_rate": 0.00011675822290846474, "loss": 1.2169, "step": 2435 }, { "epoch": 0.5028854080791426, "grad_norm": 0.2619181848445574, "learning_rate": 0.00011640342389782114, "loss": 0.9916, "step": 2440 }, { "epoch": 0.5039159109645507, "grad_norm": 0.36227995308769084, "learning_rate": 0.00011604841255963616, "loss": 1.0797, "step": 2445 }, { "epoch": 0.5049464138499588, "grad_norm": 0.4472357927883422, "learning_rate": 0.00011569319348921168, "loss": 1.2004, "step": 2450 }, { "epoch": 0.5059769167353668, "grad_norm": 0.2990090832488795, "learning_rate": 0.00011533777128453844, "loss": 0.905, "step": 2455 }, { "epoch": 0.5070074196207749, "grad_norm": 0.3421802694540464, "learning_rate": 0.00011498215054623664, "loss": 1.2502, "step": 2460 }, { "epoch": 0.5080379225061831, "grad_norm": 0.2619877856697061, "learning_rate": 0.00011462633587749629, "loss": 0.9878, "step": 2465 }, { "epoch": 0.5090684253915911, "grad_norm": 0.3573733070627249, "learning_rate": 0.00011427033188401768, "loss": 1.0799, "step": 2470 }, { "epoch": 0.5100989282769992, "grad_norm": 0.36110594973295185, "learning_rate": 0.00011391414317395167, "loss": 1.1831, "step": 2475 }, { "epoch": 0.5111294311624073, "grad_norm": 0.29726968151818156, "learning_rate": 0.00011355777435784016, "loss": 0.9268, "step": 2480 }, { "epoch": 0.5121599340478153, "grad_norm": 0.38992649702696064, "learning_rate": 0.0001132012300485564, "loss": 1.2403, "step": 2485 }, { "epoch": 0.5131904369332234, "grad_norm": 0.31217559249274757, "learning_rate": 0.00011284451486124514, "loss": 0.9683, "step": 2490 }, { "epoch": 0.5142209398186315, "grad_norm": 0.3309516167120589, "learning_rate": 0.00011248763341326307, "loss": 1.018, "step": 2495 }, { "epoch": 0.5152514427040396, "grad_norm": 0.3669152523683029, "learning_rate": 0.00011213059032411897, "loss": 1.2018, "step": 2500 }, { "epoch": 0.5162819455894476, "grad_norm": 0.26794057215849676, "learning_rate": 0.00011177339021541387, "loss": 0.8112, "step": 2505 }, { "epoch": 0.5173124484748557, "grad_norm": 0.35964442987172796, "learning_rate": 0.00011141603771078133, "loss": 1.1878, "step": 2510 }, { "epoch": 0.5183429513602638, "grad_norm": 0.26756565994743753, "learning_rate": 0.00011105853743582751, "loss": 1.0244, "step": 2515 }, { "epoch": 0.5193734542456719, "grad_norm": 0.37262291618379195, "learning_rate": 0.00011070089401807129, "loss": 1.0956, "step": 2520 }, { "epoch": 0.52040395713108, "grad_norm": 0.3862693430821137, "learning_rate": 0.0001103431120868845, "loss": 1.1777, "step": 2525 }, { "epoch": 0.5214344600164881, "grad_norm": 0.34456489334354296, "learning_rate": 0.00010998519627343182, "loss": 0.9219, "step": 2530 }, { "epoch": 0.5224649629018961, "grad_norm": 0.3363644750152242, "learning_rate": 0.00010962715121061095, "loss": 1.2219, "step": 2535 }, { "epoch": 0.5234954657873042, "grad_norm": 0.2773020803302085, "learning_rate": 0.00010926898153299259, "loss": 0.9829, "step": 2540 }, { "epoch": 0.5245259686727123, "grad_norm": 0.3387440320396881, "learning_rate": 0.00010891069187676051, "loss": 1.1028, "step": 2545 }, { "epoch": 0.5255564715581204, "grad_norm": 0.4105637833575777, "learning_rate": 0.00010855228687965138, "loss": 1.1622, "step": 2550 }, { "epoch": 0.5265869744435284, "grad_norm": 0.3203362105742117, "learning_rate": 0.000108193771180895, "loss": 0.8631, "step": 2555 }, { "epoch": 0.5276174773289365, "grad_norm": 0.3176938770218858, "learning_rate": 0.00010783514942115398, "loss": 1.1615, "step": 2560 }, { "epoch": 0.5286479802143446, "grad_norm": 0.27188600955614695, "learning_rate": 0.0001074764262424639, "loss": 1.0099, "step": 2565 }, { "epoch": 0.5296784830997526, "grad_norm": 0.3225534680387049, "learning_rate": 0.00010711760628817304, "loss": 1.1351, "step": 2570 }, { "epoch": 0.5307089859851608, "grad_norm": 0.39558907715108127, "learning_rate": 0.0001067586942028824, "loss": 1.1706, "step": 2575 }, { "epoch": 0.5317394888705689, "grad_norm": 0.28753930251222504, "learning_rate": 0.00010639969463238553, "loss": 0.8745, "step": 2580 }, { "epoch": 0.532769991755977, "grad_norm": 0.30734397473616565, "learning_rate": 0.00010604061222360828, "loss": 1.1448, "step": 2585 }, { "epoch": 0.533800494641385, "grad_norm": 0.26804645458820414, "learning_rate": 0.00010568145162454896, "loss": 0.9725, "step": 2590 }, { "epoch": 0.5348309975267931, "grad_norm": 0.3263053082797193, "learning_rate": 0.00010532221748421787, "loss": 1.0632, "step": 2595 }, { "epoch": 0.5358615004122012, "grad_norm": 0.3650347794639528, "learning_rate": 0.00010496291445257725, "loss": 1.1795, "step": 2600 }, { "epoch": 0.5358615004122012, "eval_loss": 1.0102049112319946, "eval_runtime": 2464.7805, "eval_samples_per_second": 3.246, "eval_steps_per_second": 0.203, "step": 2600 }, { "epoch": 0.5368920032976092, "grad_norm": 0.299206263893403, "learning_rate": 0.00010460354718048109, "loss": 0.8798, "step": 2605 }, { "epoch": 0.5379225061830173, "grad_norm": 0.32212757306557593, "learning_rate": 0.00010424412031961484, "loss": 1.2768, "step": 2610 }, { "epoch": 0.5389530090684254, "grad_norm": 0.2686223265118172, "learning_rate": 0.0001038846385224354, "loss": 0.9964, "step": 2615 }, { "epoch": 0.5399835119538334, "grad_norm": 0.31440179887584824, "learning_rate": 0.00010352510644211074, "loss": 1.0418, "step": 2620 }, { "epoch": 0.5410140148392415, "grad_norm": 0.3262137336119927, "learning_rate": 0.0001031655287324596, "loss": 1.2262, "step": 2625 }, { "epoch": 0.5420445177246497, "grad_norm": 0.2977115868800609, "learning_rate": 0.00010280591004789144, "loss": 0.867, "step": 2630 }, { "epoch": 0.5430750206100577, "grad_norm": 0.3481967708447838, "learning_rate": 0.00010244625504334609, "loss": 1.1666, "step": 2635 }, { "epoch": 0.5441055234954658, "grad_norm": 0.27201591179231566, "learning_rate": 0.00010208656837423351, "loss": 0.9024, "step": 2640 }, { "epoch": 0.5451360263808739, "grad_norm": 0.3829985152487614, "learning_rate": 0.00010172685469637351, "loss": 1.0713, "step": 2645 }, { "epoch": 0.546166529266282, "grad_norm": 0.36090749295144076, "learning_rate": 0.00010136711866593551, "loss": 1.1479, "step": 2650 }, { "epoch": 0.54719703215169, "grad_norm": 0.300126161900349, "learning_rate": 0.00010100736493937828, "loss": 0.8899, "step": 2655 }, { "epoch": 0.5482275350370981, "grad_norm": 0.33344984792579047, "learning_rate": 0.00010064759817338965, "loss": 1.2477, "step": 2660 }, { "epoch": 0.5492580379225062, "grad_norm": 0.2615632901768006, "learning_rate": 0.00010028782302482617, "loss": 0.9453, "step": 2665 }, { "epoch": 0.5502885408079142, "grad_norm": 0.36602450939184455, "learning_rate": 9.992804415065305e-05, "loss": 1.0671, "step": 2670 }, { "epoch": 0.5513190436933223, "grad_norm": 0.34898414138240497, "learning_rate": 9.956826620788352e-05, "loss": 1.1969, "step": 2675 }, { "epoch": 0.5523495465787304, "grad_norm": 0.28926741748856266, "learning_rate": 9.920849385351889e-05, "loss": 0.8264, "step": 2680 }, { "epoch": 0.5533800494641385, "grad_norm": 0.3470665476375764, "learning_rate": 9.884873174448811e-05, "loss": 1.2393, "step": 2685 }, { "epoch": 0.5544105523495466, "grad_norm": 0.25248382277496384, "learning_rate": 9.848898453758752e-05, "loss": 0.9752, "step": 2690 }, { "epoch": 0.5554410552349547, "grad_norm": 0.322941880174542, "learning_rate": 9.812925688942054e-05, "loss": 1.0431, "step": 2695 }, { "epoch": 0.5564715581203628, "grad_norm": 0.5004506241518043, "learning_rate": 9.776955345633739e-05, "loss": 1.1417, "step": 2700 }, { "epoch": 0.5575020610057708, "grad_norm": 0.2897297596123754, "learning_rate": 9.740987889437492e-05, "loss": 0.9032, "step": 2705 }, { "epoch": 0.5585325638911789, "grad_norm": 0.3767677442786407, "learning_rate": 9.705023785919623e-05, "loss": 1.1532, "step": 2710 }, { "epoch": 0.559563066776587, "grad_norm": 0.25503990950782773, "learning_rate": 9.669063500603049e-05, "loss": 0.9414, "step": 2715 }, { "epoch": 0.560593569661995, "grad_norm": 0.34636076916747177, "learning_rate": 9.633107498961251e-05, "loss": 1.0689, "step": 2720 }, { "epoch": 0.5616240725474031, "grad_norm": 0.39789228504761426, "learning_rate": 9.597156246412277e-05, "loss": 1.1464, "step": 2725 }, { "epoch": 0.5626545754328112, "grad_norm": 0.289581783982976, "learning_rate": 9.561210208312694e-05, "loss": 0.818, "step": 2730 }, { "epoch": 0.5636850783182193, "grad_norm": 0.31157714978145534, "learning_rate": 9.525269849951578e-05, "loss": 1.2706, "step": 2735 }, { "epoch": 0.5647155812036274, "grad_norm": 0.28372271464928434, "learning_rate": 9.489335636544476e-05, "loss": 0.9805, "step": 2740 }, { "epoch": 0.5657460840890355, "grad_norm": 0.33997031384483695, "learning_rate": 9.453408033227406e-05, "loss": 1.0937, "step": 2745 }, { "epoch": 0.5667765869744436, "grad_norm": 0.33417374019033036, "learning_rate": 9.417487505050816e-05, "loss": 1.2517, "step": 2750 }, { "epoch": 0.5678070898598516, "grad_norm": 0.3379621672677856, "learning_rate": 9.38157451697358e-05, "loss": 0.8737, "step": 2755 }, { "epoch": 0.5688375927452597, "grad_norm": 0.3589988298317977, "learning_rate": 9.345669533856961e-05, "loss": 1.1273, "step": 2760 }, { "epoch": 0.5698680956306678, "grad_norm": 0.26916205391576065, "learning_rate": 9.309773020458616e-05, "loss": 0.9603, "step": 2765 }, { "epoch": 0.5708985985160758, "grad_norm": 0.3159986677198098, "learning_rate": 9.273885441426562e-05, "loss": 1.077, "step": 2770 }, { "epoch": 0.5719291014014839, "grad_norm": 0.3412008285509142, "learning_rate": 9.238007261293176e-05, "loss": 1.1973, "step": 2775 }, { "epoch": 0.572959604286892, "grad_norm": 0.28834466612116516, "learning_rate": 9.202138944469168e-05, "loss": 0.8306, "step": 2780 }, { "epoch": 0.5739901071723, "grad_norm": 0.38911649440218604, "learning_rate": 9.16628095523758e-05, "loss": 1.141, "step": 2785 }, { "epoch": 0.5750206100577082, "grad_norm": 0.29683818561664493, "learning_rate": 9.130433757747772e-05, "loss": 0.9729, "step": 2790 }, { "epoch": 0.5760511129431163, "grad_norm": 0.3445235007246779, "learning_rate": 9.094597816009409e-05, "loss": 0.9819, "step": 2795 }, { "epoch": 0.5770816158285244, "grad_norm": 0.3661330469833344, "learning_rate": 9.058773593886469e-05, "loss": 1.1984, "step": 2800 }, { "epoch": 0.5770816158285244, "eval_loss": 1.0007954835891724, "eval_runtime": 2470.6884, "eval_samples_per_second": 3.238, "eval_steps_per_second": 0.202, "step": 2800 }, { "epoch": 0.5781121187139324, "grad_norm": 0.32153311729783396, "learning_rate": 9.022961555091226e-05, "loss": 0.8678, "step": 2805 }, { "epoch": 0.5791426215993405, "grad_norm": 0.36605355908566417, "learning_rate": 8.987162163178256e-05, "loss": 1.2198, "step": 2810 }, { "epoch": 0.5801731244847486, "grad_norm": 0.2691696483789003, "learning_rate": 8.951375881538421e-05, "loss": 0.9082, "step": 2815 }, { "epoch": 0.5812036273701566, "grad_norm": 0.35640081108055943, "learning_rate": 8.915603173392895e-05, "loss": 1.0764, "step": 2820 }, { "epoch": 0.5822341302555647, "grad_norm": 0.33493558837647336, "learning_rate": 8.87984450178715e-05, "loss": 1.1807, "step": 2825 }, { "epoch": 0.5832646331409728, "grad_norm": 0.31748865735569, "learning_rate": 8.84410032958497e-05, "loss": 0.8788, "step": 2830 }, { "epoch": 0.5842951360263808, "grad_norm": 0.36563160381436116, "learning_rate": 8.808371119462452e-05, "loss": 1.1127, "step": 2835 }, { "epoch": 0.5853256389117889, "grad_norm": 0.26343660014410963, "learning_rate": 8.772657333902027e-05, "loss": 0.9977, "step": 2840 }, { "epoch": 0.5863561417971971, "grad_norm": 0.3695249008293323, "learning_rate": 8.736959435186466e-05, "loss": 1.095, "step": 2845 }, { "epoch": 0.5873866446826052, "grad_norm": 0.35773020515251225, "learning_rate": 8.701277885392909e-05, "loss": 1.2322, "step": 2850 }, { "epoch": 0.5884171475680132, "grad_norm": 0.3269028767537368, "learning_rate": 8.665613146386854e-05, "loss": 0.8593, "step": 2855 }, { "epoch": 0.5894476504534213, "grad_norm": 0.3544711574787662, "learning_rate": 8.629965679816217e-05, "loss": 1.1952, "step": 2860 }, { "epoch": 0.5904781533388294, "grad_norm": 0.29639050674293943, "learning_rate": 8.594335947105328e-05, "loss": 0.9516, "step": 2865 }, { "epoch": 0.5915086562242374, "grad_norm": 0.3426897718147058, "learning_rate": 8.558724409448974e-05, "loss": 1.0779, "step": 2870 }, { "epoch": 0.5925391591096455, "grad_norm": 0.3740367235927869, "learning_rate": 8.523131527806423e-05, "loss": 1.1865, "step": 2875 }, { "epoch": 0.5935696619950536, "grad_norm": 0.31189806294414407, "learning_rate": 8.48755776289545e-05, "loss": 0.8847, "step": 2880 }, { "epoch": 0.5946001648804616, "grad_norm": 0.37495206238491713, "learning_rate": 8.452003575186394e-05, "loss": 1.1804, "step": 2885 }, { "epoch": 0.5956306677658697, "grad_norm": 0.2797206274129371, "learning_rate": 8.416469424896167e-05, "loss": 0.9762, "step": 2890 }, { "epoch": 0.5966611706512778, "grad_norm": 0.33348859645525075, "learning_rate": 8.380955771982332e-05, "loss": 1.0883, "step": 2895 }, { "epoch": 0.597691673536686, "grad_norm": 0.44145796098661705, "learning_rate": 8.345463076137125e-05, "loss": 1.166, "step": 2900 }, { "epoch": 0.598722176422094, "grad_norm": 0.286114945740939, "learning_rate": 8.309991796781511e-05, "loss": 0.8535, "step": 2905 }, { "epoch": 0.5997526793075021, "grad_norm": 0.34343359665934886, "learning_rate": 8.274542393059235e-05, "loss": 1.1742, "step": 2910 }, { "epoch": 0.6007831821929102, "grad_norm": 0.2624287823547147, "learning_rate": 8.239115323830889e-05, "loss": 0.955, "step": 2915 }, { "epoch": 0.6018136850783182, "grad_norm": 0.35154822360336263, "learning_rate": 8.203711047667958e-05, "loss": 1.0426, "step": 2920 }, { "epoch": 0.6028441879637263, "grad_norm": 0.3344934169585091, "learning_rate": 8.1683300228469e-05, "loss": 1.1785, "step": 2925 }, { "epoch": 0.6038746908491344, "grad_norm": 0.3209999819274115, "learning_rate": 8.132972707343192e-05, "loss": 0.8842, "step": 2930 }, { "epoch": 0.6049051937345424, "grad_norm": 0.39454283654575434, "learning_rate": 8.097639558825427e-05, "loss": 1.1693, "step": 2935 }, { "epoch": 0.6059356966199505, "grad_norm": 0.2878286096920221, "learning_rate": 8.062331034649376e-05, "loss": 0.8965, "step": 2940 }, { "epoch": 0.6069661995053586, "grad_norm": 0.33688522439048935, "learning_rate": 8.027047591852069e-05, "loss": 1.1381, "step": 2945 }, { "epoch": 0.6079967023907666, "grad_norm": 0.38189870355144323, "learning_rate": 7.991789687145873e-05, "loss": 1.1257, "step": 2950 }, { "epoch": 0.6090272052761748, "grad_norm": 0.3051536804305413, "learning_rate": 7.956557776912596e-05, "loss": 0.9044, "step": 2955 }, { "epoch": 0.6100577081615829, "grad_norm": 0.3451327149435641, "learning_rate": 7.921352317197574e-05, "loss": 1.2264, "step": 2960 }, { "epoch": 0.611088211046991, "grad_norm": 0.2827004152527063, "learning_rate": 7.886173763703757e-05, "loss": 0.9794, "step": 2965 }, { "epoch": 0.612118713932399, "grad_norm": 0.40169478150821714, "learning_rate": 7.851022571785819e-05, "loss": 1.1179, "step": 2970 }, { "epoch": 0.6131492168178071, "grad_norm": 0.4438701028140135, "learning_rate": 7.815899196444267e-05, "loss": 1.1662, "step": 2975 }, { "epoch": 0.6141797197032152, "grad_norm": 0.3254217005122457, "learning_rate": 7.780804092319547e-05, "loss": 0.8434, "step": 2980 }, { "epoch": 0.6152102225886232, "grad_norm": 0.3284145654312824, "learning_rate": 7.745737713686152e-05, "loss": 1.2109, "step": 2985 }, { "epoch": 0.6162407254740313, "grad_norm": 0.28117376745819467, "learning_rate": 7.710700514446762e-05, "loss": 0.955, "step": 2990 }, { "epoch": 0.6172712283594394, "grad_norm": 0.35764651325723473, "learning_rate": 7.675692948126345e-05, "loss": 1.032, "step": 2995 }, { "epoch": 0.6183017312448474, "grad_norm": 0.4045276967582808, "learning_rate": 7.640715467866307e-05, "loss": 1.157, "step": 3000 }, { "epoch": 0.6183017312448474, "eval_loss": 0.9929932951927185, "eval_runtime": 2471.8115, "eval_samples_per_second": 3.236, "eval_steps_per_second": 0.202, "step": 3000 }, { "epoch": 0.6193322341302555, "grad_norm": 0.3148405737434604, "learning_rate": 7.605768526418605e-05, "loss": 0.8743, "step": 3005 }, { "epoch": 0.6203627370156637, "grad_norm": 0.35877325985884195, "learning_rate": 7.57085257613991e-05, "loss": 1.2573, "step": 3010 }, { "epoch": 0.6213932399010718, "grad_norm": 0.2728312132512685, "learning_rate": 7.535968068985737e-05, "loss": 0.9663, "step": 3015 }, { "epoch": 0.6224237427864798, "grad_norm": 0.3236792895371986, "learning_rate": 7.501115456504595e-05, "loss": 1.1109, "step": 3020 }, { "epoch": 0.6234542456718879, "grad_norm": 0.40547856146336464, "learning_rate": 7.466295189832148e-05, "loss": 1.1882, "step": 3025 }, { "epoch": 0.624484748557296, "grad_norm": 0.26523329910607873, "learning_rate": 7.431507719685371e-05, "loss": 0.8277, "step": 3030 }, { "epoch": 0.625515251442704, "grad_norm": 0.3252197516332192, "learning_rate": 7.396753496356718e-05, "loss": 1.2303, "step": 3035 }, { "epoch": 0.6265457543281121, "grad_norm": 0.2664136991956755, "learning_rate": 7.362032969708297e-05, "loss": 0.9909, "step": 3040 }, { "epoch": 0.6275762572135202, "grad_norm": 0.3438706008672591, "learning_rate": 7.327346589166035e-05, "loss": 1.057, "step": 3045 }, { "epoch": 0.6286067600989282, "grad_norm": 0.40393507740091983, "learning_rate": 7.292694803713871e-05, "loss": 1.1674, "step": 3050 }, { "epoch": 0.6296372629843363, "grad_norm": 0.2966521374455327, "learning_rate": 7.258078061887947e-05, "loss": 0.8634, "step": 3055 }, { "epoch": 0.6306677658697445, "grad_norm": 0.3313252298668368, "learning_rate": 7.223496811770796e-05, "loss": 1.1705, "step": 3060 }, { "epoch": 0.6316982687551526, "grad_norm": 0.282517205910409, "learning_rate": 7.188951500985533e-05, "loss": 0.9568, "step": 3065 }, { "epoch": 0.6327287716405606, "grad_norm": 0.3423739475910178, "learning_rate": 7.154442576690083e-05, "loss": 1.1215, "step": 3070 }, { "epoch": 0.6337592745259687, "grad_norm": 0.3951313592694174, "learning_rate": 7.119970485571375e-05, "loss": 1.2196, "step": 3075 }, { "epoch": 0.6347897774113768, "grad_norm": 0.27968067577198935, "learning_rate": 7.08553567383956e-05, "loss": 0.8517, "step": 3080 }, { "epoch": 0.6358202802967848, "grad_norm": 0.3967375078625761, "learning_rate": 7.051138587222255e-05, "loss": 1.1345, "step": 3085 }, { "epoch": 0.6368507831821929, "grad_norm": 0.275638071848918, "learning_rate": 7.016779670958746e-05, "loss": 0.9428, "step": 3090 }, { "epoch": 0.637881286067601, "grad_norm": 0.3876563491000536, "learning_rate": 6.982459369794247e-05, "loss": 1.0172, "step": 3095 }, { "epoch": 0.638911788953009, "grad_norm": 0.39932395010881844, "learning_rate": 6.948178127974127e-05, "loss": 1.136, "step": 3100 }, { "epoch": 0.6399422918384171, "grad_norm": 0.344145718815036, "learning_rate": 6.913936389238174e-05, "loss": 0.9074, "step": 3105 }, { "epoch": 0.6409727947238252, "grad_norm": 0.35702868127871346, "learning_rate": 6.879734596814839e-05, "loss": 1.2245, "step": 3110 }, { "epoch": 0.6420032976092334, "grad_norm": 0.278189478468898, "learning_rate": 6.845573193415508e-05, "loss": 0.9795, "step": 3115 }, { "epoch": 0.6430338004946414, "grad_norm": 0.3820554155556976, "learning_rate": 6.811452621228766e-05, "loss": 1.0815, "step": 3120 }, { "epoch": 0.6440643033800495, "grad_norm": 0.42159702416774886, "learning_rate": 6.777373321914671e-05, "loss": 1.2058, "step": 3125 }, { "epoch": 0.6450948062654576, "grad_norm": 0.30363813905867076, "learning_rate": 6.743335736599045e-05, "loss": 0.853, "step": 3130 }, { "epoch": 0.6461253091508656, "grad_norm": 0.3626110839853577, "learning_rate": 6.709340305867762e-05, "loss": 1.1719, "step": 3135 }, { "epoch": 0.6471558120362737, "grad_norm": 0.2741980722413254, "learning_rate": 6.675387469761033e-05, "loss": 0.9857, "step": 3140 }, { "epoch": 0.6481863149216818, "grad_norm": 0.3458024372951965, "learning_rate": 6.641477667767738e-05, "loss": 1.0737, "step": 3145 }, { "epoch": 0.6492168178070898, "grad_norm": 0.3659993842021884, "learning_rate": 6.607611338819697e-05, "loss": 1.2332, "step": 3150 }, { "epoch": 0.6502473206924979, "grad_norm": 0.3134224056811165, "learning_rate": 6.573788921286028e-05, "loss": 0.833, "step": 3155 }, { "epoch": 0.651277823577906, "grad_norm": 0.31557728275482577, "learning_rate": 6.540010852967447e-05, "loss": 1.169, "step": 3160 }, { "epoch": 0.652308326463314, "grad_norm": 0.2745096877567603, "learning_rate": 6.506277571090613e-05, "loss": 0.9294, "step": 3165 }, { "epoch": 0.6533388293487222, "grad_norm": 0.37277814205936943, "learning_rate": 6.47258951230246e-05, "loss": 1.081, "step": 3170 }, { "epoch": 0.6543693322341303, "grad_norm": 0.38045711766200485, "learning_rate": 6.438947112664555e-05, "loss": 1.1493, "step": 3175 }, { "epoch": 0.6553998351195384, "grad_norm": 0.2706340588373179, "learning_rate": 6.405350807647444e-05, "loss": 0.8498, "step": 3180 }, { "epoch": 0.6564303380049464, "grad_norm": 0.3336018890725225, "learning_rate": 6.371801032125026e-05, "loss": 1.1778, "step": 3185 }, { "epoch": 0.6574608408903545, "grad_norm": 0.2913475821459678, "learning_rate": 6.338298220368912e-05, "loss": 0.9571, "step": 3190 }, { "epoch": 0.6584913437757626, "grad_norm": 0.36985494968453825, "learning_rate": 6.304842806042812e-05, "loss": 1.0778, "step": 3195 }, { "epoch": 0.6595218466611706, "grad_norm": 0.40407830600531813, "learning_rate": 6.271435222196916e-05, "loss": 1.1542, "step": 3200 }, { "epoch": 0.6595218466611706, "eval_loss": 0.9862346649169922, "eval_runtime": 2465.9653, "eval_samples_per_second": 3.244, "eval_steps_per_second": 0.203, "step": 3200 }, { "epoch": 0.6605523495465787, "grad_norm": 0.2524320710636017, "learning_rate": 6.238075901262293e-05, "loss": 0.8157, "step": 3205 }, { "epoch": 0.6615828524319868, "grad_norm": 0.3278601577235125, "learning_rate": 6.204765275045298e-05, "loss": 1.1461, "step": 3210 }, { "epoch": 0.6626133553173948, "grad_norm": 0.28938045624305053, "learning_rate": 6.171503774721966e-05, "loss": 0.9831, "step": 3215 }, { "epoch": 0.6636438582028029, "grad_norm": 0.3531127055799286, "learning_rate": 6.13829183083245e-05, "loss": 1.0438, "step": 3220 }, { "epoch": 0.6646743610882111, "grad_norm": 0.3337288007445976, "learning_rate": 6.105129873275435e-05, "loss": 1.2684, "step": 3225 }, { "epoch": 0.6657048639736192, "grad_norm": 0.30432663000679183, "learning_rate": 6.072018331302577e-05, "loss": 0.8405, "step": 3230 }, { "epoch": 0.6667353668590272, "grad_norm": 0.3481398036760903, "learning_rate": 6.038957633512957e-05, "loss": 1.1838, "step": 3235 }, { "epoch": 0.6677658697444353, "grad_norm": 0.27336340413608334, "learning_rate": 6.005948207847516e-05, "loss": 0.9207, "step": 3240 }, { "epoch": 0.6687963726298434, "grad_norm": 0.3440287779317674, "learning_rate": 5.97299048158352e-05, "loss": 1.0703, "step": 3245 }, { "epoch": 0.6698268755152514, "grad_norm": 0.39581616124415014, "learning_rate": 5.940084881329042e-05, "loss": 1.2125, "step": 3250 }, { "epoch": 0.6708573784006595, "grad_norm": 0.2935018842179493, "learning_rate": 5.907231833017424e-05, "loss": 0.8331, "step": 3255 }, { "epoch": 0.6718878812860676, "grad_norm": 0.3836289298282413, "learning_rate": 5.8744317619017755e-05, "loss": 1.2524, "step": 3260 }, { "epoch": 0.6729183841714756, "grad_norm": 0.2710041280727128, "learning_rate": 5.841685092549456e-05, "loss": 0.9405, "step": 3265 }, { "epoch": 0.6739488870568837, "grad_norm": 0.32969149065243836, "learning_rate": 5.8089922488365975e-05, "loss": 1.0393, "step": 3270 }, { "epoch": 0.6749793899422918, "grad_norm": 0.336443888291771, "learning_rate": 5.776353653942602e-05, "loss": 1.2373, "step": 3275 }, { "epoch": 0.6760098928277, "grad_norm": 0.30665104617457667, "learning_rate": 5.743769730344666e-05, "loss": 0.8681, "step": 3280 }, { "epoch": 0.677040395713108, "grad_norm": 0.3625802330226177, "learning_rate": 5.7112408998123256e-05, "loss": 1.2316, "step": 3285 }, { "epoch": 0.6780708985985161, "grad_norm": 0.2953674696158516, "learning_rate": 5.678767583401974e-05, "loss": 0.9204, "step": 3290 }, { "epoch": 0.6791014014839242, "grad_norm": 0.38225007623929924, "learning_rate": 5.646350201451438e-05, "loss": 1.066, "step": 3295 }, { "epoch": 0.6801319043693322, "grad_norm": 0.4406303775855105, "learning_rate": 5.613989173574512e-05, "loss": 1.1569, "step": 3300 }, { "epoch": 0.6811624072547403, "grad_norm": 0.2882802372851351, "learning_rate": 5.5816849186555386e-05, "loss": 0.86, "step": 3305 }, { "epoch": 0.6821929101401484, "grad_norm": 0.3780252315323207, "learning_rate": 5.549437854843995e-05, "loss": 1.1646, "step": 3310 }, { "epoch": 0.6832234130255564, "grad_norm": 0.2834828395630223, "learning_rate": 5.517248399549063e-05, "loss": 0.973, "step": 3315 }, { "epoch": 0.6842539159109645, "grad_norm": 0.3721150712296605, "learning_rate": 5.48511696943423e-05, "loss": 1.0507, "step": 3320 }, { "epoch": 0.6852844187963726, "grad_norm": 0.3748866522580312, "learning_rate": 5.4530439804119096e-05, "loss": 1.1834, "step": 3325 }, { "epoch": 0.6863149216817807, "grad_norm": 0.2887534766618149, "learning_rate": 5.4210298476380484e-05, "loss": 0.888, "step": 3330 }, { "epoch": 0.6873454245671888, "grad_norm": 0.3409152912595764, "learning_rate": 5.38907498550674e-05, "loss": 1.1606, "step": 3335 }, { "epoch": 0.6883759274525969, "grad_norm": 0.29326388422606586, "learning_rate": 5.357179807644887e-05, "loss": 0.8845, "step": 3340 }, { "epoch": 0.689406430338005, "grad_norm": 0.3589561770883435, "learning_rate": 5.3253447269068245e-05, "loss": 1.0743, "step": 3345 }, { "epoch": 0.690436933223413, "grad_norm": 0.38383916804262924, "learning_rate": 5.293570155368981e-05, "loss": 1.183, "step": 3350 }, { "epoch": 0.6914674361088211, "grad_norm": 0.3068092183719897, "learning_rate": 5.261856504324563e-05, "loss": 0.881, "step": 3355 }, { "epoch": 0.6924979389942292, "grad_norm": 0.36385771675879847, "learning_rate": 5.230204184278195e-05, "loss": 1.2467, "step": 3360 }, { "epoch": 0.6935284418796372, "grad_norm": 0.2907466758494882, "learning_rate": 5.198613604940649e-05, "loss": 0.9442, "step": 3365 }, { "epoch": 0.6945589447650453, "grad_norm": 0.3566322978977828, "learning_rate": 5.1670851752235025e-05, "loss": 1.0632, "step": 3370 }, { "epoch": 0.6955894476504534, "grad_norm": 0.36835384306712526, "learning_rate": 5.135619303233867e-05, "loss": 1.1819, "step": 3375 }, { "epoch": 0.6966199505358615, "grad_norm": 0.30530163857595344, "learning_rate": 5.104216396269109e-05, "loss": 0.8697, "step": 3380 }, { "epoch": 0.6976504534212696, "grad_norm": 0.3728398219157709, "learning_rate": 5.072876860811553e-05, "loss": 1.2065, "step": 3385 }, { "epoch": 0.6986809563066777, "grad_norm": 0.2719186126381508, "learning_rate": 5.0416011025232546e-05, "loss": 0.9332, "step": 3390 }, { "epoch": 0.6997114591920858, "grad_norm": 0.39560521280304406, "learning_rate": 5.010389526240719e-05, "loss": 1.0035, "step": 3395 }, { "epoch": 0.7007419620774938, "grad_norm": 0.45096201363438326, "learning_rate": 4.979242535969672e-05, "loss": 1.1648, "step": 3400 }, { "epoch": 0.7007419620774938, "eval_loss": 0.9801868796348572, "eval_runtime": 2471.0518, "eval_samples_per_second": 3.237, "eval_steps_per_second": 0.202, "step": 3400 }, { "epoch": 0.7017724649629019, "grad_norm": 0.3253388221391844, "learning_rate": 4.9481605348798435e-05, "loss": 0.8283, "step": 3405 }, { "epoch": 0.70280296784831, "grad_norm": 0.3382283949543721, "learning_rate": 4.917143925299728e-05, "loss": 1.1642, "step": 3410 }, { "epoch": 0.703833470733718, "grad_norm": 0.2650304501518648, "learning_rate": 4.886193108711378e-05, "loss": 0.9334, "step": 3415 }, { "epoch": 0.7048639736191261, "grad_norm": 0.3163351222719128, "learning_rate": 4.8553084857452426e-05, "loss": 1.0316, "step": 3420 }, { "epoch": 0.7058944765045342, "grad_norm": 0.4184991045727221, "learning_rate": 4.824490456174926e-05, "loss": 1.1313, "step": 3425 }, { "epoch": 0.7069249793899423, "grad_norm": 0.31621686065106913, "learning_rate": 4.7937394189120485e-05, "loss": 0.8658, "step": 3430 }, { "epoch": 0.7079554822753503, "grad_norm": 0.3252525212780396, "learning_rate": 4.763055772001086e-05, "loss": 1.1749, "step": 3435 }, { "epoch": 0.7089859851607585, "grad_norm": 0.28077212260765644, "learning_rate": 4.732439912614195e-05, "loss": 0.9195, "step": 3440 }, { "epoch": 0.7100164880461666, "grad_norm": 0.3404715984560569, "learning_rate": 4.7018922370460835e-05, "loss": 1.1004, "step": 3445 }, { "epoch": 0.7110469909315746, "grad_norm": 0.3572114714960076, "learning_rate": 4.671413140708893e-05, "loss": 1.1555, "step": 3450 }, { "epoch": 0.7120774938169827, "grad_norm": 0.31012926926152484, "learning_rate": 4.6410030181270546e-05, "loss": 0.8328, "step": 3455 }, { "epoch": 0.7131079967023908, "grad_norm": 0.3814244438557788, "learning_rate": 4.610662262932209e-05, "loss": 1.2221, "step": 3460 }, { "epoch": 0.7141384995877988, "grad_norm": 0.2587030449835817, "learning_rate": 4.5803912678580906e-05, "loss": 0.923, "step": 3465 }, { "epoch": 0.7151690024732069, "grad_norm": 0.36964920266690043, "learning_rate": 4.5501904247354474e-05, "loss": 1.0765, "step": 3470 }, { "epoch": 0.716199505358615, "grad_norm": 0.4007947343253882, "learning_rate": 4.520060124486989e-05, "loss": 1.0613, "step": 3475 }, { "epoch": 0.717230008244023, "grad_norm": 0.31355619965035286, "learning_rate": 4.4900007571222946e-05, "loss": 0.8406, "step": 3480 }, { "epoch": 0.7182605111294311, "grad_norm": 0.3709230553408382, "learning_rate": 4.460012711732795e-05, "loss": 1.202, "step": 3485 }, { "epoch": 0.7192910140148392, "grad_norm": 0.2663391335521852, "learning_rate": 4.430096376486713e-05, "loss": 0.957, "step": 3490 }, { "epoch": 0.7203215169002474, "grad_norm": 0.36149156644544367, "learning_rate": 4.4002521386240466e-05, "loss": 1.0885, "step": 3495 }, { "epoch": 0.7213520197856554, "grad_norm": 0.4116390911425264, "learning_rate": 4.3704803844515705e-05, "loss": 1.1649, "step": 3500 }, { "epoch": 0.7223825226710635, "grad_norm": 0.3337684750408035, "learning_rate": 4.3407814993378095e-05, "loss": 0.8993, "step": 3505 }, { "epoch": 0.7234130255564716, "grad_norm": 0.3496736243866778, "learning_rate": 4.311155867708071e-05, "loss": 1.1648, "step": 3510 }, { "epoch": 0.7244435284418796, "grad_norm": 0.28479774144881315, "learning_rate": 4.2816038730394656e-05, "loss": 0.9484, "step": 3515 }, { "epoch": 0.7254740313272877, "grad_norm": 0.36918418554896276, "learning_rate": 4.252125897855932e-05, "loss": 1.0583, "step": 3520 }, { "epoch": 0.7265045342126958, "grad_norm": 0.33704337919665334, "learning_rate": 4.222722323723294e-05, "loss": 1.1907, "step": 3525 }, { "epoch": 0.7275350370981039, "grad_norm": 0.3148307470632366, "learning_rate": 4.1933935312443286e-05, "loss": 0.8783, "step": 3530 }, { "epoch": 0.7285655399835119, "grad_norm": 0.3517810748272734, "learning_rate": 4.164139900053824e-05, "loss": 1.1717, "step": 3535 }, { "epoch": 0.72959604286892, "grad_norm": 0.2855842847206538, "learning_rate": 4.134961808813672e-05, "loss": 0.9112, "step": 3540 }, { "epoch": 0.7306265457543281, "grad_norm": 0.34108259740221386, "learning_rate": 4.1058596352079805e-05, "loss": 1.0142, "step": 3545 }, { "epoch": 0.7316570486397362, "grad_norm": 0.45538302782398454, "learning_rate": 4.076833755938153e-05, "loss": 1.0974, "step": 3550 }, { "epoch": 0.7326875515251443, "grad_norm": 0.31487990147891376, "learning_rate": 4.0478845467180506e-05, "loss": 0.8298, "step": 3555 }, { "epoch": 0.7337180544105524, "grad_norm": 0.36935273464286145, "learning_rate": 4.0190123822690965e-05, "loss": 1.1436, "step": 3560 }, { "epoch": 0.7347485572959604, "grad_norm": 0.27930316978222564, "learning_rate": 3.990217636315441e-05, "loss": 0.9989, "step": 3565 }, { "epoch": 0.7357790601813685, "grad_norm": 0.3714521960753657, "learning_rate": 3.961500681579129e-05, "loss": 1.1123, "step": 3570 }, { "epoch": 0.7368095630667766, "grad_norm": 0.43314908676280006, "learning_rate": 3.9328618897752566e-05, "loss": 1.1162, "step": 3575 }, { "epoch": 0.7378400659521847, "grad_norm": 0.29947090161111584, "learning_rate": 3.904301631607186e-05, "loss": 0.8351, "step": 3580 }, { "epoch": 0.7388705688375927, "grad_norm": 0.4143197770113969, "learning_rate": 3.875820276761717e-05, "loss": 1.2288, "step": 3585 }, { "epoch": 0.7399010717230008, "grad_norm": 0.2821180575460458, "learning_rate": 3.847418193904324e-05, "loss": 0.9842, "step": 3590 }, { "epoch": 0.7409315746084089, "grad_norm": 0.35222923714451126, "learning_rate": 3.8190957506743806e-05, "loss": 1.0377, "step": 3595 }, { "epoch": 0.7419620774938169, "grad_norm": 0.36695400370029363, "learning_rate": 3.790853313680389e-05, "loss": 1.1403, "step": 3600 }, { "epoch": 0.7419620774938169, "eval_loss": 0.9750496745109558, "eval_runtime": 2464.5637, "eval_samples_per_second": 3.246, "eval_steps_per_second": 0.203, "step": 3600 }, { "epoch": 0.7429925803792251, "grad_norm": 0.33168677598566876, "learning_rate": 3.7626912484952495e-05, "loss": 0.9223, "step": 3605 }, { "epoch": 0.7440230832646332, "grad_norm": 0.3685099675858659, "learning_rate": 3.734609919651523e-05, "loss": 1.22, "step": 3610 }, { "epoch": 0.7450535861500412, "grad_norm": 0.30519501707613905, "learning_rate": 3.706609690636703e-05, "loss": 0.9541, "step": 3615 }, { "epoch": 0.7460840890354493, "grad_norm": 0.3227606827521623, "learning_rate": 3.6786909238885215e-05, "loss": 1.002, "step": 3620 }, { "epoch": 0.7471145919208574, "grad_norm": 0.3907472086101187, "learning_rate": 3.650853980790262e-05, "loss": 1.1461, "step": 3625 }, { "epoch": 0.7481450948062655, "grad_norm": 0.2919486775395443, "learning_rate": 3.6230992216660664e-05, "loss": 0.8063, "step": 3630 }, { "epoch": 0.7491755976916735, "grad_norm": 0.32875886981694724, "learning_rate": 3.595427005776281e-05, "loss": 1.1514, "step": 3635 }, { "epoch": 0.7502061005770816, "grad_norm": 0.29158449116331836, "learning_rate": 3.5678376913128075e-05, "loss": 0.9759, "step": 3640 }, { "epoch": 0.7512366034624897, "grad_norm": 0.37631936130463955, "learning_rate": 3.540331635394458e-05, "loss": 0.9905, "step": 3645 }, { "epoch": 0.7522671063478977, "grad_norm": 0.3439508621054155, "learning_rate": 3.512909194062347e-05, "loss": 1.1949, "step": 3650 }, { "epoch": 0.7532976092333058, "grad_norm": 0.3146397920651041, "learning_rate": 3.485570722275264e-05, "loss": 0.8706, "step": 3655 }, { "epoch": 0.754328112118714, "grad_norm": 0.38258052872478915, "learning_rate": 3.458316573905087e-05, "loss": 1.2172, "step": 3660 }, { "epoch": 0.755358615004122, "grad_norm": 0.2793437890823054, "learning_rate": 3.4311471017322175e-05, "loss": 0.9391, "step": 3665 }, { "epoch": 0.7563891178895301, "grad_norm": 0.3643128847802364, "learning_rate": 3.4040626574409815e-05, "loss": 0.9938, "step": 3670 }, { "epoch": 0.7574196207749382, "grad_norm": 0.3709320339736749, "learning_rate": 3.377063591615113e-05, "loss": 1.1904, "step": 3675 }, { "epoch": 0.7584501236603463, "grad_norm": 0.33437847713940755, "learning_rate": 3.350150253733186e-05, "loss": 0.9233, "step": 3680 }, { "epoch": 0.7594806265457543, "grad_norm": 0.3607224445907033, "learning_rate": 3.3233229921641064e-05, "loss": 1.2125, "step": 3685 }, { "epoch": 0.7605111294311624, "grad_norm": 0.27198443746283474, "learning_rate": 3.296582154162604e-05, "loss": 0.9196, "step": 3690 }, { "epoch": 0.7615416323165705, "grad_norm": 0.3668276085455235, "learning_rate": 3.2699280858647327e-05, "loss": 1.0335, "step": 3695 }, { "epoch": 0.7625721352019785, "grad_norm": 0.431636600953154, "learning_rate": 3.2433611322833845e-05, "loss": 1.1378, "step": 3700 }, { "epoch": 0.7636026380873866, "grad_norm": 0.3284008231342369, "learning_rate": 3.216881637303839e-05, "loss": 0.8308, "step": 3705 }, { "epoch": 0.7646331409727948, "grad_norm": 0.3443882838132438, "learning_rate": 3.190489943679297e-05, "loss": 1.1797, "step": 3710 }, { "epoch": 0.7656636438582028, "grad_norm": 0.30472428853778355, "learning_rate": 3.164186393026445e-05, "loss": 0.8953, "step": 3715 }, { "epoch": 0.7666941467436109, "grad_norm": 0.346855761423586, "learning_rate": 3.137971325821054e-05, "loss": 1.1678, "step": 3720 }, { "epoch": 0.767724649629019, "grad_norm": 0.3586318210457773, "learning_rate": 3.111845081393542e-05, "loss": 1.1801, "step": 3725 }, { "epoch": 0.768755152514427, "grad_norm": 0.3012111522622655, "learning_rate": 3.0858079979245965e-05, "loss": 0.8259, "step": 3730 }, { "epoch": 0.7697856553998351, "grad_norm": 0.3675771233203778, "learning_rate": 3.059860412440811e-05, "loss": 1.1379, "step": 3735 }, { "epoch": 0.7708161582852432, "grad_norm": 0.2997972891151249, "learning_rate": 3.0340026608102902e-05, "loss": 0.9012, "step": 3740 }, { "epoch": 0.7718466611706513, "grad_norm": 0.3350924890911528, "learning_rate": 3.008235077738334e-05, "loss": 1.0476, "step": 3745 }, { "epoch": 0.7728771640560593, "grad_norm": 0.42578144522976874, "learning_rate": 2.9825579967630846e-05, "loss": 1.2014, "step": 3750 }, { "epoch": 0.7739076669414674, "grad_norm": 0.31905668249953223, "learning_rate": 2.956971750251215e-05, "loss": 0.9089, "step": 3755 }, { "epoch": 0.7749381698268755, "grad_norm": 0.3320270260748608, "learning_rate": 2.9314766693936356e-05, "loss": 1.1712, "step": 3760 }, { "epoch": 0.7759686727122836, "grad_norm": 0.26880506131338233, "learning_rate": 2.906073084201191e-05, "loss": 0.9532, "step": 3765 }, { "epoch": 0.7769991755976917, "grad_norm": 0.360515318139803, "learning_rate": 2.8807613235004037e-05, "loss": 1.0235, "step": 3770 }, { "epoch": 0.7780296784830998, "grad_norm": 0.40012910311865413, "learning_rate": 2.855541714929206e-05, "loss": 1.1292, "step": 3775 }, { "epoch": 0.7790601813685079, "grad_norm": 0.3056042121876613, "learning_rate": 2.8304145849327036e-05, "loss": 0.7814, "step": 3780 }, { "epoch": 0.7800906842539159, "grad_norm": 0.3475474680120366, "learning_rate": 2.8053802587589538e-05, "loss": 1.1668, "step": 3785 }, { "epoch": 0.781121187139324, "grad_norm": 0.2843659913958508, "learning_rate": 2.7804390604547557e-05, "loss": 0.9237, "step": 3790 }, { "epoch": 0.7821516900247321, "grad_norm": 0.36497922605390365, "learning_rate": 2.7555913128614398e-05, "loss": 1.0276, "step": 3795 }, { "epoch": 0.7831821929101401, "grad_norm": 0.39023528771913, "learning_rate": 2.7308373376107142e-05, "loss": 1.1268, "step": 3800 }, { "epoch": 0.7831821929101401, "eval_loss": 0.9705116748809814, "eval_runtime": 2662.1546, "eval_samples_per_second": 3.005, "eval_steps_per_second": 0.188, "step": 3800 }, { "epoch": 0.7842126957955482, "grad_norm": 0.3292637316481058, "learning_rate": 2.706177455120482e-05, "loss": 0.8586, "step": 3805 }, { "epoch": 0.7852431986809563, "grad_norm": 0.3463786306953725, "learning_rate": 2.681611984590696e-05, "loss": 1.1979, "step": 3810 }, { "epoch": 0.7862737015663643, "grad_norm": 0.3033390300653325, "learning_rate": 2.6571412439992437e-05, "loss": 0.9448, "step": 3815 }, { "epoch": 0.7873042044517725, "grad_norm": 0.3744585100233523, "learning_rate": 2.6327655500978076e-05, "loss": 1.0825, "step": 3820 }, { "epoch": 0.7883347073371806, "grad_norm": 0.3888535219196645, "learning_rate": 2.608485218407779e-05, "loss": 1.1602, "step": 3825 }, { "epoch": 0.7893652102225887, "grad_norm": 0.3281836015354183, "learning_rate": 2.5843005632161787e-05, "loss": 0.8071, "step": 3830 }, { "epoch": 0.7903957131079967, "grad_norm": 0.44764397719301374, "learning_rate": 2.5602118975715683e-05, "loss": 1.223, "step": 3835 }, { "epoch": 0.7914262159934048, "grad_norm": 0.2605243101554053, "learning_rate": 2.5362195332800253e-05, "loss": 0.9269, "step": 3840 }, { "epoch": 0.7924567188788129, "grad_norm": 0.34490759623743283, "learning_rate": 2.5123237809010836e-05, "loss": 0.9951, "step": 3845 }, { "epoch": 0.7934872217642209, "grad_norm": 0.45219071339330563, "learning_rate": 2.4885249497437223e-05, "loss": 1.1591, "step": 3850 }, { "epoch": 0.794517724649629, "grad_norm": 0.3701824230659728, "learning_rate": 2.4648233478623705e-05, "loss": 0.9237, "step": 3855 }, { "epoch": 0.7955482275350371, "grad_norm": 0.3540574000256143, "learning_rate": 2.4412192820529034e-05, "loss": 1.1936, "step": 3860 }, { "epoch": 0.7965787304204451, "grad_norm": 0.2979316113245556, "learning_rate": 2.4177130578486885e-05, "loss": 0.8818, "step": 3865 }, { "epoch": 0.7976092333058532, "grad_norm": 0.3972288442858004, "learning_rate": 2.3943049795166126e-05, "loss": 1.0352, "step": 3870 }, { "epoch": 0.7986397361912614, "grad_norm": 0.43218350490127366, "learning_rate": 2.370995350053157e-05, "loss": 1.141, "step": 3875 }, { "epoch": 0.7996702390766695, "grad_norm": 0.33536259992407524, "learning_rate": 2.3477844711804708e-05, "loss": 0.8845, "step": 3880 }, { "epoch": 0.8007007419620775, "grad_norm": 0.3715539599255406, "learning_rate": 2.3246726433424716e-05, "loss": 1.1807, "step": 3885 }, { "epoch": 0.8017312448474856, "grad_norm": 0.27142620043338195, "learning_rate": 2.301660165700936e-05, "loss": 0.9382, "step": 3890 }, { "epoch": 0.8027617477328937, "grad_norm": 0.3200117930180246, "learning_rate": 2.2787473361316592e-05, "loss": 1.1648, "step": 3895 }, { "epoch": 0.8037922506183017, "grad_norm": 0.3832079088616207, "learning_rate": 2.2559344512205705e-05, "loss": 1.1815, "step": 3900 }, { "epoch": 0.8048227535037098, "grad_norm": 0.28512724726672417, "learning_rate": 2.233221806259903e-05, "loss": 0.7237, "step": 3905 }, { "epoch": 0.8058532563891179, "grad_norm": 0.347779304749182, "learning_rate": 2.2106096952443888e-05, "loss": 1.1468, "step": 3910 }, { "epoch": 0.8068837592745259, "grad_norm": 0.27344000922062256, "learning_rate": 2.188098410867424e-05, "loss": 0.967, "step": 3915 }, { "epoch": 0.807914262159934, "grad_norm": 0.35776849097052316, "learning_rate": 2.165688244517299e-05, "loss": 1.0351, "step": 3920 }, { "epoch": 0.8089447650453421, "grad_norm": 0.6146898482449163, "learning_rate": 2.143379486273428e-05, "loss": 1.1856, "step": 3925 }, { "epoch": 0.8099752679307503, "grad_norm": 0.37385831712675915, "learning_rate": 2.1211724249025787e-05, "loss": 0.9441, "step": 3930 }, { "epoch": 0.8110057708161583, "grad_norm": 0.33568983400788555, "learning_rate": 2.099067347855157e-05, "loss": 1.2175, "step": 3935 }, { "epoch": 0.8120362737015664, "grad_norm": 0.2905285106790405, "learning_rate": 2.077064541261462e-05, "loss": 0.9725, "step": 3940 }, { "epoch": 0.8130667765869745, "grad_norm": 0.3292103830025177, "learning_rate": 2.0551642899279975e-05, "loss": 1.0363, "step": 3945 }, { "epoch": 0.8140972794723825, "grad_norm": 0.4380964265869986, "learning_rate": 2.0333668773337866e-05, "loss": 1.1014, "step": 3950 }, { "epoch": 0.8151277823577906, "grad_norm": 0.3297438808755647, "learning_rate": 2.0116725856266926e-05, "loss": 0.8749, "step": 3955 }, { "epoch": 0.8161582852431987, "grad_norm": 0.3404142147430593, "learning_rate": 1.9900816956197698e-05, "loss": 1.2154, "step": 3960 }, { "epoch": 0.8171887881286067, "grad_norm": 0.28990266176476354, "learning_rate": 1.9685944867876373e-05, "loss": 0.9402, "step": 3965 }, { "epoch": 0.8182192910140148, "grad_norm": 0.3750595800822031, "learning_rate": 1.9472112372628536e-05, "loss": 1.0682, "step": 3970 }, { "epoch": 0.8192497938994229, "grad_norm": 0.4415444668426194, "learning_rate": 1.9259322238323095e-05, "loss": 1.1201, "step": 3975 }, { "epoch": 0.8202802967848309, "grad_norm": 0.31061384334227865, "learning_rate": 1.9047577219336665e-05, "loss": 0.8734, "step": 3980 }, { "epoch": 0.8213107996702391, "grad_norm": 0.3577711991948567, "learning_rate": 1.8836880056517658e-05, "loss": 1.1896, "step": 3985 }, { "epoch": 0.8223413025556472, "grad_norm": 0.2862925377080917, "learning_rate": 1.862723347715103e-05, "loss": 0.9206, "step": 3990 }, { "epoch": 0.8233718054410553, "grad_norm": 0.36692123404130295, "learning_rate": 1.841864019492282e-05, "loss": 1.0504, "step": 3995 }, { "epoch": 0.8244023083264633, "grad_norm": 0.3600663907207253, "learning_rate": 1.821110290988509e-05, "loss": 1.2122, "step": 4000 }, { "epoch": 0.8244023083264633, "eval_loss": 0.9671783447265625, "eval_runtime": 2672.6929, "eval_samples_per_second": 2.993, "eval_steps_per_second": 0.187, "step": 4000 }, { "epoch": 0.8254328112118714, "grad_norm": 0.3081888529721894, "learning_rate": 1.8004624308421026e-05, "loss": 0.7821, "step": 4005 }, { "epoch": 0.8264633140972795, "grad_norm": 0.36648973755342457, "learning_rate": 1.7799207063210044e-05, "loss": 1.2144, "step": 4010 }, { "epoch": 0.8274938169826875, "grad_norm": 0.2898792198270504, "learning_rate": 1.759485383319326e-05, "loss": 0.9427, "step": 4015 }, { "epoch": 0.8285243198680956, "grad_norm": 0.38174814126983564, "learning_rate": 1.7391567263539144e-05, "loss": 1.0154, "step": 4020 }, { "epoch": 0.8295548227535037, "grad_norm": 0.4742619716693875, "learning_rate": 1.7189349985609115e-05, "loss": 1.0813, "step": 4025 }, { "epoch": 0.8305853256389117, "grad_norm": 0.3664614977301218, "learning_rate": 1.6988204616923666e-05, "loss": 0.8368, "step": 4030 }, { "epoch": 0.8316158285243199, "grad_norm": 0.43325455353643755, "learning_rate": 1.6788133761128312e-05, "loss": 1.1778, "step": 4035 }, { "epoch": 0.832646331409728, "grad_norm": 0.29420614649546306, "learning_rate": 1.658914000795999e-05, "loss": 0.922, "step": 4040 }, { "epoch": 0.8336768342951361, "grad_norm": 0.3437287490008131, "learning_rate": 1.639122593321357e-05, "loss": 1.0901, "step": 4045 }, { "epoch": 0.8347073371805441, "grad_norm": 0.41394098029113957, "learning_rate": 1.6194394098708377e-05, "loss": 1.1766, "step": 4050 }, { "epoch": 0.8357378400659522, "grad_norm": 0.3399407346712763, "learning_rate": 1.59986470522551e-05, "loss": 0.8491, "step": 4055 }, { "epoch": 0.8367683429513603, "grad_norm": 0.3655997093950039, "learning_rate": 1.580398732762297e-05, "loss": 1.1981, "step": 4060 }, { "epoch": 0.8377988458367683, "grad_norm": 0.30216864477326527, "learning_rate": 1.5610417444506664e-05, "loss": 0.9703, "step": 4065 }, { "epoch": 0.8388293487221764, "grad_norm": 0.3993955860213197, "learning_rate": 1.541793990849387e-05, "loss": 1.0866, "step": 4070 }, { "epoch": 0.8398598516075845, "grad_norm": 0.40765142147078065, "learning_rate": 1.522655721103291e-05, "loss": 1.1578, "step": 4075 }, { "epoch": 0.8408903544929925, "grad_norm": 0.3132479277395683, "learning_rate": 1.5036271829400294e-05, "loss": 0.8317, "step": 4080 }, { "epoch": 0.8419208573784006, "grad_norm": 0.353057034757265, "learning_rate": 1.4847086226668872e-05, "loss": 1.1387, "step": 4085 }, { "epoch": 0.8429513602638088, "grad_norm": 0.28714788090256477, "learning_rate": 1.4659002851675774e-05, "loss": 0.9363, "step": 4090 }, { "epoch": 0.8439818631492169, "grad_norm": 0.3558290938292747, "learning_rate": 1.447202413899078e-05, "loss": 0.9974, "step": 4095 }, { "epoch": 0.8450123660346249, "grad_norm": 0.36577045780008416, "learning_rate": 1.4286152508884898e-05, "loss": 1.1466, "step": 4100 }, { "epoch": 0.846042868920033, "grad_norm": 0.3010779731099561, "learning_rate": 1.4101390367298861e-05, "loss": 0.7895, "step": 4105 }, { "epoch": 0.8470733718054411, "grad_norm": 0.3541525630578856, "learning_rate": 1.3917740105812094e-05, "loss": 1.1429, "step": 4110 }, { "epoch": 0.8481038746908491, "grad_norm": 0.28819494987192507, "learning_rate": 1.3735204101611776e-05, "loss": 0.9335, "step": 4115 }, { "epoch": 0.8491343775762572, "grad_norm": 0.34823589325876386, "learning_rate": 1.355378471746196e-05, "loss": 1.0577, "step": 4120 }, { "epoch": 0.8501648804616653, "grad_norm": 0.34950447449734157, "learning_rate": 1.3373484301673145e-05, "loss": 1.1231, "step": 4125 }, { "epoch": 0.8511953833470733, "grad_norm": 0.29975822569477556, "learning_rate": 1.3194305188071732e-05, "loss": 0.8424, "step": 4130 }, { "epoch": 0.8522258862324814, "grad_norm": 0.390420435835868, "learning_rate": 1.301624969596985e-05, "loss": 1.1329, "step": 4135 }, { "epoch": 0.8532563891178895, "grad_norm": 0.2911183426984063, "learning_rate": 1.2839320130135468e-05, "loss": 0.9841, "step": 4140 }, { "epoch": 0.8542868920032977, "grad_norm": 0.3500369252129129, "learning_rate": 1.266351878076234e-05, "loss": 1.0513, "step": 4145 }, { "epoch": 0.8553173948887057, "grad_norm": 0.4017098058926308, "learning_rate": 1.2488847923440483e-05, "loss": 1.1624, "step": 4150 }, { "epoch": 0.8563478977741138, "grad_norm": 0.3441542578051488, "learning_rate": 1.2315309819126852e-05, "loss": 0.8611, "step": 4155 }, { "epoch": 0.8573784006595219, "grad_norm": 0.36811687971995544, "learning_rate": 1.2142906714115787e-05, "loss": 1.2298, "step": 4160 }, { "epoch": 0.8584089035449299, "grad_norm": 0.2934040538277352, "learning_rate": 1.197164084001009e-05, "loss": 0.9825, "step": 4165 }, { "epoch": 0.859439406430338, "grad_norm": 0.3571442419781514, "learning_rate": 1.1801514413692239e-05, "loss": 1.0296, "step": 4170 }, { "epoch": 0.8604699093157461, "grad_norm": 0.4388041060818487, "learning_rate": 1.1632529637295475e-05, "loss": 1.1665, "step": 4175 }, { "epoch": 0.8615004122011541, "grad_norm": 0.3211936804672997, "learning_rate": 1.1464688698175497e-05, "loss": 0.8568, "step": 4180 }, { "epoch": 0.8625309150865622, "grad_norm": 0.38975409701266794, "learning_rate": 1.1297993768881998e-05, "loss": 1.1854, "step": 4185 }, { "epoch": 0.8635614179719703, "grad_norm": 0.27934297465783847, "learning_rate": 1.113244700713063e-05, "loss": 0.8697, "step": 4190 }, { "epoch": 0.8645919208573783, "grad_norm": 0.39072549944659307, "learning_rate": 1.0968050555775067e-05, "loss": 1.0164, "step": 4195 }, { "epoch": 0.8656224237427865, "grad_norm": 0.4998393287633088, "learning_rate": 1.0804806542779223e-05, "loss": 1.0571, "step": 4200 }, { "epoch": 0.8656224237427865, "eval_loss": 0.9649366140365601, "eval_runtime": 2477.8624, "eval_samples_per_second": 3.229, "eval_steps_per_second": 0.202, "step": 4200 }, { "epoch": 0.8666529266281946, "grad_norm": 0.30161833311373715, "learning_rate": 1.0642717081189735e-05, "loss": 0.8134, "step": 4205 }, { "epoch": 0.8676834295136027, "grad_norm": 0.3714099835276263, "learning_rate": 1.0481784269108664e-05, "loss": 1.2136, "step": 4210 }, { "epoch": 0.8687139323990107, "grad_norm": 0.29858155300200195, "learning_rate": 1.032201018966621e-05, "loss": 0.9334, "step": 4215 }, { "epoch": 0.8697444352844188, "grad_norm": 0.36365152527356986, "learning_rate": 1.0163396910993883e-05, "loss": 1.0167, "step": 4220 }, { "epoch": 0.8707749381698269, "grad_norm": 0.3926164945656461, "learning_rate": 1.0005946486197648e-05, "loss": 1.1597, "step": 4225 }, { "epoch": 0.8718054410552349, "grad_norm": 0.31631364428210557, "learning_rate": 9.849660953331363e-06, "loss": 0.8626, "step": 4230 }, { "epoch": 0.872835943940643, "grad_norm": 0.397012339843304, "learning_rate": 9.694542335370437e-06, "loss": 1.1559, "step": 4235 }, { "epoch": 0.8738664468260511, "grad_norm": 0.27478860653538173, "learning_rate": 9.540592640185597e-06, "loss": 0.9427, "step": 4240 }, { "epoch": 0.8748969497114591, "grad_norm": 0.34552863820949387, "learning_rate": 9.387813860516915e-06, "loss": 1.0141, "step": 4245 }, { "epoch": 0.8759274525968672, "grad_norm": 0.4001075016221677, "learning_rate": 9.236207973948063e-06, "loss": 1.1693, "step": 4250 }, { "epoch": 0.8769579554822754, "grad_norm": 0.3089362146851165, "learning_rate": 9.085776942880608e-06, "loss": 0.8552, "step": 4255 }, { "epoch": 0.8779884583676835, "grad_norm": 0.3341234310512727, "learning_rate": 8.936522714508678e-06, "loss": 1.1562, "step": 4260 }, { "epoch": 0.8790189612530915, "grad_norm": 0.2798328581056808, "learning_rate": 8.788447220793806e-06, "loss": 0.99, "step": 4265 }, { "epoch": 0.8800494641384996, "grad_norm": 0.39837992045978476, "learning_rate": 8.641552378439776e-06, "loss": 1.1073, "step": 4270 }, { "epoch": 0.8810799670239077, "grad_norm": 0.4106276064285754, "learning_rate": 8.495840088868024e-06, "loss": 1.1004, "step": 4275 }, { "epoch": 0.8821104699093157, "grad_norm": 0.37910153433606, "learning_rate": 8.351312238192787e-06, "loss": 0.9667, "step": 4280 }, { "epoch": 0.8831409727947238, "grad_norm": 0.3835389510278448, "learning_rate": 8.207970697196866e-06, "loss": 1.1637, "step": 4285 }, { "epoch": 0.8841714756801319, "grad_norm": 0.2996106330164154, "learning_rate": 8.065817321307367e-06, "loss": 0.8998, "step": 4290 }, { "epoch": 0.88520197856554, "grad_norm": 0.37985003315229504, "learning_rate": 7.924853950571642e-06, "loss": 0.9981, "step": 4295 }, { "epoch": 0.886232481450948, "grad_norm": 0.3734475653137268, "learning_rate": 7.78508240963347e-06, "loss": 1.1901, "step": 4300 }, { "epoch": 0.8872629843363561, "grad_norm": 0.32917519272254847, "learning_rate": 7.646504507709563e-06, "loss": 0.8026, "step": 4305 }, { "epoch": 0.8882934872217643, "grad_norm": 0.3184623209600695, "learning_rate": 7.50912203856593e-06, "loss": 1.1705, "step": 4310 }, { "epoch": 0.8893239901071723, "grad_norm": 0.27564137851589965, "learning_rate": 7.372936780494877e-06, "loss": 0.8968, "step": 4315 }, { "epoch": 0.8903544929925804, "grad_norm": 0.37744723978180933, "learning_rate": 7.237950496291856e-06, "loss": 0.9825, "step": 4320 }, { "epoch": 0.8913849958779885, "grad_norm": 0.3497470810206515, "learning_rate": 7.104164933232649e-06, "loss": 1.1498, "step": 4325 }, { "epoch": 0.8924154987633965, "grad_norm": 0.314664899690108, "learning_rate": 6.971581823050832e-06, "loss": 0.8124, "step": 4330 }, { "epoch": 0.8934460016488046, "grad_norm": 0.3437097611977056, "learning_rate": 6.840202881915325e-06, "loss": 1.1555, "step": 4335 }, { "epoch": 0.8944765045342127, "grad_norm": 0.28333067118500105, "learning_rate": 6.710029810408092e-06, "loss": 0.899, "step": 4340 }, { "epoch": 0.8955070074196207, "grad_norm": 0.39572557163512656, "learning_rate": 6.581064293502293e-06, "loss": 1.0465, "step": 4345 }, { "epoch": 0.8965375103050288, "grad_norm": 0.37753981586149277, "learning_rate": 6.453308000540304e-06, "loss": 1.1223, "step": 4350 }, { "epoch": 0.8975680131904369, "grad_norm": 0.33675472826608077, "learning_rate": 6.326762585212209e-06, "loss": 0.8248, "step": 4355 }, { "epoch": 0.8985985160758451, "grad_norm": 0.3825197592387351, "learning_rate": 6.20142968553441e-06, "loss": 1.1401, "step": 4360 }, { "epoch": 0.8996290189612531, "grad_norm": 0.2838180681849475, "learning_rate": 6.077310923828328e-06, "loss": 0.909, "step": 4365 }, { "epoch": 0.9006595218466612, "grad_norm": 0.37686082536565757, "learning_rate": 5.954407906699511e-06, "loss": 1.0217, "step": 4370 }, { "epoch": 0.9016900247320693, "grad_norm": 0.3380783448392427, "learning_rate": 5.8327222250167735e-06, "loss": 1.1523, "step": 4375 }, { "epoch": 0.9027205276174773, "grad_norm": 0.2883596670864561, "learning_rate": 5.71225545389158e-06, "loss": 0.8144, "step": 4380 }, { "epoch": 0.9037510305028854, "grad_norm": 0.3483906272756692, "learning_rate": 5.59300915265778e-06, "loss": 1.1855, "step": 4385 }, { "epoch": 0.9047815333882935, "grad_norm": 0.2639750686093266, "learning_rate": 5.4749848648512624e-06, "loss": 0.9381, "step": 4390 }, { "epoch": 0.9058120362737015, "grad_norm": 0.3972431149683815, "learning_rate": 5.358184118190068e-06, "loss": 1.0683, "step": 4395 }, { "epoch": 0.9068425391591096, "grad_norm": 0.35801168305170344, "learning_rate": 5.242608424554651e-06, "loss": 1.0903, "step": 4400 }, { "epoch": 0.9068425391591096, "eval_loss": 0.9635400176048279, "eval_runtime": 2476.6467, "eval_samples_per_second": 3.23, "eval_steps_per_second": 0.202, "step": 4400 }, { "epoch": 0.9078730420445177, "grad_norm": 0.33090712708760794, "learning_rate": 5.128259279968195e-06, "loss": 0.8644, "step": 4405 }, { "epoch": 0.9089035449299258, "grad_norm": 0.35066225756340424, "learning_rate": 5.01513816457736e-06, "loss": 1.1482, "step": 4410 }, { "epoch": 0.9099340478153339, "grad_norm": 0.2777858386287618, "learning_rate": 4.903246542633033e-06, "loss": 0.9529, "step": 4415 }, { "epoch": 0.910964550700742, "grad_norm": 0.3807172458090473, "learning_rate": 4.7925858624714215e-06, "loss": 1.0499, "step": 4420 }, { "epoch": 0.9119950535861501, "grad_norm": 0.36543715138130545, "learning_rate": 4.683157556495343e-06, "loss": 1.1883, "step": 4425 }, { "epoch": 0.9130255564715581, "grad_norm": 0.3339553261899812, "learning_rate": 4.574963041155622e-06, "loss": 0.8319, "step": 4430 }, { "epoch": 0.9140560593569662, "grad_norm": 0.3202714496531148, "learning_rate": 4.468003716932734e-06, "loss": 1.1678, "step": 4435 }, { "epoch": 0.9150865622423743, "grad_norm": 0.27715040842554794, "learning_rate": 4.362280968318777e-06, "loss": 0.9364, "step": 4440 }, { "epoch": 0.9161170651277823, "grad_norm": 0.3628969851723974, "learning_rate": 4.257796163799455e-06, "loss": 1.0704, "step": 4445 }, { "epoch": 0.9171475680131904, "grad_norm": 0.4255339930477384, "learning_rate": 4.154550655836409e-06, "loss": 1.1622, "step": 4450 }, { "epoch": 0.9181780708985985, "grad_norm": 0.3098446765156209, "learning_rate": 4.052545780849715e-06, "loss": 0.8363, "step": 4455 }, { "epoch": 0.9192085737840066, "grad_norm": 0.38626454996818416, "learning_rate": 3.9517828592005475e-06, "loss": 1.1148, "step": 4460 }, { "epoch": 0.9202390766694146, "grad_norm": 0.2786681923266397, "learning_rate": 3.852263195174155e-06, "loss": 0.8743, "step": 4465 }, { "epoch": 0.9212695795548228, "grad_norm": 0.3627972033581097, "learning_rate": 3.7539880769628998e-06, "loss": 1.0305, "step": 4470 }, { "epoch": 0.9223000824402309, "grad_norm": 0.3545740104488335, "learning_rate": 3.6569587766496216e-06, "loss": 1.1188, "step": 4475 }, { "epoch": 0.9233305853256389, "grad_norm": 0.30405956916616017, "learning_rate": 3.561176550191203e-06, "loss": 0.7902, "step": 4480 }, { "epoch": 0.924361088211047, "grad_norm": 0.40805109311840626, "learning_rate": 3.46664263740224e-06, "loss": 1.0949, "step": 4485 }, { "epoch": 0.9253915910964551, "grad_norm": 0.2937050314911077, "learning_rate": 3.3733582619390523e-06, "loss": 0.9339, "step": 4490 }, { "epoch": 0.9264220939818631, "grad_norm": 0.37795976499150646, "learning_rate": 3.281324631283833e-06, "loss": 1.0108, "step": 4495 }, { "epoch": 0.9274525968672712, "grad_norm": 0.4168578135164906, "learning_rate": 3.1905429367289795e-06, "loss": 1.1267, "step": 4500 }, { "epoch": 0.9284830997526793, "grad_norm": 0.30324094735934326, "learning_rate": 3.101014353361753e-06, "loss": 0.8109, "step": 4505 }, { "epoch": 0.9295136026380874, "grad_norm": 0.3929210483289662, "learning_rate": 3.012740040048978e-06, "loss": 1.1848, "step": 4510 }, { "epoch": 0.9305441055234954, "grad_norm": 0.2889829253469363, "learning_rate": 2.9257211394220773e-06, "loss": 0.8944, "step": 4515 }, { "epoch": 0.9315746084089035, "grad_norm": 0.39922043302261206, "learning_rate": 2.8399587778623505e-06, "loss": 1.0599, "step": 4520 }, { "epoch": 0.9326051112943117, "grad_norm": 0.41401354380939953, "learning_rate": 2.755454065486263e-06, "loss": 1.1731, "step": 4525 }, { "epoch": 0.9336356141797197, "grad_norm": 0.3816124740915107, "learning_rate": 2.672208096131157e-06, "loss": 0.9285, "step": 4530 }, { "epoch": 0.9346661170651278, "grad_norm": 0.365839138684449, "learning_rate": 2.5902219473411204e-06, "loss": 1.1613, "step": 4535 }, { "epoch": 0.9356966199505359, "grad_norm": 0.30150334355809666, "learning_rate": 2.509496680352963e-06, "loss": 0.9157, "step": 4540 }, { "epoch": 0.936727122835944, "grad_norm": 0.36964220055921937, "learning_rate": 2.430033340082516e-06, "loss": 1.0569, "step": 4545 }, { "epoch": 0.937757625721352, "grad_norm": 0.41058988578726896, "learning_rate": 2.3518329551111217e-06, "loss": 1.1936, "step": 4550 }, { "epoch": 0.9387881286067601, "grad_norm": 0.32143724004908797, "learning_rate": 2.2748965376723e-06, "loss": 0.8617, "step": 4555 }, { "epoch": 0.9398186314921682, "grad_norm": 0.3695703528509068, "learning_rate": 2.199225083638656e-06, "loss": 1.2086, "step": 4560 }, { "epoch": 0.9408491343775762, "grad_norm": 0.28544602650039125, "learning_rate": 2.1248195725089624e-06, "loss": 0.9673, "step": 4565 }, { "epoch": 0.9418796372629843, "grad_norm": 0.33475042033286156, "learning_rate": 2.0516809673955083e-06, "loss": 1.0888, "step": 4570 }, { "epoch": 0.9429101401483924, "grad_norm": 0.40597870855708795, "learning_rate": 1.9798102150116573e-06, "loss": 1.1073, "step": 4575 }, { "epoch": 0.9439406430338005, "grad_norm": 0.35068928153409934, "learning_rate": 1.909208245659522e-06, "loss": 0.8919, "step": 4580 }, { "epoch": 0.9449711459192086, "grad_norm": 0.3520873665570541, "learning_rate": 1.8398759732179637e-06, "loss": 1.1762, "step": 4585 }, { "epoch": 0.9460016488046167, "grad_norm": 0.31075650924614495, "learning_rate": 1.7718142951307914e-06, "loss": 0.8997, "step": 4590 }, { "epoch": 0.9470321516900247, "grad_norm": 0.3520700228079484, "learning_rate": 1.705024092395091e-06, "loss": 1.0955, "step": 4595 }, { "epoch": 0.9480626545754328, "grad_norm": 0.3680157698495691, "learning_rate": 1.6395062295498698e-06, "loss": 1.178, "step": 4600 }, { "epoch": 0.9480626545754328, "eval_loss": 0.9628809690475464, "eval_runtime": 2669.1038, "eval_samples_per_second": 2.997, "eval_steps_per_second": 0.187, "step": 4600 }, { "epoch": 0.9490931574608409, "grad_norm": 0.31126006459380184, "learning_rate": 1.5752615546647975e-06, "loss": 0.8694, "step": 4605 }, { "epoch": 0.950123660346249, "grad_norm": 0.34126438848006335, "learning_rate": 1.5122908993293273e-06, "loss": 1.1496, "step": 4610 }, { "epoch": 0.951154163231657, "grad_norm": 0.28191122396530277, "learning_rate": 1.4505950786418255e-06, "loss": 0.9136, "step": 4615 }, { "epoch": 0.9521846661170651, "grad_norm": 0.3757522319369373, "learning_rate": 1.3901748911991253e-06, "loss": 1.0589, "step": 4620 }, { "epoch": 0.9532151690024732, "grad_norm": 0.3812139506827204, "learning_rate": 1.331031119086079e-06, "loss": 1.1577, "step": 4625 }, { "epoch": 0.9542456718878813, "grad_norm": 0.3285095663923165, "learning_rate": 1.2731645278655445e-06, "loss": 0.9256, "step": 4630 }, { "epoch": 0.9552761747732894, "grad_norm": 0.3434798919525598, "learning_rate": 1.2165758665683924e-06, "loss": 1.1373, "step": 4635 }, { "epoch": 0.9563066776586975, "grad_norm": 0.3017676114908958, "learning_rate": 1.1612658676838473e-06, "loss": 0.8882, "step": 4640 }, { "epoch": 0.9573371805441055, "grad_norm": 0.3795284009217967, "learning_rate": 1.107235247150018e-06, "loss": 1.0639, "step": 4645 }, { "epoch": 0.9583676834295136, "grad_norm": 0.4845153985581749, "learning_rate": 1.0544847043445938e-06, "loss": 1.0753, "step": 4650 }, { "epoch": 0.9593981863149217, "grad_norm": 0.32363946255340464, "learning_rate": 1.0030149220758288e-06, "loss": 0.8307, "step": 4655 }, { "epoch": 0.9604286892003298, "grad_norm": 0.4184299714975136, "learning_rate": 9.528265665736502e-07, "loss": 1.1371, "step": 4660 }, { "epoch": 0.9614591920857378, "grad_norm": 0.28512105148343453, "learning_rate": 9.039202874811192e-07, "loss": 0.9634, "step": 4665 }, { "epoch": 0.9624896949711459, "grad_norm": 0.360971597617964, "learning_rate": 8.562967178459391e-07, "loss": 1.0805, "step": 4670 }, { "epoch": 0.963520197856554, "grad_norm": 0.3822471953295701, "learning_rate": 8.099564741123166e-07, "loss": 1.2147, "step": 4675 }, { "epoch": 0.964550700741962, "grad_norm": 0.32385746632379775, "learning_rate": 7.649001561129354e-07, "loss": 0.8677, "step": 4680 }, { "epoch": 0.9655812036273702, "grad_norm": 0.33325536515100207, "learning_rate": 7.211283470612395e-07, "loss": 1.1736, "step": 4685 }, { "epoch": 0.9666117065127783, "grad_norm": 0.2789936033919106, "learning_rate": 6.786416135438512e-07, "loss": 0.9563, "step": 4690 }, { "epoch": 0.9676422093981863, "grad_norm": 0.3347741972532808, "learning_rate": 6.374405055132537e-07, "loss": 0.9439, "step": 4695 }, { "epoch": 0.9686727122835944, "grad_norm": 0.3903192561742831, "learning_rate": 5.975255562806647e-07, "loss": 1.1797, "step": 4700 }, { "epoch": 0.9697032151690025, "grad_norm": 0.32771315967905756, "learning_rate": 5.58897282509141e-07, "loss": 0.8953, "step": 4705 }, { "epoch": 0.9707337180544106, "grad_norm": 0.34884009544181055, "learning_rate": 5.215561842068728e-07, "loss": 1.1761, "step": 4710 }, { "epoch": 0.9717642209398186, "grad_norm": 0.2775010942276059, "learning_rate": 4.855027447207338e-07, "loss": 0.9457, "step": 4715 }, { "epoch": 0.9727947238252267, "grad_norm": 0.382431683738394, "learning_rate": 4.507374307299972e-07, "loss": 1.0359, "step": 4720 }, { "epoch": 0.9738252267106348, "grad_norm": 0.4253814777599987, "learning_rate": 4.172606922403399e-07, "loss": 1.1746, "step": 4725 }, { "epoch": 0.9748557295960428, "grad_norm": 0.3119359666789732, "learning_rate": 3.8507296257798145e-07, "loss": 0.8048, "step": 4730 }, { "epoch": 0.9758862324814509, "grad_norm": 0.35641657318401637, "learning_rate": 3.541746583840655e-07, "loss": 1.1508, "step": 4735 }, { "epoch": 0.9769167353668591, "grad_norm": 0.28277759207142494, "learning_rate": 3.24566179609298e-07, "loss": 0.9406, "step": 4740 }, { "epoch": 0.9779472382522671, "grad_norm": 0.3649978639842664, "learning_rate": 2.9624790950875113e-07, "loss": 1.0074, "step": 4745 }, { "epoch": 0.9789777411376752, "grad_norm": 0.43930720949024876, "learning_rate": 2.692202146369338e-07, "loss": 1.1133, "step": 4750 }, { "epoch": 0.9800082440230833, "grad_norm": 0.3568172083915493, "learning_rate": 2.434834448429957e-07, "loss": 0.8548, "step": 4755 }, { "epoch": 0.9810387469084914, "grad_norm": 0.3512721884204812, "learning_rate": 2.1903793326621957e-07, "loss": 1.1658, "step": 4760 }, { "epoch": 0.9820692497938994, "grad_norm": 0.28260427199429944, "learning_rate": 1.9588399633173605e-07, "loss": 0.8872, "step": 4765 }, { "epoch": 0.9830997526793075, "grad_norm": 0.3342814004364708, "learning_rate": 1.740219337463822e-07, "loss": 1.0368, "step": 4770 }, { "epoch": 0.9841302555647156, "grad_norm": 0.39410468491852235, "learning_rate": 1.534520284948715e-07, "loss": 1.1301, "step": 4775 }, { "epoch": 0.9851607584501236, "grad_norm": 0.32300550472176315, "learning_rate": 1.3417454683608554e-07, "loss": 0.8525, "step": 4780 }, { "epoch": 0.9861912613355317, "grad_norm": 0.38783208707768235, "learning_rate": 1.1618973829966572e-07, "loss": 1.2139, "step": 4785 }, { "epoch": 0.9872217642209398, "grad_norm": 0.27657069522605154, "learning_rate": 9.949783568272697e-08, "loss": 0.9094, "step": 4790 }, { "epoch": 0.988252267106348, "grad_norm": 0.37764224067778107, "learning_rate": 8.409905504693782e-08, "loss": 1.0543, "step": 4795 }, { "epoch": 0.989282769991756, "grad_norm": 0.5230932784753239, "learning_rate": 6.999359571561171e-08, "loss": 1.1661, "step": 4800 }, { "epoch": 0.989282769991756, "eval_loss": 0.9627586603164673, "eval_runtime": 2560.4353, "eval_samples_per_second": 3.124, "eval_steps_per_second": 0.195, "step": 4800 }, { "epoch": 0.9903132728771641, "grad_norm": 0.2751657982671328, "learning_rate": 5.718164027121997e-08, "loss": 0.7531, "step": 4805 }, { "epoch": 0.9913437757625722, "grad_norm": 0.3764599692340869, "learning_rate": 4.566335455299387e-08, "loss": 1.1338, "step": 4810 }, { "epoch": 0.9923742786479802, "grad_norm": 0.2992197664538524, "learning_rate": 3.5438887654737355e-08, "loss": 0.8897, "step": 4815 }, { "epoch": 0.9934047815333883, "grad_norm": 0.3535257610205098, "learning_rate": 2.6508371922984166e-08, "loss": 1.0454, "step": 4820 }, { "epoch": 0.9944352844187964, "grad_norm": 0.37026464651118146, "learning_rate": 1.887192295521034e-08, "loss": 1.1346, "step": 4825 }, { "epoch": 0.9954657873042044, "grad_norm": 0.3271412553655858, "learning_rate": 1.252963959834652e-08, "loss": 0.8182, "step": 4830 }, { "epoch": 0.9964962901896125, "grad_norm": 0.4040096861009905, "learning_rate": 7.481603947556703e-09, "loss": 1.0549, "step": 4835 }, { "epoch": 0.9975267930750206, "grad_norm": 0.28077083214803567, "learning_rate": 3.727881345105821e-09, "loss": 0.918, "step": 4840 }, { "epoch": 0.9985572959604286, "grad_norm": 0.38213498100295196, "learning_rate": 1.2685203795492762e-09, "loss": 1.0451, "step": 4845 }, { "epoch": 0.9995877988458368, "grad_norm": 0.4052030227111039, "learning_rate": 1.0355288510011107e-10, "loss": 1.1878, "step": 4850 }, { "epoch": 1.0, "step": 4852, "total_flos": 3.101569705521971e+16, "train_loss": 1.1041986294158312, "train_runtime": 123226.4419, "train_samples_per_second": 0.63, "train_steps_per_second": 0.039 } ], "logging_steps": 5, "max_steps": 4852, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.101569705521971e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }