{ "best_metric": 0.07686587423086166, "best_model_checkpoint": "/teamspace/studios/this_studio/save/LLama_End/checkpoint-3500", "epoch": 1.905539676058255, "eval_steps": 500, "global_step": 3500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027221995372260785, "grad_norm": 0.3611927926540375, "learning_rate": 5.000000000000001e-07, "loss": 0.7652, "step": 5 }, { "epoch": 0.005444399074452157, "grad_norm": 0.3633479177951813, "learning_rate": 1.0000000000000002e-06, "loss": 0.7798, "step": 10 }, { "epoch": 0.008166598611678236, "grad_norm": 0.31540995836257935, "learning_rate": 1.5e-06, "loss": 0.7635, "step": 15 }, { "epoch": 0.010888798148904314, "grad_norm": 0.3314383924007416, "learning_rate": 2.0000000000000003e-06, "loss": 0.7475, "step": 20 }, { "epoch": 0.013610997686130393, "grad_norm": 0.391197144985199, "learning_rate": 2.5e-06, "loss": 0.7827, "step": 25 }, { "epoch": 0.016333197223356473, "grad_norm": 0.39311718940734863, "learning_rate": 3e-06, "loss": 0.7985, "step": 30 }, { "epoch": 0.01905539676058255, "grad_norm": 0.34564024209976196, "learning_rate": 3.5000000000000004e-06, "loss": 0.7608, "step": 35 }, { "epoch": 0.021777596297808628, "grad_norm": 0.43378740549087524, "learning_rate": 4.000000000000001e-06, "loss": 0.7566, "step": 40 }, { "epoch": 0.02449979583503471, "grad_norm": 0.38521063327789307, "learning_rate": 4.5e-06, "loss": 0.75, "step": 45 }, { "epoch": 0.027221995372260787, "grad_norm": 0.4021587371826172, "learning_rate": 5e-06, "loss": 0.7492, "step": 50 }, { "epoch": 0.029944194909486865, "grad_norm": 0.44062522053718567, "learning_rate": 5.500000000000001e-06, "loss": 0.7208, "step": 55 }, { "epoch": 0.032666394446712946, "grad_norm": 0.37620142102241516, "learning_rate": 6e-06, "loss": 0.6911, "step": 60 }, { "epoch": 0.03538859398393902, "grad_norm": 0.27264639735221863, "learning_rate": 6.5000000000000004e-06, "loss": 0.6624, "step": 65 }, { "epoch": 0.0381107935211651, "grad_norm": 0.2566782832145691, "learning_rate": 7.000000000000001e-06, "loss": 0.6326, "step": 70 }, { "epoch": 0.04083299305839118, "grad_norm": 0.27515003085136414, "learning_rate": 7.5e-06, "loss": 0.6135, "step": 75 }, { "epoch": 0.043555192595617256, "grad_norm": 0.2592063248157501, "learning_rate": 8.000000000000001e-06, "loss": 0.5585, "step": 80 }, { "epoch": 0.046277392132843334, "grad_norm": 0.22615216672420502, "learning_rate": 8.500000000000002e-06, "loss": 0.5482, "step": 85 }, { "epoch": 0.04899959167006942, "grad_norm": 0.21914538741111755, "learning_rate": 9e-06, "loss": 0.5488, "step": 90 }, { "epoch": 0.051721791207295496, "grad_norm": 0.23039759695529938, "learning_rate": 9.5e-06, "loss": 0.4816, "step": 95 }, { "epoch": 0.054443990744521574, "grad_norm": 0.21080121397972107, "learning_rate": 1e-05, "loss": 0.4686, "step": 100 }, { "epoch": 0.05716619028174765, "grad_norm": 0.19633150100708008, "learning_rate": 1.05e-05, "loss": 0.4426, "step": 105 }, { "epoch": 0.05988838981897373, "grad_norm": 0.20176054537296295, "learning_rate": 1.1000000000000001e-05, "loss": 0.4131, "step": 110 }, { "epoch": 0.06261058935619981, "grad_norm": 0.1965443640947342, "learning_rate": 1.1500000000000002e-05, "loss": 0.394, "step": 115 }, { "epoch": 0.06533278889342589, "grad_norm": 0.18852515518665314, "learning_rate": 1.2e-05, "loss": 0.3621, "step": 120 }, { "epoch": 0.06805498843065197, "grad_norm": 0.1972607970237732, "learning_rate": 1.25e-05, "loss": 0.3436, "step": 125 }, { "epoch": 0.07077718796787805, "grad_norm": 0.1883157640695572, "learning_rate": 1.3000000000000001e-05, "loss": 0.3235, "step": 130 }, { "epoch": 0.07349938750510412, "grad_norm": 0.20938365161418915, "learning_rate": 1.3500000000000001e-05, "loss": 0.3093, "step": 135 }, { "epoch": 0.0762215870423302, "grad_norm": 0.2136233150959015, "learning_rate": 1.4000000000000001e-05, "loss": 0.281, "step": 140 }, { "epoch": 0.07894378657955628, "grad_norm": 0.23373591899871826, "learning_rate": 1.45e-05, "loss": 0.2645, "step": 145 }, { "epoch": 0.08166598611678236, "grad_norm": 0.21917635202407837, "learning_rate": 1.5e-05, "loss": 0.2399, "step": 150 }, { "epoch": 0.08438818565400844, "grad_norm": 0.2619529068470001, "learning_rate": 1.55e-05, "loss": 0.2381, "step": 155 }, { "epoch": 0.08711038519123451, "grad_norm": 0.2083514928817749, "learning_rate": 1.6000000000000003e-05, "loss": 0.2139, "step": 160 }, { "epoch": 0.08983258472846059, "grad_norm": 0.22957220673561096, "learning_rate": 1.65e-05, "loss": 0.1956, "step": 165 }, { "epoch": 0.09255478426568667, "grad_norm": 0.270378977060318, "learning_rate": 1.7000000000000003e-05, "loss": 0.1906, "step": 170 }, { "epoch": 0.09527698380291276, "grad_norm": 0.24123062193393707, "learning_rate": 1.75e-05, "loss": 0.1811, "step": 175 }, { "epoch": 0.09799918334013884, "grad_norm": 0.2726229429244995, "learning_rate": 1.8e-05, "loss": 0.1756, "step": 180 }, { "epoch": 0.10072138287736492, "grad_norm": 0.24878880381584167, "learning_rate": 1.85e-05, "loss": 0.1745, "step": 185 }, { "epoch": 0.10344358241459099, "grad_norm": 0.22229672968387604, "learning_rate": 1.9e-05, "loss": 0.1641, "step": 190 }, { "epoch": 0.10616578195181707, "grad_norm": 0.25277239084243774, "learning_rate": 1.9500000000000003e-05, "loss": 0.1478, "step": 195 }, { "epoch": 0.10888798148904315, "grad_norm": 0.23793649673461914, "learning_rate": 2e-05, "loss": 0.1621, "step": 200 }, { "epoch": 0.11161018102626923, "grad_norm": 0.22567716240882874, "learning_rate": 2.05e-05, "loss": 0.1416, "step": 205 }, { "epoch": 0.1143323805634953, "grad_norm": 0.24615788459777832, "learning_rate": 2.1e-05, "loss": 0.1418, "step": 210 }, { "epoch": 0.11705458010072138, "grad_norm": 0.23229427635669708, "learning_rate": 2.15e-05, "loss": 0.137, "step": 215 }, { "epoch": 0.11977677963794746, "grad_norm": 0.2679113447666168, "learning_rate": 2.2000000000000003e-05, "loss": 0.1186, "step": 220 }, { "epoch": 0.12249897917517354, "grad_norm": 0.24723300337791443, "learning_rate": 2.25e-05, "loss": 0.1243, "step": 225 }, { "epoch": 0.12522117871239963, "grad_norm": 0.24604055285453796, "learning_rate": 2.3000000000000003e-05, "loss": 0.1472, "step": 230 }, { "epoch": 0.1279433782496257, "grad_norm": 0.2207736223936081, "learning_rate": 2.35e-05, "loss": 0.1258, "step": 235 }, { "epoch": 0.13066557778685178, "grad_norm": 0.2685292065143585, "learning_rate": 2.4e-05, "loss": 0.1407, "step": 240 }, { "epoch": 0.13338777732407786, "grad_norm": 0.21564778685569763, "learning_rate": 2.45e-05, "loss": 0.1355, "step": 245 }, { "epoch": 0.13610997686130394, "grad_norm": 0.22922654449939728, "learning_rate": 2.5e-05, "loss": 0.1298, "step": 250 }, { "epoch": 0.13883217639853002, "grad_norm": 0.21827565133571625, "learning_rate": 2.5500000000000003e-05, "loss": 0.1256, "step": 255 }, { "epoch": 0.1415543759357561, "grad_norm": 0.2129678726196289, "learning_rate": 2.6000000000000002e-05, "loss": 0.1435, "step": 260 }, { "epoch": 0.14427657547298217, "grad_norm": 0.196010559797287, "learning_rate": 2.6500000000000004e-05, "loss": 0.1353, "step": 265 }, { "epoch": 0.14699877501020825, "grad_norm": 0.22913923859596252, "learning_rate": 2.7000000000000002e-05, "loss": 0.1187, "step": 270 }, { "epoch": 0.14972097454743433, "grad_norm": 0.26281264424324036, "learning_rate": 2.7500000000000004e-05, "loss": 0.1697, "step": 275 }, { "epoch": 0.1524431740846604, "grad_norm": 0.28295812010765076, "learning_rate": 2.8000000000000003e-05, "loss": 0.1229, "step": 280 }, { "epoch": 0.15516537362188648, "grad_norm": 0.23690921068191528, "learning_rate": 2.8499999999999998e-05, "loss": 0.1086, "step": 285 }, { "epoch": 0.15788757315911256, "grad_norm": 0.22863948345184326, "learning_rate": 2.9e-05, "loss": 0.1064, "step": 290 }, { "epoch": 0.16060977269633864, "grad_norm": 0.1989349126815796, "learning_rate": 2.95e-05, "loss": 0.1145, "step": 295 }, { "epoch": 0.16333197223356472, "grad_norm": 0.2325228750705719, "learning_rate": 3e-05, "loss": 0.1181, "step": 300 }, { "epoch": 0.1660541717707908, "grad_norm": 0.2611011266708374, "learning_rate": 3.05e-05, "loss": 0.1093, "step": 305 }, { "epoch": 0.16877637130801687, "grad_norm": 0.23333916068077087, "learning_rate": 3.1e-05, "loss": 0.1119, "step": 310 }, { "epoch": 0.17149857084524295, "grad_norm": 0.22769390046596527, "learning_rate": 3.15e-05, "loss": 0.1079, "step": 315 }, { "epoch": 0.17422077038246903, "grad_norm": 0.23360049724578857, "learning_rate": 3.2000000000000005e-05, "loss": 0.1035, "step": 320 }, { "epoch": 0.1769429699196951, "grad_norm": 0.2405271828174591, "learning_rate": 3.2500000000000004e-05, "loss": 0.1208, "step": 325 }, { "epoch": 0.17966516945692118, "grad_norm": 0.2959926128387451, "learning_rate": 3.3e-05, "loss": 0.113, "step": 330 }, { "epoch": 0.18238736899414726, "grad_norm": 0.2246614694595337, "learning_rate": 3.35e-05, "loss": 0.1082, "step": 335 }, { "epoch": 0.18510956853137334, "grad_norm": 0.23407892882823944, "learning_rate": 3.4000000000000007e-05, "loss": 0.103, "step": 340 }, { "epoch": 0.18783176806859944, "grad_norm": 0.21018989384174347, "learning_rate": 3.45e-05, "loss": 0.1084, "step": 345 }, { "epoch": 0.19055396760582552, "grad_norm": 0.1964080035686493, "learning_rate": 3.5e-05, "loss": 0.1021, "step": 350 }, { "epoch": 0.1932761671430516, "grad_norm": 0.2037973254919052, "learning_rate": 3.55e-05, "loss": 0.1088, "step": 355 }, { "epoch": 0.19599836668027767, "grad_norm": 0.2156139761209488, "learning_rate": 3.6e-05, "loss": 0.1104, "step": 360 }, { "epoch": 0.19872056621750375, "grad_norm": 0.21694843471050262, "learning_rate": 3.65e-05, "loss": 0.1073, "step": 365 }, { "epoch": 0.20144276575472983, "grad_norm": 0.21144092082977295, "learning_rate": 3.7e-05, "loss": 0.1365, "step": 370 }, { "epoch": 0.2041649652919559, "grad_norm": 0.24178194999694824, "learning_rate": 3.7500000000000003e-05, "loss": 0.0984, "step": 375 }, { "epoch": 0.20688716482918199, "grad_norm": 0.20173633098602295, "learning_rate": 3.8e-05, "loss": 0.1289, "step": 380 }, { "epoch": 0.20960936436640806, "grad_norm": 0.2253100723028183, "learning_rate": 3.85e-05, "loss": 0.1055, "step": 385 }, { "epoch": 0.21233156390363414, "grad_norm": 0.2524998188018799, "learning_rate": 3.9000000000000006e-05, "loss": 0.1203, "step": 390 }, { "epoch": 0.21505376344086022, "grad_norm": 0.21095526218414307, "learning_rate": 3.9500000000000005e-05, "loss": 0.0965, "step": 395 }, { "epoch": 0.2177759629780863, "grad_norm": 0.1919548511505127, "learning_rate": 4e-05, "loss": 0.0961, "step": 400 }, { "epoch": 0.22049816251531237, "grad_norm": 0.2088468074798584, "learning_rate": 4.05e-05, "loss": 0.0945, "step": 405 }, { "epoch": 0.22322036205253845, "grad_norm": 0.22967375814914703, "learning_rate": 4.1e-05, "loss": 0.0991, "step": 410 }, { "epoch": 0.22594256158976453, "grad_norm": 0.19900205731391907, "learning_rate": 4.15e-05, "loss": 0.1069, "step": 415 }, { "epoch": 0.2286647611269906, "grad_norm": 0.22003936767578125, "learning_rate": 4.2e-05, "loss": 0.1037, "step": 420 }, { "epoch": 0.23138696066421668, "grad_norm": 0.21360959112644196, "learning_rate": 4.25e-05, "loss": 0.0952, "step": 425 }, { "epoch": 0.23410916020144276, "grad_norm": 0.1812208890914917, "learning_rate": 4.3e-05, "loss": 0.0962, "step": 430 }, { "epoch": 0.23683135973866884, "grad_norm": 0.19043520092964172, "learning_rate": 4.35e-05, "loss": 0.1075, "step": 435 }, { "epoch": 0.23955355927589492, "grad_norm": 0.2013886421918869, "learning_rate": 4.4000000000000006e-05, "loss": 0.104, "step": 440 }, { "epoch": 0.242275758813121, "grad_norm": 0.1831788420677185, "learning_rate": 4.4500000000000004e-05, "loss": 0.1012, "step": 445 }, { "epoch": 0.24499795835034707, "grad_norm": 0.1884724646806717, "learning_rate": 4.5e-05, "loss": 0.1115, "step": 450 }, { "epoch": 0.24772015788757315, "grad_norm": 0.1994744837284088, "learning_rate": 4.55e-05, "loss": 0.1006, "step": 455 }, { "epoch": 0.25044235742479926, "grad_norm": 0.1783529371023178, "learning_rate": 4.600000000000001e-05, "loss": 0.0902, "step": 460 }, { "epoch": 0.25316455696202533, "grad_norm": 0.180820032954216, "learning_rate": 4.6500000000000005e-05, "loss": 0.0937, "step": 465 }, { "epoch": 0.2558867564992514, "grad_norm": 0.19279232621192932, "learning_rate": 4.7e-05, "loss": 0.0944, "step": 470 }, { "epoch": 0.2586089560364775, "grad_norm": 0.1725597232580185, "learning_rate": 4.75e-05, "loss": 0.104, "step": 475 }, { "epoch": 0.26133115557370357, "grad_norm": 0.17777124047279358, "learning_rate": 4.8e-05, "loss": 0.0889, "step": 480 }, { "epoch": 0.26405335511092964, "grad_norm": 0.18942312896251678, "learning_rate": 4.85e-05, "loss": 0.0964, "step": 485 }, { "epoch": 0.2667755546481557, "grad_norm": 0.17807991802692413, "learning_rate": 4.9e-05, "loss": 0.1038, "step": 490 }, { "epoch": 0.2694977541853818, "grad_norm": 0.21202729642391205, "learning_rate": 4.9500000000000004e-05, "loss": 0.0941, "step": 495 }, { "epoch": 0.2722199537226079, "grad_norm": 0.18360304832458496, "learning_rate": 5e-05, "loss": 0.1013, "step": 500 }, { "epoch": 0.2722199537226079, "eval_loss": 0.09772992134094238, "eval_runtime": 271.0226, "eval_samples_per_second": 2.214, "eval_steps_per_second": 0.553, "step": 500 }, { "epoch": 0.27494215325983395, "grad_norm": 0.1730509251356125, "learning_rate": 4.999969346326857e-05, "loss": 0.098, "step": 505 }, { "epoch": 0.27766435279706003, "grad_norm": 0.17720815539360046, "learning_rate": 4.9998773860591444e-05, "loss": 0.0912, "step": 510 }, { "epoch": 0.2803865523342861, "grad_norm": 0.16912730038166046, "learning_rate": 4.9997241214519986e-05, "loss": 0.0885, "step": 515 }, { "epoch": 0.2831087518715122, "grad_norm": 0.166640505194664, "learning_rate": 4.999509556263919e-05, "loss": 0.1013, "step": 520 }, { "epoch": 0.28583095140873827, "grad_norm": 0.1893271952867508, "learning_rate": 4.999233695756673e-05, "loss": 0.1005, "step": 525 }, { "epoch": 0.28855315094596434, "grad_norm": 0.1966632455587387, "learning_rate": 4.998896546695172e-05, "loss": 0.0958, "step": 530 }, { "epoch": 0.2912753504831904, "grad_norm": 0.16383662819862366, "learning_rate": 4.998498117347302e-05, "loss": 0.0971, "step": 535 }, { "epoch": 0.2939975500204165, "grad_norm": 0.21534255146980286, "learning_rate": 4.998038417483721e-05, "loss": 0.1057, "step": 540 }, { "epoch": 0.2967197495576426, "grad_norm": 0.16162091493606567, "learning_rate": 4.9975174583776196e-05, "loss": 0.0943, "step": 545 }, { "epoch": 0.29944194909486865, "grad_norm": 0.1837451457977295, "learning_rate": 4.996935252804448e-05, "loss": 0.0968, "step": 550 }, { "epoch": 0.30216414863209473, "grad_norm": 0.15664127469062805, "learning_rate": 4.9962918150415946e-05, "loss": 0.0975, "step": 555 }, { "epoch": 0.3048863481693208, "grad_norm": 0.16413035988807678, "learning_rate": 4.995587160868047e-05, "loss": 0.0925, "step": 560 }, { "epoch": 0.3076085477065469, "grad_norm": 0.16700997948646545, "learning_rate": 4.994821307563995e-05, "loss": 0.097, "step": 565 }, { "epoch": 0.31033074724377296, "grad_norm": 0.16121503710746765, "learning_rate": 4.9939942739104105e-05, "loss": 0.0996, "step": 570 }, { "epoch": 0.31305294678099904, "grad_norm": 0.16212740540504456, "learning_rate": 4.993106080188592e-05, "loss": 0.0888, "step": 575 }, { "epoch": 0.3157751463182251, "grad_norm": 0.1662980020046234, "learning_rate": 4.9921567481796585e-05, "loss": 0.0994, "step": 580 }, { "epoch": 0.3184973458554512, "grad_norm": 0.1845940202474594, "learning_rate": 4.9911463011640195e-05, "loss": 0.0976, "step": 585 }, { "epoch": 0.3212195453926773, "grad_norm": 0.17079226672649384, "learning_rate": 4.9900747639208044e-05, "loss": 0.0947, "step": 590 }, { "epoch": 0.32394174492990335, "grad_norm": 0.1642676740884781, "learning_rate": 4.9889421627272575e-05, "loss": 0.1039, "step": 595 }, { "epoch": 0.32666394446712943, "grad_norm": 0.20938318967819214, "learning_rate": 4.987748525358087e-05, "loss": 0.1005, "step": 600 }, { "epoch": 0.3293861440043555, "grad_norm": 0.17017363011837006, "learning_rate": 4.9864938810847884e-05, "loss": 0.0941, "step": 605 }, { "epoch": 0.3321083435415816, "grad_norm": 0.17874999344348907, "learning_rate": 4.985178260674927e-05, "loss": 0.1055, "step": 610 }, { "epoch": 0.33483054307880766, "grad_norm": 0.16853144764900208, "learning_rate": 4.98380169639138e-05, "loss": 0.0941, "step": 615 }, { "epoch": 0.33755274261603374, "grad_norm": 0.18321168422698975, "learning_rate": 4.98236422199155e-05, "loss": 0.0991, "step": 620 }, { "epoch": 0.3402749421532598, "grad_norm": 0.16565477848052979, "learning_rate": 4.980865872726532e-05, "loss": 0.0979, "step": 625 }, { "epoch": 0.3429971416904859, "grad_norm": 0.19292525947093964, "learning_rate": 4.9793066853402536e-05, "loss": 0.0957, "step": 630 }, { "epoch": 0.345719341227712, "grad_norm": 0.1700926125049591, "learning_rate": 4.977686698068572e-05, "loss": 0.0889, "step": 635 }, { "epoch": 0.34844154076493805, "grad_norm": 0.15266568958759308, "learning_rate": 4.976005950638334e-05, "loss": 0.0928, "step": 640 }, { "epoch": 0.35116374030216413, "grad_norm": 0.15712082386016846, "learning_rate": 4.974264484266406e-05, "loss": 0.0927, "step": 645 }, { "epoch": 0.3538859398393902, "grad_norm": 0.15186965465545654, "learning_rate": 4.972462341658661e-05, "loss": 0.0944, "step": 650 }, { "epoch": 0.3566081393766163, "grad_norm": 0.18069899082183838, "learning_rate": 4.970599567008931e-05, "loss": 0.0998, "step": 655 }, { "epoch": 0.35933033891384236, "grad_norm": 0.14533384144306183, "learning_rate": 4.968676205997925e-05, "loss": 0.0981, "step": 660 }, { "epoch": 0.36205253845106844, "grad_norm": 0.16364873945713043, "learning_rate": 4.966692305792106e-05, "loss": 0.0932, "step": 665 }, { "epoch": 0.3647747379882945, "grad_norm": 0.18472006916999817, "learning_rate": 4.9646479150425376e-05, "loss": 0.0866, "step": 670 }, { "epoch": 0.3674969375255206, "grad_norm": 0.15847539901733398, "learning_rate": 4.962543083883687e-05, "loss": 0.0837, "step": 675 }, { "epoch": 0.3702191370627467, "grad_norm": 0.155991330742836, "learning_rate": 4.9603778639322004e-05, "loss": 0.0906, "step": 680 }, { "epoch": 0.37294133659997275, "grad_norm": 0.16835445165634155, "learning_rate": 4.958152308285633e-05, "loss": 0.0949, "step": 685 }, { "epoch": 0.3756635361371989, "grad_norm": 0.18460558354854584, "learning_rate": 4.95586647152115e-05, "loss": 0.0931, "step": 690 }, { "epoch": 0.37838573567442496, "grad_norm": 0.17255567014217377, "learning_rate": 4.9535204096941854e-05, "loss": 0.087, "step": 695 }, { "epoch": 0.38110793521165104, "grad_norm": 0.15889868140220642, "learning_rate": 4.951114180337069e-05, "loss": 0.0965, "step": 700 }, { "epoch": 0.3838301347488771, "grad_norm": 0.1403263658285141, "learning_rate": 4.948647842457615e-05, "loss": 0.0862, "step": 705 }, { "epoch": 0.3865523342861032, "grad_norm": 0.16774417459964752, "learning_rate": 4.946121456537676e-05, "loss": 0.0911, "step": 710 }, { "epoch": 0.38927453382332927, "grad_norm": 0.16498145461082458, "learning_rate": 4.9435350845316575e-05, "loss": 0.0851, "step": 715 }, { "epoch": 0.39199673336055535, "grad_norm": 0.15889093279838562, "learning_rate": 4.9408887898650036e-05, "loss": 0.0836, "step": 720 }, { "epoch": 0.3947189328977814, "grad_norm": 0.14469870924949646, "learning_rate": 4.9381826374326336e-05, "loss": 0.1033, "step": 725 }, { "epoch": 0.3974411324350075, "grad_norm": 0.1625497043132782, "learning_rate": 4.935416693597358e-05, "loss": 0.0862, "step": 730 }, { "epoch": 0.4001633319722336, "grad_norm": 0.15590594708919525, "learning_rate": 4.932591026188247e-05, "loss": 0.0946, "step": 735 }, { "epoch": 0.40288553150945966, "grad_norm": 0.1706922948360443, "learning_rate": 4.929705704498969e-05, "loss": 0.0913, "step": 740 }, { "epoch": 0.40560773104668574, "grad_norm": 0.15704989433288574, "learning_rate": 4.9267607992860906e-05, "loss": 0.0888, "step": 745 }, { "epoch": 0.4083299305839118, "grad_norm": 0.14272324740886688, "learning_rate": 4.9237563827673416e-05, "loss": 0.0845, "step": 750 }, { "epoch": 0.4110521301211379, "grad_norm": 0.1592000275850296, "learning_rate": 4.9206925286198426e-05, "loss": 0.09, "step": 755 }, { "epoch": 0.41377432965836397, "grad_norm": 0.16091205179691315, "learning_rate": 4.9175693119783013e-05, "loss": 0.0855, "step": 760 }, { "epoch": 0.41649652919559005, "grad_norm": 0.13851270079612732, "learning_rate": 4.914386809433167e-05, "loss": 0.0946, "step": 765 }, { "epoch": 0.4192187287328161, "grad_norm": 0.15238268673419952, "learning_rate": 4.911145099028753e-05, "loss": 0.0861, "step": 770 }, { "epoch": 0.4219409282700422, "grad_norm": 0.15689148008823395, "learning_rate": 4.9078442602613265e-05, "loss": 0.088, "step": 775 }, { "epoch": 0.4246631278072683, "grad_norm": 0.17214879393577576, "learning_rate": 4.9044843740771505e-05, "loss": 0.0882, "step": 780 }, { "epoch": 0.42738532734449436, "grad_norm": 0.15415912866592407, "learning_rate": 4.901065522870511e-05, "loss": 0.0978, "step": 785 }, { "epoch": 0.43010752688172044, "grad_norm": 0.165154829621315, "learning_rate": 4.897587790481683e-05, "loss": 0.0833, "step": 790 }, { "epoch": 0.4328297264189465, "grad_norm": 0.15241703391075134, "learning_rate": 4.894051262194885e-05, "loss": 0.0944, "step": 795 }, { "epoch": 0.4355519259561726, "grad_norm": 0.13514836132526398, "learning_rate": 4.8904560247361833e-05, "loss": 0.1036, "step": 800 }, { "epoch": 0.43827412549339867, "grad_norm": 0.14687085151672363, "learning_rate": 4.886802166271364e-05, "loss": 0.1042, "step": 805 }, { "epoch": 0.44099632503062475, "grad_norm": 0.143589586019516, "learning_rate": 4.8830897764037744e-05, "loss": 0.0771, "step": 810 }, { "epoch": 0.4437185245678508, "grad_norm": 0.16582971811294556, "learning_rate": 4.879318946172123e-05, "loss": 0.0992, "step": 815 }, { "epoch": 0.4464407241050769, "grad_norm": 0.1646450310945511, "learning_rate": 4.875489768048247e-05, "loss": 0.0959, "step": 820 }, { "epoch": 0.449162923642303, "grad_norm": 0.12446028739213943, "learning_rate": 4.871602335934847e-05, "loss": 0.0817, "step": 825 }, { "epoch": 0.45188512317952906, "grad_norm": 0.15374915301799774, "learning_rate": 4.867656745163182e-05, "loss": 0.0763, "step": 830 }, { "epoch": 0.45460732271675514, "grad_norm": 0.1652667224407196, "learning_rate": 4.8636530924907296e-05, "loss": 0.1026, "step": 835 }, { "epoch": 0.4573295222539812, "grad_norm": 0.13503149151802063, "learning_rate": 4.85959147609882e-05, "loss": 0.0986, "step": 840 }, { "epoch": 0.4600517217912073, "grad_norm": 0.16977562010288239, "learning_rate": 4.855471995590222e-05, "loss": 0.1099, "step": 845 }, { "epoch": 0.46277392132843337, "grad_norm": 0.14991120994091034, "learning_rate": 4.851294751986702e-05, "loss": 0.0839, "step": 850 }, { "epoch": 0.46549612086565945, "grad_norm": 0.15737441182136536, "learning_rate": 4.84705984772655e-05, "loss": 0.097, "step": 855 }, { "epoch": 0.4682183204028855, "grad_norm": 0.14336347579956055, "learning_rate": 4.8427673866620615e-05, "loss": 0.0837, "step": 860 }, { "epoch": 0.4709405199401116, "grad_norm": 0.12957650423049927, "learning_rate": 4.8384174740569944e-05, "loss": 0.0871, "step": 865 }, { "epoch": 0.4736627194773377, "grad_norm": 0.15204955637454987, "learning_rate": 4.83401021658399e-05, "loss": 0.0905, "step": 870 }, { "epoch": 0.47638491901456376, "grad_norm": 0.1311759650707245, "learning_rate": 4.82954572232195e-05, "loss": 0.1187, "step": 875 }, { "epoch": 0.47910711855178983, "grad_norm": 0.14466483891010284, "learning_rate": 4.825024100753395e-05, "loss": 0.0892, "step": 880 }, { "epoch": 0.4818293180890159, "grad_norm": 0.14845815300941467, "learning_rate": 4.820445462761771e-05, "loss": 0.0865, "step": 885 }, { "epoch": 0.484551517626242, "grad_norm": 0.15656766295433044, "learning_rate": 4.815809920628738e-05, "loss": 0.0899, "step": 890 }, { "epoch": 0.48727371716346807, "grad_norm": 0.11957818269729614, "learning_rate": 4.8111175880314084e-05, "loss": 0.0767, "step": 895 }, { "epoch": 0.48999591670069415, "grad_norm": 0.14076441526412964, "learning_rate": 4.806368580039566e-05, "loss": 0.0832, "step": 900 }, { "epoch": 0.4927181162379202, "grad_norm": 0.15563461184501648, "learning_rate": 4.801563013112844e-05, "loss": 0.0873, "step": 905 }, { "epoch": 0.4954403157751463, "grad_norm": 0.13865019381046295, "learning_rate": 4.7967010050978635e-05, "loss": 0.1083, "step": 910 }, { "epoch": 0.4981625153123724, "grad_norm": 0.1487104445695877, "learning_rate": 4.791782675225348e-05, "loss": 0.0853, "step": 915 }, { "epoch": 0.5008847148495985, "grad_norm": 0.14385932683944702, "learning_rate": 4.7868081441071975e-05, "loss": 0.0882, "step": 920 }, { "epoch": 0.5036069143868246, "grad_norm": 0.144570991396904, "learning_rate": 4.781777533733534e-05, "loss": 0.0859, "step": 925 }, { "epoch": 0.5063291139240507, "grad_norm": 0.1305367350578308, "learning_rate": 4.776690967469708e-05, "loss": 0.0788, "step": 930 }, { "epoch": 0.5090513134612767, "grad_norm": 0.14406158030033112, "learning_rate": 4.771548570053268e-05, "loss": 0.075, "step": 935 }, { "epoch": 0.5117735129985028, "grad_norm": 0.14757996797561646, "learning_rate": 4.766350467590911e-05, "loss": 0.0991, "step": 940 }, { "epoch": 0.5144957125357289, "grad_norm": 0.13629554212093353, "learning_rate": 4.7610967875553846e-05, "loss": 0.082, "step": 945 }, { "epoch": 0.517217912072955, "grad_norm": 0.1259639412164688, "learning_rate": 4.755787658782361e-05, "loss": 0.0848, "step": 950 }, { "epoch": 0.519940111610181, "grad_norm": 0.14720089733600616, "learning_rate": 4.750423211467278e-05, "loss": 0.083, "step": 955 }, { "epoch": 0.5226623111474071, "grad_norm": 0.13509656488895416, "learning_rate": 4.745003577162148e-05, "loss": 0.0856, "step": 960 }, { "epoch": 0.5253845106846332, "grad_norm": 0.15491896867752075, "learning_rate": 4.7395288887723296e-05, "loss": 0.0953, "step": 965 }, { "epoch": 0.5281067102218593, "grad_norm": 0.14023731648921967, "learning_rate": 4.73399928055327e-05, "loss": 0.0887, "step": 970 }, { "epoch": 0.5308289097590854, "grad_norm": 0.13751854002475739, "learning_rate": 4.728414888107211e-05, "loss": 0.0908, "step": 975 }, { "epoch": 0.5335511092963114, "grad_norm": 0.16226938366889954, "learning_rate": 4.722775848379866e-05, "loss": 0.0898, "step": 980 }, { "epoch": 0.5362733088335375, "grad_norm": 0.12918660044670105, "learning_rate": 4.717082299657058e-05, "loss": 0.0985, "step": 985 }, { "epoch": 0.5389955083707636, "grad_norm": 0.14950741827487946, "learning_rate": 4.711334381561333e-05, "loss": 0.0894, "step": 990 }, { "epoch": 0.5417177079079897, "grad_norm": 0.158179372549057, "learning_rate": 4.7055322350485344e-05, "loss": 0.0823, "step": 995 }, { "epoch": 0.5444399074452158, "grad_norm": 0.14156126976013184, "learning_rate": 4.699676002404342e-05, "loss": 0.0851, "step": 1000 }, { "epoch": 0.5444399074452158, "eval_loss": 0.08648888021707535, "eval_runtime": 254.9955, "eval_samples_per_second": 2.353, "eval_steps_per_second": 0.588, "step": 1000 }, { "epoch": 0.5471621069824418, "grad_norm": 0.14035391807556152, "learning_rate": 4.6937658272407905e-05, "loss": 0.0827, "step": 1005 }, { "epoch": 0.5498843065196679, "grad_norm": 0.1346469521522522, "learning_rate": 4.6878018544927415e-05, "loss": 0.0986, "step": 1010 }, { "epoch": 0.552606506056894, "grad_norm": 0.16363918781280518, "learning_rate": 4.681784230414332e-05, "loss": 0.0883, "step": 1015 }, { "epoch": 0.5553287055941201, "grad_norm": 0.14541316032409668, "learning_rate": 4.6757131025753886e-05, "loss": 0.0881, "step": 1020 }, { "epoch": 0.5580509051313461, "grad_norm": 0.1540316641330719, "learning_rate": 4.6695886198578034e-05, "loss": 0.0825, "step": 1025 }, { "epoch": 0.5607731046685722, "grad_norm": 0.15897440910339355, "learning_rate": 4.6634109324518914e-05, "loss": 0.0978, "step": 1030 }, { "epoch": 0.5634953042057983, "grad_norm": 0.13335077464580536, "learning_rate": 4.657180191852701e-05, "loss": 0.1565, "step": 1035 }, { "epoch": 0.5662175037430244, "grad_norm": 0.1566026508808136, "learning_rate": 4.6508965508562995e-05, "loss": 0.0927, "step": 1040 }, { "epoch": 0.5689397032802505, "grad_norm": 0.1556750237941742, "learning_rate": 4.644560163556031e-05, "loss": 0.0839, "step": 1045 }, { "epoch": 0.5716619028174765, "grad_norm": 0.15351781249046326, "learning_rate": 4.638171185338729e-05, "loss": 0.0812, "step": 1050 }, { "epoch": 0.5743841023547026, "grad_norm": 0.16494281589984894, "learning_rate": 4.6317297728809147e-05, "loss": 0.0889, "step": 1055 }, { "epoch": 0.5771063018919287, "grad_norm": 0.15591537952423096, "learning_rate": 4.62523608414495e-05, "loss": 0.0858, "step": 1060 }, { "epoch": 0.5798285014291548, "grad_norm": 0.1396636664867401, "learning_rate": 4.618690278375164e-05, "loss": 0.0856, "step": 1065 }, { "epoch": 0.5825507009663808, "grad_norm": 0.1384498029947281, "learning_rate": 4.61209251609395e-05, "loss": 0.082, "step": 1070 }, { "epoch": 0.5852729005036069, "grad_norm": 0.13901008665561676, "learning_rate": 4.605442959097826e-05, "loss": 0.0821, "step": 1075 }, { "epoch": 0.587995100040833, "grad_norm": 0.1609126627445221, "learning_rate": 4.5987417704534697e-05, "loss": 0.1273, "step": 1080 }, { "epoch": 0.5907172995780591, "grad_norm": 0.1475275307893753, "learning_rate": 4.591989114493718e-05, "loss": 0.0807, "step": 1085 }, { "epoch": 0.5934394991152852, "grad_norm": 0.14419402182102203, "learning_rate": 4.5851851568135376e-05, "loss": 0.0737, "step": 1090 }, { "epoch": 0.5961616986525112, "grad_norm": 0.13583903014659882, "learning_rate": 4.5783300642659644e-05, "loss": 0.0904, "step": 1095 }, { "epoch": 0.5988838981897373, "grad_norm": 0.1405334770679474, "learning_rate": 4.571424004958012e-05, "loss": 0.0863, "step": 1100 }, { "epoch": 0.6016060977269634, "grad_norm": 0.1326403170824051, "learning_rate": 4.564467148246548e-05, "loss": 0.0839, "step": 1105 }, { "epoch": 0.6043282972641895, "grad_norm": 0.14316369593143463, "learning_rate": 4.557459664734141e-05, "loss": 0.0862, "step": 1110 }, { "epoch": 0.6070504968014155, "grad_norm": 0.14497336745262146, "learning_rate": 4.550401726264879e-05, "loss": 0.0733, "step": 1115 }, { "epoch": 0.6097726963386416, "grad_norm": 0.15112735331058502, "learning_rate": 4.5432935059201544e-05, "loss": 0.0712, "step": 1120 }, { "epoch": 0.6124948958758677, "grad_norm": 0.1457735002040863, "learning_rate": 4.536135178014415e-05, "loss": 0.0808, "step": 1125 }, { "epoch": 0.6152170954130938, "grad_norm": 0.14395588636398315, "learning_rate": 4.528926918090898e-05, "loss": 0.0863, "step": 1130 }, { "epoch": 0.6179392949503199, "grad_norm": 0.13603851199150085, "learning_rate": 4.521668902917317e-05, "loss": 0.0848, "step": 1135 }, { "epoch": 0.6206614944875459, "grad_norm": 0.16027645766735077, "learning_rate": 4.514361310481533e-05, "loss": 0.0929, "step": 1140 }, { "epoch": 0.623383694024772, "grad_norm": 0.13826999068260193, "learning_rate": 4.507004319987185e-05, "loss": 0.0914, "step": 1145 }, { "epoch": 0.6261058935619981, "grad_norm": 0.14140823483467102, "learning_rate": 4.499598111849299e-05, "loss": 0.0885, "step": 1150 }, { "epoch": 0.6288280930992242, "grad_norm": 0.13926826417446136, "learning_rate": 4.492142867689861e-05, "loss": 0.083, "step": 1155 }, { "epoch": 0.6315502926364502, "grad_norm": 0.148858904838562, "learning_rate": 4.484638770333367e-05, "loss": 0.0866, "step": 1160 }, { "epoch": 0.6342724921736763, "grad_norm": 0.12866345047950745, "learning_rate": 4.4770860038023335e-05, "loss": 0.0991, "step": 1165 }, { "epoch": 0.6369946917109024, "grad_norm": 0.13434378802776337, "learning_rate": 4.4694847533127903e-05, "loss": 0.079, "step": 1170 }, { "epoch": 0.6397168912481285, "grad_norm": 0.13786840438842773, "learning_rate": 4.461835205269736e-05, "loss": 0.0842, "step": 1175 }, { "epoch": 0.6424390907853545, "grad_norm": 0.12938237190246582, "learning_rate": 4.454137547262566e-05, "loss": 0.0771, "step": 1180 }, { "epoch": 0.6451612903225806, "grad_norm": 0.13730603456497192, "learning_rate": 4.446391968060475e-05, "loss": 0.0831, "step": 1185 }, { "epoch": 0.6478834898598067, "grad_norm": 0.142304465174675, "learning_rate": 4.4385986576078254e-05, "loss": 0.0918, "step": 1190 }, { "epoch": 0.6506056893970328, "grad_norm": 0.1353188157081604, "learning_rate": 4.43075780701949e-05, "loss": 0.0897, "step": 1195 }, { "epoch": 0.6533278889342589, "grad_norm": 0.1252220869064331, "learning_rate": 4.422869608576167e-05, "loss": 0.1353, "step": 1200 }, { "epoch": 0.6560500884714849, "grad_norm": 0.14703655242919922, "learning_rate": 4.4149342557196605e-05, "loss": 0.2083, "step": 1205 }, { "epoch": 0.658772288008711, "grad_norm": 0.13858124613761902, "learning_rate": 4.406951943048141e-05, "loss": 0.0824, "step": 1210 }, { "epoch": 0.6614944875459371, "grad_norm": 0.1428638994693756, "learning_rate": 4.3989228663113714e-05, "loss": 0.0774, "step": 1215 }, { "epoch": 0.6642166870831632, "grad_norm": 0.13073915243148804, "learning_rate": 4.3908472224059064e-05, "loss": 0.0788, "step": 1220 }, { "epoch": 0.6669388866203892, "grad_norm": 0.14101268351078033, "learning_rate": 4.3827252093702656e-05, "loss": 0.0882, "step": 1225 }, { "epoch": 0.6696610861576153, "grad_norm": 0.13420827686786652, "learning_rate": 4.374557026380075e-05, "loss": 0.0783, "step": 1230 }, { "epoch": 0.6723832856948414, "grad_norm": 0.12414020299911499, "learning_rate": 4.366342873743185e-05, "loss": 0.0874, "step": 1235 }, { "epoch": 0.6751054852320675, "grad_norm": 0.12837223708629608, "learning_rate": 4.358082952894753e-05, "loss": 0.0863, "step": 1240 }, { "epoch": 0.6778276847692936, "grad_norm": 0.14144425094127655, "learning_rate": 4.349777466392313e-05, "loss": 0.0784, "step": 1245 }, { "epoch": 0.6805498843065196, "grad_norm": 0.13347384333610535, "learning_rate": 4.341426617910798e-05, "loss": 0.0823, "step": 1250 }, { "epoch": 0.6832720838437457, "grad_norm": 0.1294572502374649, "learning_rate": 4.3330306122375516e-05, "loss": 0.0787, "step": 1255 }, { "epoch": 0.6859942833809718, "grad_norm": 0.13768264651298523, "learning_rate": 4.324589655267306e-05, "loss": 0.0812, "step": 1260 }, { "epoch": 0.6887164829181979, "grad_norm": 0.14431719481945038, "learning_rate": 4.3161039539971295e-05, "loss": 0.0917, "step": 1265 }, { "epoch": 0.691438682455424, "grad_norm": 0.13428707420825958, "learning_rate": 4.307573716521353e-05, "loss": 0.0866, "step": 1270 }, { "epoch": 0.69416088199265, "grad_norm": 0.14333197474479675, "learning_rate": 4.298999152026465e-05, "loss": 0.0822, "step": 1275 }, { "epoch": 0.6968830815298761, "grad_norm": 0.13545657694339752, "learning_rate": 4.2903804707859835e-05, "loss": 0.1482, "step": 1280 }, { "epoch": 0.6996052810671022, "grad_norm": 0.13254009187221527, "learning_rate": 4.281717884155298e-05, "loss": 0.0792, "step": 1285 }, { "epoch": 0.7023274806043283, "grad_norm": 0.12849317491054535, "learning_rate": 4.2730116045664905e-05, "loss": 0.0909, "step": 1290 }, { "epoch": 0.7050496801415543, "grad_norm": 0.15678727626800537, "learning_rate": 4.264261845523116e-05, "loss": 0.0902, "step": 1295 }, { "epoch": 0.7077718796787804, "grad_norm": 0.1492096483707428, "learning_rate": 4.255468821594981e-05, "loss": 0.0796, "step": 1300 }, { "epoch": 0.7104940792160065, "grad_norm": 0.12939919531345367, "learning_rate": 4.2466327484128685e-05, "loss": 0.0929, "step": 1305 }, { "epoch": 0.7132162787532326, "grad_norm": 0.13585573434829712, "learning_rate": 4.2377538426632595e-05, "loss": 0.0997, "step": 1310 }, { "epoch": 0.7159384782904586, "grad_norm": 0.15193824470043182, "learning_rate": 4.228832322083013e-05, "loss": 0.0806, "step": 1315 }, { "epoch": 0.7186606778276847, "grad_norm": 0.1575016975402832, "learning_rate": 4.2198684054540285e-05, "loss": 0.0877, "step": 1320 }, { "epoch": 0.7213828773649108, "grad_norm": 0.12813854217529297, "learning_rate": 4.210862312597884e-05, "loss": 0.0802, "step": 1325 }, { "epoch": 0.7241050769021369, "grad_norm": 0.13729339838027954, "learning_rate": 4.201814264370441e-05, "loss": 0.0816, "step": 1330 }, { "epoch": 0.726827276439363, "grad_norm": 0.12651780247688293, "learning_rate": 4.192724482656428e-05, "loss": 0.0812, "step": 1335 }, { "epoch": 0.729549475976589, "grad_norm": 0.13558164238929749, "learning_rate": 4.1835931903640046e-05, "loss": 0.0926, "step": 1340 }, { "epoch": 0.7322716755138151, "grad_norm": 0.1335664838552475, "learning_rate": 4.17442061141929e-05, "loss": 0.078, "step": 1345 }, { "epoch": 0.7349938750510412, "grad_norm": 0.12852634489536285, "learning_rate": 4.165206970760874e-05, "loss": 0.0805, "step": 1350 }, { "epoch": 0.7377160745882673, "grad_norm": 0.14529520273208618, "learning_rate": 4.1559524943342985e-05, "loss": 0.0837, "step": 1355 }, { "epoch": 0.7404382741254933, "grad_norm": 0.14452920854091644, "learning_rate": 4.1466574090865225e-05, "loss": 0.0857, "step": 1360 }, { "epoch": 0.7431604736627194, "grad_norm": 0.13917720317840576, "learning_rate": 4.1373219429603473e-05, "loss": 0.0814, "step": 1365 }, { "epoch": 0.7458826731999455, "grad_norm": 0.11816851049661636, "learning_rate": 4.127946324888836e-05, "loss": 0.0743, "step": 1370 }, { "epoch": 0.7486048727371717, "grad_norm": 0.14422592520713806, "learning_rate": 4.118530784789694e-05, "loss": 0.0768, "step": 1375 }, { "epoch": 0.7513270722743978, "grad_norm": 0.14289607107639313, "learning_rate": 4.109075553559633e-05, "loss": 0.0806, "step": 1380 }, { "epoch": 0.7540492718116238, "grad_norm": 0.14432035386562347, "learning_rate": 4.099580863068706e-05, "loss": 0.0915, "step": 1385 }, { "epoch": 0.7567714713488499, "grad_norm": 0.16059063374996185, "learning_rate": 4.0900469461546235e-05, "loss": 0.0799, "step": 1390 }, { "epoch": 0.759493670886076, "grad_norm": 0.14347511529922485, "learning_rate": 4.0804740366170454e-05, "loss": 0.0899, "step": 1395 }, { "epoch": 0.7622158704233021, "grad_norm": 0.12717047333717346, "learning_rate": 4.070862369211843e-05, "loss": 0.0827, "step": 1400 }, { "epoch": 0.7649380699605282, "grad_norm": 0.12884029746055603, "learning_rate": 4.061212179645345e-05, "loss": 0.0856, "step": 1405 }, { "epoch": 0.7676602694977542, "grad_norm": 0.14441439509391785, "learning_rate": 4.051523704568557e-05, "loss": 0.1416, "step": 1410 }, { "epoch": 0.7703824690349803, "grad_norm": 0.12841658294200897, "learning_rate": 4.0417971815713584e-05, "loss": 0.0827, "step": 1415 }, { "epoch": 0.7731046685722064, "grad_norm": 0.14411351084709167, "learning_rate": 4.032032849176673e-05, "loss": 0.0808, "step": 1420 }, { "epoch": 0.7758268681094325, "grad_norm": 0.13570953905582428, "learning_rate": 4.022230946834624e-05, "loss": 0.08, "step": 1425 }, { "epoch": 0.7785490676466585, "grad_norm": 0.13861487805843353, "learning_rate": 4.012391714916661e-05, "loss": 0.0867, "step": 1430 }, { "epoch": 0.7812712671838846, "grad_norm": 0.14143118262290955, "learning_rate": 4.0025153947096624e-05, "loss": 0.0865, "step": 1435 }, { "epoch": 0.7839934667211107, "grad_norm": 0.1405879408121109, "learning_rate": 3.992602228410023e-05, "loss": 0.0983, "step": 1440 }, { "epoch": 0.7867156662583368, "grad_norm": 0.14435367286205292, "learning_rate": 3.982652459117707e-05, "loss": 0.0829, "step": 1445 }, { "epoch": 0.7894378657955629, "grad_norm": 0.14033889770507812, "learning_rate": 3.972666330830299e-05, "loss": 0.0903, "step": 1450 }, { "epoch": 0.7921600653327889, "grad_norm": 0.13285259902477264, "learning_rate": 3.9626440884370065e-05, "loss": 0.0748, "step": 1455 }, { "epoch": 0.794882264870015, "grad_norm": 0.1237478256225586, "learning_rate": 3.952585977712664e-05, "loss": 0.083, "step": 1460 }, { "epoch": 0.7976044644072411, "grad_norm": 0.14227712154388428, "learning_rate": 3.942492245311703e-05, "loss": 0.0926, "step": 1465 }, { "epoch": 0.8003266639444672, "grad_norm": 0.13259918987751007, "learning_rate": 3.9323631387621015e-05, "loss": 0.0841, "step": 1470 }, { "epoch": 0.8030488634816932, "grad_norm": 0.12796998023986816, "learning_rate": 3.9221989064593175e-05, "loss": 0.076, "step": 1475 }, { "epoch": 0.8057710630189193, "grad_norm": 0.13424307107925415, "learning_rate": 3.9119997976601954e-05, "loss": 0.097, "step": 1480 }, { "epoch": 0.8084932625561454, "grad_norm": 0.12660135328769684, "learning_rate": 3.9017660624768515e-05, "loss": 0.0834, "step": 1485 }, { "epoch": 0.8112154620933715, "grad_norm": 0.12413927167654037, "learning_rate": 3.891497951870545e-05, "loss": 0.0765, "step": 1490 }, { "epoch": 0.8139376616305976, "grad_norm": 0.1326495110988617, "learning_rate": 3.8811957176455214e-05, "loss": 0.0832, "step": 1495 }, { "epoch": 0.8166598611678236, "grad_norm": 0.133337140083313, "learning_rate": 3.870859612442837e-05, "loss": 0.0819, "step": 1500 }, { "epoch": 0.8166598611678236, "eval_loss": 0.08255165070295334, "eval_runtime": 238.131, "eval_samples_per_second": 2.52, "eval_steps_per_second": 0.63, "step": 1500 }, { "epoch": 0.8193820607050497, "grad_norm": 0.14871715009212494, "learning_rate": 3.8604898897341644e-05, "loss": 0.1878, "step": 1505 }, { "epoch": 0.8221042602422758, "grad_norm": 0.1360073983669281, "learning_rate": 3.850086803815576e-05, "loss": 0.0879, "step": 1510 }, { "epoch": 0.8248264597795019, "grad_norm": 0.12314160168170929, "learning_rate": 3.8396506098013076e-05, "loss": 0.0767, "step": 1515 }, { "epoch": 0.8275486593167279, "grad_norm": 0.14723584055900574, "learning_rate": 3.829181563617503e-05, "loss": 0.0972, "step": 1520 }, { "epoch": 0.830270858853954, "grad_norm": 0.14206384122371674, "learning_rate": 3.81867992199594e-05, "loss": 0.0839, "step": 1525 }, { "epoch": 0.8329930583911801, "grad_norm": 0.11804749816656113, "learning_rate": 3.808145942467729e-05, "loss": 0.0923, "step": 1530 }, { "epoch": 0.8357152579284062, "grad_norm": 0.12313738465309143, "learning_rate": 3.797579883357002e-05, "loss": 0.0872, "step": 1535 }, { "epoch": 0.8384374574656323, "grad_norm": 0.13978321850299835, "learning_rate": 3.7869820037745776e-05, "loss": 0.1243, "step": 1540 }, { "epoch": 0.8411596570028583, "grad_norm": 0.12310109287500381, "learning_rate": 3.776352563611604e-05, "loss": 0.0839, "step": 1545 }, { "epoch": 0.8438818565400844, "grad_norm": 0.13480614125728607, "learning_rate": 3.765691823533191e-05, "loss": 0.0873, "step": 1550 }, { "epoch": 0.8466040560773105, "grad_norm": 0.13653361797332764, "learning_rate": 3.75500004497201e-05, "loss": 0.0771, "step": 1555 }, { "epoch": 0.8493262556145366, "grad_norm": 0.1283886879682541, "learning_rate": 3.74427749012189e-05, "loss": 0.0816, "step": 1560 }, { "epoch": 0.8520484551517626, "grad_norm": 0.12764351069927216, "learning_rate": 3.733524421931385e-05, "loss": 0.0859, "step": 1565 }, { "epoch": 0.8547706546889887, "grad_norm": 0.1281007081270218, "learning_rate": 3.722741104097323e-05, "loss": 0.0806, "step": 1570 }, { "epoch": 0.8574928542262148, "grad_norm": 0.1311092972755432, "learning_rate": 3.711927801058347e-05, "loss": 0.0792, "step": 1575 }, { "epoch": 0.8602150537634409, "grad_norm": 0.11376714706420898, "learning_rate": 3.7010847779884204e-05, "loss": 0.0763, "step": 1580 }, { "epoch": 0.862937253300667, "grad_norm": 0.14167912304401398, "learning_rate": 3.690212300790333e-05, "loss": 0.0871, "step": 1585 }, { "epoch": 0.865659452837893, "grad_norm": 0.13373591005802155, "learning_rate": 3.679310636089174e-05, "loss": 0.1189, "step": 1590 }, { "epoch": 0.8683816523751191, "grad_norm": 0.14792831242084503, "learning_rate": 3.668380051225794e-05, "loss": 0.0811, "step": 1595 }, { "epoch": 0.8711038519123452, "grad_norm": 0.14200051128864288, "learning_rate": 3.657420814250258e-05, "loss": 0.086, "step": 1600 }, { "epoch": 0.8738260514495713, "grad_norm": 0.1501314491033554, "learning_rate": 3.646433193915257e-05, "loss": 0.0885, "step": 1605 }, { "epoch": 0.8765482509867973, "grad_norm": 0.10102449357509613, "learning_rate": 3.635417459669532e-05, "loss": 0.0738, "step": 1610 }, { "epoch": 0.8792704505240234, "grad_norm": 0.12374872714281082, "learning_rate": 3.624373881651254e-05, "loss": 0.0883, "step": 1615 }, { "epoch": 0.8819926500612495, "grad_norm": 0.1358879953622818, "learning_rate": 3.6133027306814085e-05, "loss": 0.0869, "step": 1620 }, { "epoch": 0.8847148495984756, "grad_norm": 0.12947793304920197, "learning_rate": 3.6022042782571494e-05, "loss": 0.0868, "step": 1625 }, { "epoch": 0.8874370491357016, "grad_norm": 0.14391177892684937, "learning_rate": 3.591078796545144e-05, "loss": 0.0829, "step": 1630 }, { "epoch": 0.8901592486729277, "grad_norm": 0.13038809597492218, "learning_rate": 3.579926558374897e-05, "loss": 0.0888, "step": 1635 }, { "epoch": 0.8928814482101538, "grad_norm": 0.15558017790317535, "learning_rate": 3.5687478372320576e-05, "loss": 0.0853, "step": 1640 }, { "epoch": 0.8956036477473799, "grad_norm": 0.1262005716562271, "learning_rate": 3.557542907251718e-05, "loss": 0.0763, "step": 1645 }, { "epoch": 0.898325847284606, "grad_norm": 0.1222100704908371, "learning_rate": 3.546312043211687e-05, "loss": 0.0835, "step": 1650 }, { "epoch": 0.901048046821832, "grad_norm": 0.14368890225887299, "learning_rate": 3.535055520525753e-05, "loss": 0.0795, "step": 1655 }, { "epoch": 0.9037702463590581, "grad_norm": 0.13927970826625824, "learning_rate": 3.52377361523693e-05, "loss": 0.0838, "step": 1660 }, { "epoch": 0.9064924458962842, "grad_norm": 0.14974816143512726, "learning_rate": 3.512466604010688e-05, "loss": 0.0834, "step": 1665 }, { "epoch": 0.9092146454335103, "grad_norm": 0.14425551891326904, "learning_rate": 3.501134764128167e-05, "loss": 0.0933, "step": 1670 }, { "epoch": 0.9119368449707363, "grad_norm": 0.13006910681724548, "learning_rate": 3.4897783734793794e-05, "loss": 0.0829, "step": 1675 }, { "epoch": 0.9146590445079624, "grad_norm": 0.10913355648517609, "learning_rate": 3.478397710556397e-05, "loss": 0.0775, "step": 1680 }, { "epoch": 0.9173812440451885, "grad_norm": 0.13194523751735687, "learning_rate": 3.466993054446515e-05, "loss": 0.0762, "step": 1685 }, { "epoch": 0.9201034435824146, "grad_norm": 0.1317387819290161, "learning_rate": 3.4555646848254136e-05, "loss": 0.094, "step": 1690 }, { "epoch": 0.9228256431196407, "grad_norm": 0.1169929951429367, "learning_rate": 3.4441128819503e-05, "loss": 0.0839, "step": 1695 }, { "epoch": 0.9255478426568667, "grad_norm": 0.13555140793323517, "learning_rate": 3.4326379266530314e-05, "loss": 0.0831, "step": 1700 }, { "epoch": 0.9282700421940928, "grad_norm": 0.1173190027475357, "learning_rate": 3.421140100333231e-05, "loss": 0.0842, "step": 1705 }, { "epoch": 0.9309922417313189, "grad_norm": 0.1393524408340454, "learning_rate": 3.409619684951386e-05, "loss": 0.1127, "step": 1710 }, { "epoch": 0.933714441268545, "grad_norm": 0.12765948474407196, "learning_rate": 3.3980769630219354e-05, "loss": 0.0781, "step": 1715 }, { "epoch": 0.936436640805771, "grad_norm": 0.11775423586368561, "learning_rate": 3.386512217606339e-05, "loss": 0.071, "step": 1720 }, { "epoch": 0.9391588403429971, "grad_norm": 0.1447712928056717, "learning_rate": 3.3749257323061376e-05, "loss": 0.0852, "step": 1725 }, { "epoch": 0.9418810398802232, "grad_norm": 0.14077867567539215, "learning_rate": 3.3633177912559984e-05, "loss": 0.0793, "step": 1730 }, { "epoch": 0.9446032394174493, "grad_norm": 0.13743150234222412, "learning_rate": 3.3516886791167444e-05, "loss": 0.0804, "step": 1735 }, { "epoch": 0.9473254389546754, "grad_norm": 0.13769537210464478, "learning_rate": 3.34003868106838e-05, "loss": 0.0865, "step": 1740 }, { "epoch": 0.9500476384919014, "grad_norm": 0.1354909986257553, "learning_rate": 3.328368082803088e-05, "loss": 0.0793, "step": 1745 }, { "epoch": 0.9527698380291275, "grad_norm": 0.12832538783550262, "learning_rate": 3.316677170518235e-05, "loss": 0.0769, "step": 1750 }, { "epoch": 0.9554920375663536, "grad_norm": 0.1272445023059845, "learning_rate": 3.304966230909342e-05, "loss": 0.0931, "step": 1755 }, { "epoch": 0.9582142371035797, "grad_norm": 0.14885401725769043, "learning_rate": 3.293235551163063e-05, "loss": 0.0835, "step": 1760 }, { "epoch": 0.9609364366408057, "grad_norm": 0.13052543997764587, "learning_rate": 3.2814854189501346e-05, "loss": 0.0866, "step": 1765 }, { "epoch": 0.9636586361780318, "grad_norm": 0.12114793807268143, "learning_rate": 3.269716122418326e-05, "loss": 0.0851, "step": 1770 }, { "epoch": 0.9663808357152579, "grad_norm": 0.13619346916675568, "learning_rate": 3.2579279501853744e-05, "loss": 0.0834, "step": 1775 }, { "epoch": 0.969103035252484, "grad_norm": 0.13156457245349884, "learning_rate": 3.246121191331902e-05, "loss": 0.0868, "step": 1780 }, { "epoch": 0.9718252347897101, "grad_norm": 0.1379031538963318, "learning_rate": 3.234296135394329e-05, "loss": 0.081, "step": 1785 }, { "epoch": 0.9745474343269361, "grad_norm": 0.1381215900182724, "learning_rate": 3.222453072357777e-05, "loss": 0.0758, "step": 1790 }, { "epoch": 0.9772696338641622, "grad_norm": 0.11675938963890076, "learning_rate": 3.2105922926489504e-05, "loss": 0.1002, "step": 1795 }, { "epoch": 0.9799918334013883, "grad_norm": 0.12285912781953812, "learning_rate": 3.1987140871290236e-05, "loss": 0.0736, "step": 1800 }, { "epoch": 0.9827140329386144, "grad_norm": 0.13735820353031158, "learning_rate": 3.1868187470864984e-05, "loss": 0.0779, "step": 1805 }, { "epoch": 0.9854362324758404, "grad_norm": 0.13905146718025208, "learning_rate": 3.1749065642300674e-05, "loss": 0.0833, "step": 1810 }, { "epoch": 0.9881584320130665, "grad_norm": 0.13085448741912842, "learning_rate": 3.1629778306814585e-05, "loss": 0.0859, "step": 1815 }, { "epoch": 0.9908806315502926, "grad_norm": 0.13941799104213715, "learning_rate": 3.151032838968271e-05, "loss": 0.0845, "step": 1820 }, { "epoch": 0.9936028310875187, "grad_norm": 0.12749053537845612, "learning_rate": 3.139071882016802e-05, "loss": 0.0935, "step": 1825 }, { "epoch": 0.9963250306247448, "grad_norm": 0.12027429789304733, "learning_rate": 3.127095253144864e-05, "loss": 0.077, "step": 1830 }, { "epoch": 0.9990472301619708, "grad_norm": 0.13250844180583954, "learning_rate": 3.11510324605459e-05, "loss": 0.0795, "step": 1835 }, { "epoch": 1.001769429699197, "grad_norm": 0.12925724685192108, "learning_rate": 3.103096154825233e-05, "loss": 0.1021, "step": 1840 }, { "epoch": 1.004491629236423, "grad_norm": 0.1351374089717865, "learning_rate": 3.0910742739059527e-05, "loss": 0.0833, "step": 1845 }, { "epoch": 1.0072138287736492, "grad_norm": 0.13750001788139343, "learning_rate": 3.0790378981085956e-05, "loss": 0.076, "step": 1850 }, { "epoch": 1.0099360283108751, "grad_norm": 0.1203581839799881, "learning_rate": 3.0669873226004655e-05, "loss": 0.0769, "step": 1855 }, { "epoch": 1.0126582278481013, "grad_norm": 0.12651003897190094, "learning_rate": 3.054922842897084e-05, "loss": 0.0729, "step": 1860 }, { "epoch": 1.0153804273853273, "grad_norm": 0.14429447054862976, "learning_rate": 3.0428447548549467e-05, "loss": 0.0772, "step": 1865 }, { "epoch": 1.0181026269225535, "grad_norm": 0.13154245913028717, "learning_rate": 3.030753354664262e-05, "loss": 0.1681, "step": 1870 }, { "epoch": 1.0208248264597795, "grad_norm": 0.11439166218042374, "learning_rate": 3.018648938841695e-05, "loss": 0.0753, "step": 1875 }, { "epoch": 1.0235470259970056, "grad_norm": 0.12332076579332352, "learning_rate": 3.00653180422309e-05, "loss": 0.0746, "step": 1880 }, { "epoch": 1.0262692255342316, "grad_norm": 0.12956227362155914, "learning_rate": 2.994402247956194e-05, "loss": 0.0733, "step": 1885 }, { "epoch": 1.0289914250714578, "grad_norm": 0.13542629778385162, "learning_rate": 2.9822605674933696e-05, "loss": 0.076, "step": 1890 }, { "epoch": 1.0317136246086838, "grad_norm": 0.14228489995002747, "learning_rate": 2.9701070605843e-05, "loss": 0.0756, "step": 1895 }, { "epoch": 1.03443582414591, "grad_norm": 0.13206136226654053, "learning_rate": 2.957942025268689e-05, "loss": 0.0733, "step": 1900 }, { "epoch": 1.037158023683136, "grad_norm": 0.14194877445697784, "learning_rate": 2.945765759868949e-05, "loss": 0.0837, "step": 1905 }, { "epoch": 1.039880223220362, "grad_norm": 0.1430080085992813, "learning_rate": 2.933578562982888e-05, "loss": 0.0814, "step": 1910 }, { "epoch": 1.042602422757588, "grad_norm": 0.12564821541309357, "learning_rate": 2.9213807334763854e-05, "loss": 0.0748, "step": 1915 }, { "epoch": 1.0453246222948143, "grad_norm": 0.12261082231998444, "learning_rate": 2.9091725704760638e-05, "loss": 0.0668, "step": 1920 }, { "epoch": 1.0480468218320402, "grad_norm": 0.14175981283187866, "learning_rate": 2.8969543733619554e-05, "loss": 0.0742, "step": 1925 }, { "epoch": 1.0507690213692664, "grad_norm": 0.1397016942501068, "learning_rate": 2.884726441760155e-05, "loss": 0.0716, "step": 1930 }, { "epoch": 1.0534912209064924, "grad_norm": 0.1348811835050583, "learning_rate": 2.8724890755354783e-05, "loss": 0.0882, "step": 1935 }, { "epoch": 1.0562134204437186, "grad_norm": 0.11297186464071274, "learning_rate": 2.8602425747841057e-05, "loss": 0.0715, "step": 1940 }, { "epoch": 1.0589356199809445, "grad_norm": 0.1416822224855423, "learning_rate": 2.8479872398262198e-05, "loss": 0.0805, "step": 1945 }, { "epoch": 1.0616578195181707, "grad_norm": 0.147059828042984, "learning_rate": 2.8357233711986487e-05, "loss": 0.0764, "step": 1950 }, { "epoch": 1.0643800190553967, "grad_norm": 0.137193962931633, "learning_rate": 2.8234512696474875e-05, "loss": 0.0733, "step": 1955 }, { "epoch": 1.0671022185926229, "grad_norm": 0.12843886017799377, "learning_rate": 2.8111712361207283e-05, "loss": 0.0751, "step": 1960 }, { "epoch": 1.0698244181298489, "grad_norm": 0.1351711004972458, "learning_rate": 2.7988835717608784e-05, "loss": 0.0776, "step": 1965 }, { "epoch": 1.072546617667075, "grad_norm": 0.1511557698249817, "learning_rate": 2.7865885778975743e-05, "loss": 0.0764, "step": 1970 }, { "epoch": 1.075268817204301, "grad_norm": 0.12794414162635803, "learning_rate": 2.774286556040196e-05, "loss": 0.0792, "step": 1975 }, { "epoch": 1.0779910167415272, "grad_norm": 0.13065066933631897, "learning_rate": 2.7619778078704685e-05, "loss": 0.0706, "step": 1980 }, { "epoch": 1.0807132162787532, "grad_norm": 0.13252925872802734, "learning_rate": 2.7496626352350664e-05, "loss": 0.0826, "step": 1985 }, { "epoch": 1.0834354158159794, "grad_norm": 0.1248321682214737, "learning_rate": 2.7373413401382104e-05, "loss": 0.076, "step": 1990 }, { "epoch": 1.0861576153532053, "grad_norm": 0.14091134071350098, "learning_rate": 2.7250142247342637e-05, "loss": 0.1298, "step": 1995 }, { "epoch": 1.0888798148904315, "grad_norm": 0.13955925405025482, "learning_rate": 2.7126815913203178e-05, "loss": 0.0753, "step": 2000 }, { "epoch": 1.0888798148904315, "eval_loss": 0.08035612851381302, "eval_runtime": 228.5599, "eval_samples_per_second": 2.625, "eval_steps_per_second": 0.656, "step": 2000 }, { "epoch": 1.0916020144276575, "grad_norm": 0.1300753802061081, "learning_rate": 2.7003437423287857e-05, "loss": 0.1036, "step": 2005 }, { "epoch": 1.0943242139648837, "grad_norm": 0.12375082820653915, "learning_rate": 2.688000980319979e-05, "loss": 0.0744, "step": 2010 }, { "epoch": 1.0970464135021096, "grad_norm": 0.14749085903167725, "learning_rate": 2.6756536079746907e-05, "loss": 0.0738, "step": 2015 }, { "epoch": 1.0997686130393358, "grad_norm": 0.13898104429244995, "learning_rate": 2.663301928086774e-05, "loss": 0.0701, "step": 2020 }, { "epoch": 1.1024908125765618, "grad_norm": 0.12244424223899841, "learning_rate": 2.6509462435557152e-05, "loss": 0.0707, "step": 2025 }, { "epoch": 1.105213012113788, "grad_norm": 0.13717930018901825, "learning_rate": 2.6385868573792072e-05, "loss": 0.07, "step": 2030 }, { "epoch": 1.107935211651014, "grad_norm": 0.16028448939323425, "learning_rate": 2.6262240726457167e-05, "loss": 0.0815, "step": 2035 }, { "epoch": 1.1106574111882401, "grad_norm": 0.13453496992588043, "learning_rate": 2.6138581925270533e-05, "loss": 0.0802, "step": 2040 }, { "epoch": 1.113379610725466, "grad_norm": 0.14346693456172943, "learning_rate": 2.6014895202709354e-05, "loss": 0.0777, "step": 2045 }, { "epoch": 1.1161018102626923, "grad_norm": 0.12791860103607178, "learning_rate": 2.5891183591935515e-05, "loss": 0.0798, "step": 2050 }, { "epoch": 1.1188240097999183, "grad_norm": 0.1344967633485794, "learning_rate": 2.5767450126721255e-05, "loss": 0.0679, "step": 2055 }, { "epoch": 1.1215462093371444, "grad_norm": 0.1335950344800949, "learning_rate": 2.564369784137472e-05, "loss": 0.0742, "step": 2060 }, { "epoch": 1.1242684088743704, "grad_norm": 0.12762115895748138, "learning_rate": 2.5519929770665597e-05, "loss": 0.0885, "step": 2065 }, { "epoch": 1.1269906084115966, "grad_norm": 0.1343313604593277, "learning_rate": 2.539614894975067e-05, "loss": 0.0672, "step": 2070 }, { "epoch": 1.1297128079488226, "grad_norm": 0.13929401338100433, "learning_rate": 2.5272358414099413e-05, "loss": 0.0765, "step": 2075 }, { "epoch": 1.1324350074860488, "grad_norm": 0.1336519867181778, "learning_rate": 2.5148561199419516e-05, "loss": 0.0728, "step": 2080 }, { "epoch": 1.1351572070232747, "grad_norm": 0.13433049619197845, "learning_rate": 2.5024760341582453e-05, "loss": 0.0785, "step": 2085 }, { "epoch": 1.137879406560501, "grad_norm": 0.12692618370056152, "learning_rate": 2.490095887654906e-05, "loss": 0.0971, "step": 2090 }, { "epoch": 1.1406016060977269, "grad_norm": 0.1449315845966339, "learning_rate": 2.4777159840295046e-05, "loss": 0.0669, "step": 2095 }, { "epoch": 1.143323805634953, "grad_norm": 0.14388057589530945, "learning_rate": 2.4653366268736565e-05, "loss": 0.0776, "step": 2100 }, { "epoch": 1.146046005172179, "grad_norm": 0.12486547976732254, "learning_rate": 2.452958119765577e-05, "loss": 0.0782, "step": 2105 }, { "epoch": 1.1487682047094052, "grad_norm": 0.135404571890831, "learning_rate": 2.4405807662626338e-05, "loss": 0.0792, "step": 2110 }, { "epoch": 1.1514904042466312, "grad_norm": 0.14270424842834473, "learning_rate": 2.4282048698939062e-05, "loss": 0.0729, "step": 2115 }, { "epoch": 1.1542126037838574, "grad_norm": 0.13348889350891113, "learning_rate": 2.41583073415274e-05, "loss": 0.0842, "step": 2120 }, { "epoch": 1.1569348033210836, "grad_norm": 0.15408015251159668, "learning_rate": 2.403458662489304e-05, "loss": 0.0903, "step": 2125 }, { "epoch": 1.1596570028583095, "grad_norm": 0.12234120815992355, "learning_rate": 2.3910889583031533e-05, "loss": 0.0793, "step": 2130 }, { "epoch": 1.1623792023955355, "grad_norm": 0.15008710324764252, "learning_rate": 2.3787219249357803e-05, "loss": 0.0759, "step": 2135 }, { "epoch": 1.1651014019327617, "grad_norm": 0.133488729596138, "learning_rate": 2.3663578656631858e-05, "loss": 0.0783, "step": 2140 }, { "epoch": 1.1678236014699879, "grad_norm": 0.12147729843854904, "learning_rate": 2.3539970836884347e-05, "loss": 0.0816, "step": 2145 }, { "epoch": 1.1705458010072138, "grad_norm": 0.14398206770420074, "learning_rate": 2.3416398821342238e-05, "loss": 0.0739, "step": 2150 }, { "epoch": 1.1732680005444398, "grad_norm": 0.14384213089942932, "learning_rate": 2.329286564035446e-05, "loss": 0.0764, "step": 2155 }, { "epoch": 1.175990200081666, "grad_norm": 0.1491377055644989, "learning_rate": 2.3169374323317624e-05, "loss": 0.0716, "step": 2160 }, { "epoch": 1.1787123996188922, "grad_norm": 0.13119584321975708, "learning_rate": 2.3045927898601702e-05, "loss": 0.0737, "step": 2165 }, { "epoch": 1.1814345991561181, "grad_norm": 0.13395731151103973, "learning_rate": 2.292252939347577e-05, "loss": 0.0674, "step": 2170 }, { "epoch": 1.1841567986933441, "grad_norm": 0.1500498354434967, "learning_rate": 2.27991818340338e-05, "loss": 0.0872, "step": 2175 }, { "epoch": 1.1868789982305703, "grad_norm": 0.15766474604606628, "learning_rate": 2.2675888245120382e-05, "loss": 0.1069, "step": 2180 }, { "epoch": 1.1896011977677965, "grad_norm": 0.1266118288040161, "learning_rate": 2.255265165025663e-05, "loss": 0.0753, "step": 2185 }, { "epoch": 1.1923233973050225, "grad_norm": 0.12611445784568787, "learning_rate": 2.2429475071565987e-05, "loss": 0.0704, "step": 2190 }, { "epoch": 1.1950455968422484, "grad_norm": 0.1382066011428833, "learning_rate": 2.2306361529700125e-05, "loss": 0.081, "step": 2195 }, { "epoch": 1.1977677963794746, "grad_norm": 0.13451717793941498, "learning_rate": 2.218331404376484e-05, "loss": 0.0813, "step": 2200 }, { "epoch": 1.2004899959167008, "grad_norm": 0.1681748777627945, "learning_rate": 2.2060335631246075e-05, "loss": 0.0793, "step": 2205 }, { "epoch": 1.2032121954539268, "grad_norm": 0.14408931136131287, "learning_rate": 2.1937429307935887e-05, "loss": 0.0763, "step": 2210 }, { "epoch": 1.2059343949911527, "grad_norm": 0.14158952236175537, "learning_rate": 2.1814598087858476e-05, "loss": 0.0882, "step": 2215 }, { "epoch": 1.208656594528379, "grad_norm": 0.12508870661258698, "learning_rate": 2.16918449831963e-05, "loss": 0.0762, "step": 2220 }, { "epoch": 1.2113787940656051, "grad_norm": 0.13746081292629242, "learning_rate": 2.15691730042162e-05, "loss": 0.0802, "step": 2225 }, { "epoch": 1.214100993602831, "grad_norm": 0.12565705180168152, "learning_rate": 2.144658515919557e-05, "loss": 0.071, "step": 2230 }, { "epoch": 1.216823193140057, "grad_norm": 0.13257178664207458, "learning_rate": 2.1324084454348592e-05, "loss": 0.069, "step": 2235 }, { "epoch": 1.2195453926772832, "grad_norm": 0.15064120292663574, "learning_rate": 2.1201673893752534e-05, "loss": 0.0793, "step": 2240 }, { "epoch": 1.2222675922145094, "grad_norm": 0.1452379673719406, "learning_rate": 2.107935647927404e-05, "loss": 0.0653, "step": 2245 }, { "epoch": 1.2249897917517354, "grad_norm": 0.16073830425739288, "learning_rate": 2.095713521049554e-05, "loss": 0.0726, "step": 2250 }, { "epoch": 1.2277119912889614, "grad_norm": 0.15698719024658203, "learning_rate": 2.0835013084641704e-05, "loss": 0.0781, "step": 2255 }, { "epoch": 1.2304341908261875, "grad_norm": 0.13535834848880768, "learning_rate": 2.07129930965059e-05, "loss": 0.0754, "step": 2260 }, { "epoch": 1.2331563903634137, "grad_norm": 0.13440349698066711, "learning_rate": 2.0591078238376803e-05, "loss": 0.1275, "step": 2265 }, { "epoch": 1.2358785899006397, "grad_norm": 0.15323837101459503, "learning_rate": 2.0469271499964993e-05, "loss": 0.0738, "step": 2270 }, { "epoch": 1.238600789437866, "grad_norm": 0.14659421145915985, "learning_rate": 2.034757586832961e-05, "loss": 0.0719, "step": 2275 }, { "epoch": 1.2413229889750919, "grad_norm": 0.14592771232128143, "learning_rate": 2.022599432780515e-05, "loss": 0.0793, "step": 2280 }, { "epoch": 1.244045188512318, "grad_norm": 0.12547121942043304, "learning_rate": 2.0104529859928254e-05, "loss": 0.0754, "step": 2285 }, { "epoch": 1.246767388049544, "grad_norm": 0.13453277945518494, "learning_rate": 1.9983185443364615e-05, "loss": 0.0679, "step": 2290 }, { "epoch": 1.2494895875867702, "grad_norm": 0.13268929719924927, "learning_rate": 1.9861964053835885e-05, "loss": 0.0778, "step": 2295 }, { "epoch": 1.2522117871239962, "grad_norm": 0.14037257432937622, "learning_rate": 1.974086866404675e-05, "loss": 0.0874, "step": 2300 }, { "epoch": 1.2549339866612224, "grad_norm": 0.1446313112974167, "learning_rate": 1.961990224361201e-05, "loss": 0.0708, "step": 2305 }, { "epoch": 1.2576561861984483, "grad_norm": 0.15174619853496552, "learning_rate": 1.949906775898375e-05, "loss": 0.0841, "step": 2310 }, { "epoch": 1.2603783857356743, "grad_norm": 0.14297954738140106, "learning_rate": 1.9378368173378618e-05, "loss": 0.0802, "step": 2315 }, { "epoch": 1.2631005852729005, "grad_norm": 0.1344052255153656, "learning_rate": 1.9257806446705116e-05, "loss": 0.0675, "step": 2320 }, { "epoch": 1.2658227848101267, "grad_norm": 0.14959311485290527, "learning_rate": 1.913738553549106e-05, "loss": 0.0848, "step": 2325 }, { "epoch": 1.2685449843473526, "grad_norm": 0.12404653429985046, "learning_rate": 1.9017108392811065e-05, "loss": 0.0687, "step": 2330 }, { "epoch": 1.2712671838845788, "grad_norm": 0.13011127710342407, "learning_rate": 1.8896977968214078e-05, "loss": 0.113, "step": 2335 }, { "epoch": 1.2739893834218048, "grad_norm": 0.14289557933807373, "learning_rate": 1.877699720765114e-05, "loss": 0.0771, "step": 2340 }, { "epoch": 1.276711582959031, "grad_norm": 0.14011026918888092, "learning_rate": 1.8657169053403052e-05, "loss": 0.0771, "step": 2345 }, { "epoch": 1.279433782496257, "grad_norm": 0.1479315608739853, "learning_rate": 1.8537496444008283e-05, "loss": 0.07, "step": 2350 }, { "epoch": 1.2821559820334831, "grad_norm": 0.1444329172372818, "learning_rate": 1.841798231419087e-05, "loss": 0.0815, "step": 2355 }, { "epoch": 1.284878181570709, "grad_norm": 0.11986621469259262, "learning_rate": 1.8298629594788467e-05, "loss": 0.0689, "step": 2360 }, { "epoch": 1.2876003811079353, "grad_norm": 0.13101747632026672, "learning_rate": 1.817944121268048e-05, "loss": 0.0728, "step": 2365 }, { "epoch": 1.2903225806451613, "grad_norm": 0.1301647424697876, "learning_rate": 1.8060420090716265e-05, "loss": 0.0789, "step": 2370 }, { "epoch": 1.2930447801823874, "grad_norm": 0.14230471849441528, "learning_rate": 1.794156914764349e-05, "loss": 0.0702, "step": 2375 }, { "epoch": 1.2957669797196134, "grad_norm": 0.15737561881542206, "learning_rate": 1.7822891298036515e-05, "loss": 0.0781, "step": 2380 }, { "epoch": 1.2984891792568396, "grad_norm": 0.16721278429031372, "learning_rate": 1.7704389452224944e-05, "loss": 0.0805, "step": 2385 }, { "epoch": 1.3012113787940656, "grad_norm": 0.1321476846933365, "learning_rate": 1.7586066516222276e-05, "loss": 0.0768, "step": 2390 }, { "epoch": 1.3039335783312918, "grad_norm": 0.15134398639202118, "learning_rate": 1.7467925391654585e-05, "loss": 0.0825, "step": 2395 }, { "epoch": 1.3066557778685177, "grad_norm": 0.14100222289562225, "learning_rate": 1.73499689756894e-05, "loss": 0.0823, "step": 2400 }, { "epoch": 1.309377977405744, "grad_norm": 0.1558840423822403, "learning_rate": 1.7232200160964657e-05, "loss": 0.0832, "step": 2405 }, { "epoch": 1.3121001769429699, "grad_norm": 0.14366985857486725, "learning_rate": 1.7114621835517773e-05, "loss": 0.0791, "step": 2410 }, { "epoch": 1.314822376480196, "grad_norm": 0.12937745451927185, "learning_rate": 1.699723688271477e-05, "loss": 0.0703, "step": 2415 }, { "epoch": 1.317544576017422, "grad_norm": 0.12883788347244263, "learning_rate": 1.6880048181179652e-05, "loss": 0.0797, "step": 2420 }, { "epoch": 1.3202667755546482, "grad_norm": 0.13123267889022827, "learning_rate": 1.6763058604723723e-05, "loss": 0.0727, "step": 2425 }, { "epoch": 1.3229889750918742, "grad_norm": 0.10518030822277069, "learning_rate": 1.6646271022275185e-05, "loss": 0.066, "step": 2430 }, { "epoch": 1.3257111746291004, "grad_norm": 0.13036532700061798, "learning_rate": 1.6529688297808726e-05, "loss": 0.078, "step": 2435 }, { "epoch": 1.3284333741663263, "grad_norm": 0.14793965220451355, "learning_rate": 1.6413313290275355e-05, "loss": 0.0779, "step": 2440 }, { "epoch": 1.3311555737035525, "grad_norm": 0.15071183443069458, "learning_rate": 1.629714885353221e-05, "loss": 0.0835, "step": 2445 }, { "epoch": 1.3338777732407785, "grad_norm": 0.13086527585983276, "learning_rate": 1.618119783627263e-05, "loss": 0.0759, "step": 2450 }, { "epoch": 1.3365999727780047, "grad_norm": 0.13066627085208893, "learning_rate": 1.6065463081956292e-05, "loss": 0.0745, "step": 2455 }, { "epoch": 1.3393221723152307, "grad_norm": 0.15003521740436554, "learning_rate": 1.5949947428739448e-05, "loss": 0.084, "step": 2460 }, { "epoch": 1.3420443718524568, "grad_norm": 0.12341570854187012, "learning_rate": 1.5834653709405368e-05, "loss": 0.0707, "step": 2465 }, { "epoch": 1.3447665713896828, "grad_norm": 0.1332896649837494, "learning_rate": 1.571958475129484e-05, "loss": 0.0701, "step": 2470 }, { "epoch": 1.347488770926909, "grad_norm": 0.12761184573173523, "learning_rate": 1.5604743376236847e-05, "loss": 0.0708, "step": 2475 }, { "epoch": 1.350210970464135, "grad_norm": 0.12873777747154236, "learning_rate": 1.549013240047937e-05, "loss": 0.0731, "step": 2480 }, { "epoch": 1.3529331700013612, "grad_norm": 0.1278139352798462, "learning_rate": 1.537575463462031e-05, "loss": 0.0712, "step": 2485 }, { "epoch": 1.3556553695385871, "grad_norm": 0.16153199970722198, "learning_rate": 1.526161288353861e-05, "loss": 0.0825, "step": 2490 }, { "epoch": 1.3583775690758133, "grad_norm": 0.13839781284332275, "learning_rate": 1.5147709946325395e-05, "loss": 0.0769, "step": 2495 }, { "epoch": 1.3610997686130393, "grad_norm": 0.1404309719800949, "learning_rate": 1.5034048616215402e-05, "loss": 0.0737, "step": 2500 }, { "epoch": 1.3610997686130393, "eval_loss": 0.07860012352466583, "eval_runtime": 248.5723, "eval_samples_per_second": 2.414, "eval_steps_per_second": 0.603, "step": 2500 }, { "epoch": 1.3638219681502655, "grad_norm": 0.14883238077163696, "learning_rate": 1.4920631680518432e-05, "loss": 0.0695, "step": 2505 }, { "epoch": 1.3665441676874914, "grad_norm": 0.14661747217178345, "learning_rate": 1.4807461920551028e-05, "loss": 0.0752, "step": 2510 }, { "epoch": 1.3692663672247176, "grad_norm": 0.13054049015045166, "learning_rate": 1.469454211156826e-05, "loss": 0.0698, "step": 2515 }, { "epoch": 1.3719885667619436, "grad_norm": 0.13900673389434814, "learning_rate": 1.4581875022695653e-05, "loss": 0.0897, "step": 2520 }, { "epoch": 1.3747107662991698, "grad_norm": 0.1521766632795334, "learning_rate": 1.4469463416861307e-05, "loss": 0.0751, "step": 2525 }, { "epoch": 1.3774329658363957, "grad_norm": 0.14469105005264282, "learning_rate": 1.4357310050728115e-05, "loss": 0.0807, "step": 2530 }, { "epoch": 1.380155165373622, "grad_norm": 0.132577583193779, "learning_rate": 1.4245417674626183e-05, "loss": 0.0693, "step": 2535 }, { "epoch": 1.382877364910848, "grad_norm": 0.12606105208396912, "learning_rate": 1.4133789032485367e-05, "loss": 0.0739, "step": 2540 }, { "epoch": 1.385599564448074, "grad_norm": 0.1424499899148941, "learning_rate": 1.4022426861767998e-05, "loss": 0.0806, "step": 2545 }, { "epoch": 1.3883217639853, "grad_norm": 0.15580779314041138, "learning_rate": 1.391133389340174e-05, "loss": 0.0737, "step": 2550 }, { "epoch": 1.3910439635225262, "grad_norm": 0.14672575891017914, "learning_rate": 1.3800512851712635e-05, "loss": 0.0746, "step": 2555 }, { "epoch": 1.3937661630597522, "grad_norm": 0.15802086889743805, "learning_rate": 1.3689966454358256e-05, "loss": 0.0768, "step": 2560 }, { "epoch": 1.3964883625969784, "grad_norm": 0.1357397735118866, "learning_rate": 1.3579697412261117e-05, "loss": 0.0724, "step": 2565 }, { "epoch": 1.3992105621342044, "grad_norm": 0.14769281446933746, "learning_rate": 1.3469708429542157e-05, "loss": 0.0738, "step": 2570 }, { "epoch": 1.4019327616714305, "grad_norm": 0.1663895845413208, "learning_rate": 1.3360002203454442e-05, "loss": 0.0739, "step": 2575 }, { "epoch": 1.4046549612086565, "grad_norm": 0.13682472705841064, "learning_rate": 1.325058142431701e-05, "loss": 0.0732, "step": 2580 }, { "epoch": 1.4073771607458827, "grad_norm": 0.1464831680059433, "learning_rate": 1.3141448775448875e-05, "loss": 0.0781, "step": 2585 }, { "epoch": 1.4100993602831087, "grad_norm": 0.14444297552108765, "learning_rate": 1.3032606933103305e-05, "loss": 0.0742, "step": 2590 }, { "epoch": 1.4128215598203349, "grad_norm": 0.14417661726474762, "learning_rate": 1.2924058566402098e-05, "loss": 0.0761, "step": 2595 }, { "epoch": 1.4155437593575608, "grad_norm": 0.12622784078121185, "learning_rate": 1.2815806337270186e-05, "loss": 0.0693, "step": 2600 }, { "epoch": 1.418265958894787, "grad_norm": 0.1415053904056549, "learning_rate": 1.270785290037031e-05, "loss": 0.0771, "step": 2605 }, { "epoch": 1.4209881584320132, "grad_norm": 0.14631399512290955, "learning_rate": 1.260020090303797e-05, "loss": 0.1018, "step": 2610 }, { "epoch": 1.4237103579692392, "grad_norm": 0.1438084840774536, "learning_rate": 1.2492852985216483e-05, "loss": 0.0708, "step": 2615 }, { "epoch": 1.4264325575064651, "grad_norm": 0.12973198294639587, "learning_rate": 1.2385811779392236e-05, "loss": 0.0795, "step": 2620 }, { "epoch": 1.4291547570436913, "grad_norm": 0.15162228047847748, "learning_rate": 1.2279079910530147e-05, "loss": 0.0811, "step": 2625 }, { "epoch": 1.4318769565809175, "grad_norm": 0.14159157872200012, "learning_rate": 1.2172659996009254e-05, "loss": 0.0762, "step": 2630 }, { "epoch": 1.4345991561181435, "grad_norm": 0.13606242835521698, "learning_rate": 1.2066554645558578e-05, "loss": 0.0739, "step": 2635 }, { "epoch": 1.4373213556553694, "grad_norm": 0.15046873688697815, "learning_rate": 1.1960766461193124e-05, "loss": 0.0663, "step": 2640 }, { "epoch": 1.4400435551925956, "grad_norm": 0.12745556235313416, "learning_rate": 1.1855298037150022e-05, "loss": 0.073, "step": 2645 }, { "epoch": 1.4427657547298218, "grad_norm": 0.156600683927536, "learning_rate": 1.1750151959824961e-05, "loss": 0.0746, "step": 2650 }, { "epoch": 1.4454879542670478, "grad_norm": 0.13598720729351044, "learning_rate": 1.1645330807708713e-05, "loss": 0.0687, "step": 2655 }, { "epoch": 1.4482101538042738, "grad_norm": 0.13264238834381104, "learning_rate": 1.1540837151323951e-05, "loss": 0.0814, "step": 2660 }, { "epoch": 1.4509323533415, "grad_norm": 0.13126635551452637, "learning_rate": 1.143667355316219e-05, "loss": 0.0768, "step": 2665 }, { "epoch": 1.4536545528787261, "grad_norm": 0.14652326703071594, "learning_rate": 1.1332842567620941e-05, "loss": 0.0731, "step": 2670 }, { "epoch": 1.456376752415952, "grad_norm": 0.14913715422153473, "learning_rate": 1.1229346740941088e-05, "loss": 0.0893, "step": 2675 }, { "epoch": 1.459098951953178, "grad_norm": 0.15281103551387787, "learning_rate": 1.1126188611144406e-05, "loss": 0.0981, "step": 2680 }, { "epoch": 1.4618211514904043, "grad_norm": 0.14536678791046143, "learning_rate": 1.102337070797137e-05, "loss": 0.0775, "step": 2685 }, { "epoch": 1.4645433510276304, "grad_norm": 0.14888562262058258, "learning_rate": 1.0920895552819118e-05, "loss": 0.0828, "step": 2690 }, { "epoch": 1.4672655505648564, "grad_norm": 0.12814417481422424, "learning_rate": 1.0818765658679576e-05, "loss": 0.0766, "step": 2695 }, { "epoch": 1.4699877501020824, "grad_norm": 0.12727950513362885, "learning_rate": 1.0716983530077843e-05, "loss": 0.0776, "step": 2700 }, { "epoch": 1.4727099496393086, "grad_norm": 0.12727922201156616, "learning_rate": 1.0615551663010806e-05, "loss": 0.0689, "step": 2705 }, { "epoch": 1.4754321491765348, "grad_norm": 0.14102789759635925, "learning_rate": 1.051447254488591e-05, "loss": 0.0634, "step": 2710 }, { "epoch": 1.4781543487137607, "grad_norm": 0.14436770975589752, "learning_rate": 1.0413748654460149e-05, "loss": 0.0796, "step": 2715 }, { "epoch": 1.4808765482509867, "grad_norm": 0.1291683316230774, "learning_rate": 1.0313382461779306e-05, "loss": 0.0755, "step": 2720 }, { "epoch": 1.4835987477882129, "grad_norm": 0.12600617110729218, "learning_rate": 1.0213376428117333e-05, "loss": 0.0658, "step": 2725 }, { "epoch": 1.486320947325439, "grad_norm": 0.14813588559627533, "learning_rate": 1.0113733005916057e-05, "loss": 0.075, "step": 2730 }, { "epoch": 1.489043146862665, "grad_norm": 0.15321391820907593, "learning_rate": 1.0014454638724982e-05, "loss": 0.0774, "step": 2735 }, { "epoch": 1.491765346399891, "grad_norm": 0.13844044506549835, "learning_rate": 9.915543761141432e-06, "loss": 0.074, "step": 2740 }, { "epoch": 1.4944875459371172, "grad_norm": 0.14296895265579224, "learning_rate": 9.81700279875075e-06, "loss": 0.0796, "step": 2745 }, { "epoch": 1.4972097454743434, "grad_norm": 0.14961597323417664, "learning_rate": 9.718834168066904e-06, "loss": 0.0713, "step": 2750 }, { "epoch": 1.4999319450115693, "grad_norm": 0.1663280427455902, "learning_rate": 9.62104027647319e-06, "loss": 0.077, "step": 2755 }, { "epoch": 1.5026541445487953, "grad_norm": 0.14651760458946228, "learning_rate": 9.523623522163197e-06, "loss": 0.0719, "step": 2760 }, { "epoch": 1.5053763440860215, "grad_norm": 0.14970749616622925, "learning_rate": 9.426586294082013e-06, "loss": 0.0776, "step": 2765 }, { "epoch": 1.5080985436232477, "grad_norm": 0.15019264817237854, "learning_rate": 9.329930971867595e-06, "loss": 0.0771, "step": 2770 }, { "epoch": 1.5108207431604737, "grad_norm": 0.14440083503723145, "learning_rate": 9.233659925792477e-06, "loss": 0.072, "step": 2775 }, { "epoch": 1.5135429426976996, "grad_norm": 0.138129323720932, "learning_rate": 9.137775516705604e-06, "loss": 0.0758, "step": 2780 }, { "epoch": 1.5162651422349258, "grad_norm": 0.14640666544437408, "learning_rate": 9.042280095974434e-06, "loss": 0.0844, "step": 2785 }, { "epoch": 1.518987341772152, "grad_norm": 0.14141102135181427, "learning_rate": 8.947176005427324e-06, "loss": 0.0798, "step": 2790 }, { "epoch": 1.521709541309378, "grad_norm": 0.14003820717334747, "learning_rate": 8.852465577296015e-06, "loss": 0.0761, "step": 2795 }, { "epoch": 1.524431740846604, "grad_norm": 0.13643573224544525, "learning_rate": 8.75815113415852e-06, "loss": 0.0753, "step": 2800 }, { "epoch": 1.5271539403838301, "grad_norm": 0.13099689781665802, "learning_rate": 8.66423498888213e-06, "loss": 0.0764, "step": 2805 }, { "epoch": 1.5298761399210563, "grad_norm": 0.16020213067531586, "learning_rate": 8.570719444566702e-06, "loss": 0.0733, "step": 2810 }, { "epoch": 1.5325983394582823, "grad_norm": 0.1546473205089569, "learning_rate": 8.477606794488183e-06, "loss": 0.0839, "step": 2815 }, { "epoch": 1.5353205389955082, "grad_norm": 0.14650870859622955, "learning_rate": 8.384899322042356e-06, "loss": 0.0837, "step": 2820 }, { "epoch": 1.5380427385327344, "grad_norm": 0.11506952345371246, "learning_rate": 8.29259930068887e-06, "loss": 0.0775, "step": 2825 }, { "epoch": 1.5407649380699606, "grad_norm": 0.14615637063980103, "learning_rate": 8.200708993895476e-06, "loss": 0.0744, "step": 2830 }, { "epoch": 1.5434871376071866, "grad_norm": 0.13769914209842682, "learning_rate": 8.109230655082517e-06, "loss": 0.0795, "step": 2835 }, { "epoch": 1.5462093371444126, "grad_norm": 0.13748113811016083, "learning_rate": 8.018166527567672e-06, "loss": 0.0675, "step": 2840 }, { "epoch": 1.5489315366816387, "grad_norm": 0.15049228072166443, "learning_rate": 7.927518844510941e-06, "loss": 0.0703, "step": 2845 }, { "epoch": 1.551653736218865, "grad_norm": 0.13132914900779724, "learning_rate": 7.837289828859884e-06, "loss": 0.0742, "step": 2850 }, { "epoch": 1.554375935756091, "grad_norm": 0.14202824234962463, "learning_rate": 7.7474816932951e-06, "loss": 0.0699, "step": 2855 }, { "epoch": 1.5570981352933169, "grad_norm": 0.16140979528427124, "learning_rate": 7.658096640175985e-06, "loss": 0.0777, "step": 2860 }, { "epoch": 1.559820334830543, "grad_norm": 0.1300312578678131, "learning_rate": 7.569136861486686e-06, "loss": 0.0648, "step": 2865 }, { "epoch": 1.5625425343677692, "grad_norm": 0.14492963254451752, "learning_rate": 7.4806045387823935e-06, "loss": 0.0713, "step": 2870 }, { "epoch": 1.5652647339049952, "grad_norm": 0.13518249988555908, "learning_rate": 7.392501843135816e-06, "loss": 0.0649, "step": 2875 }, { "epoch": 1.5679869334422212, "grad_norm": 0.14563652873039246, "learning_rate": 7.304830935083934e-06, "loss": 0.0724, "step": 2880 }, { "epoch": 1.5707091329794474, "grad_norm": 0.1495407372713089, "learning_rate": 7.2175939645750454e-06, "loss": 0.0695, "step": 2885 }, { "epoch": 1.5734313325166736, "grad_norm": 0.15107296407222748, "learning_rate": 7.130793070916006e-06, "loss": 0.0762, "step": 2890 }, { "epoch": 1.5761535320538995, "grad_norm": 0.13882042467594147, "learning_rate": 7.0444303827197994e-06, "loss": 0.0698, "step": 2895 }, { "epoch": 1.5788757315911255, "grad_norm": 0.15497738122940063, "learning_rate": 6.9585080178533155e-06, "loss": 0.0729, "step": 2900 }, { "epoch": 1.5815979311283517, "grad_norm": 0.14168910682201385, "learning_rate": 6.873028083385435e-06, "loss": 0.0681, "step": 2905 }, { "epoch": 1.5843201306655779, "grad_norm": 0.1269499659538269, "learning_rate": 6.787992675535318e-06, "loss": 0.0803, "step": 2910 }, { "epoch": 1.5870423302028038, "grad_norm": 0.14480414986610413, "learning_rate": 6.703403879621048e-06, "loss": 0.0719, "step": 2915 }, { "epoch": 1.5897645297400298, "grad_norm": 0.14816336333751678, "learning_rate": 6.619263770008471e-06, "loss": 0.0764, "step": 2920 }, { "epoch": 1.592486729277256, "grad_norm": 0.13959960639476776, "learning_rate": 6.535574410060321e-06, "loss": 0.0777, "step": 2925 }, { "epoch": 1.5952089288144822, "grad_norm": 0.13289447128772736, "learning_rate": 6.452337852085621e-06, "loss": 0.077, "step": 2930 }, { "epoch": 1.5979311283517081, "grad_norm": 0.14249089360237122, "learning_rate": 6.369556137289373e-06, "loss": 0.0678, "step": 2935 }, { "epoch": 1.600653327888934, "grad_norm": 0.14189301431179047, "learning_rate": 6.28723129572247e-06, "loss": 0.077, "step": 2940 }, { "epoch": 1.6033755274261603, "grad_norm": 0.13121193647384644, "learning_rate": 6.2053653462319475e-06, "loss": 0.0651, "step": 2945 }, { "epoch": 1.6060977269633865, "grad_norm": 0.13624070584774017, "learning_rate": 6.123960296411449e-06, "loss": 0.0726, "step": 2950 }, { "epoch": 1.6088199265006125, "grad_norm": 0.1546383649110794, "learning_rate": 6.043018142552009e-06, "loss": 0.0747, "step": 2955 }, { "epoch": 1.6115421260378384, "grad_norm": 0.14421793818473816, "learning_rate": 5.962540869593081e-06, "loss": 0.0818, "step": 2960 }, { "epoch": 1.6142643255750646, "grad_norm": 0.13135267794132233, "learning_rate": 5.882530451073886e-06, "loss": 0.0724, "step": 2965 }, { "epoch": 1.6169865251122908, "grad_norm": 0.14589843153953552, "learning_rate": 5.8029888490850005e-06, "loss": 0.0769, "step": 2970 }, { "epoch": 1.6197087246495168, "grad_norm": 0.1423359513282776, "learning_rate": 5.723918014220236e-06, "loss": 0.0662, "step": 2975 }, { "epoch": 1.6224309241867427, "grad_norm": 0.12941160798072815, "learning_rate": 5.645319885528824e-06, "loss": 0.0938, "step": 2980 }, { "epoch": 1.625153123723969, "grad_norm": 0.16445757448673248, "learning_rate": 5.5671963904678185e-06, "loss": 0.0752, "step": 2985 }, { "epoch": 1.627875323261195, "grad_norm": 0.15879502892494202, "learning_rate": 5.489549444854908e-06, "loss": 0.0875, "step": 2990 }, { "epoch": 1.630597522798421, "grad_norm": 0.1418047547340393, "learning_rate": 5.412380952821358e-06, "loss": 0.0778, "step": 2995 }, { "epoch": 1.633319722335647, "grad_norm": 0.1362149566411972, "learning_rate": 5.33569280676536e-06, "loss": 0.0692, "step": 3000 }, { "epoch": 1.633319722335647, "eval_loss": 0.07731131464242935, "eval_runtime": 262.646, "eval_samples_per_second": 2.284, "eval_steps_per_second": 0.571, "step": 3000 }, { "epoch": 1.6360419218728732, "grad_norm": 0.14646309614181519, "learning_rate": 5.259486887305592e-06, "loss": 0.068, "step": 3005 }, { "epoch": 1.6387641214100994, "grad_norm": 0.13968642055988312, "learning_rate": 5.183765063235138e-06, "loss": 0.0727, "step": 3010 }, { "epoch": 1.6414863209473254, "grad_norm": 0.1375684291124344, "learning_rate": 5.108529191475636e-06, "loss": 0.0794, "step": 3015 }, { "epoch": 1.6442085204845516, "grad_norm": 0.1326666921377182, "learning_rate": 5.033781117031738e-06, "loss": 0.072, "step": 3020 }, { "epoch": 1.6469307200217775, "grad_norm": 0.14453214406967163, "learning_rate": 4.959522672945891e-06, "loss": 0.0763, "step": 3025 }, { "epoch": 1.6496529195590037, "grad_norm": 0.15132352709770203, "learning_rate": 4.885755680253334e-06, "loss": 0.1175, "step": 3030 }, { "epoch": 1.65237511909623, "grad_norm": 0.15359079837799072, "learning_rate": 4.812481947937497e-06, "loss": 0.0837, "step": 3035 }, { "epoch": 1.6550973186334559, "grad_norm": 0.13241231441497803, "learning_rate": 4.739703272885626e-06, "loss": 0.0738, "step": 3040 }, { "epoch": 1.6578195181706818, "grad_norm": 0.14873500168323517, "learning_rate": 4.667421439844691e-06, "loss": 0.0708, "step": 3045 }, { "epoch": 1.660541717707908, "grad_norm": 0.13607364892959595, "learning_rate": 4.5956382213776295e-06, "loss": 0.1151, "step": 3050 }, { "epoch": 1.6632639172451342, "grad_norm": 0.15463878214359283, "learning_rate": 4.524355377819897e-06, "loss": 0.0831, "step": 3055 }, { "epoch": 1.6659861167823602, "grad_norm": 0.1272389441728592, "learning_rate": 4.45357465723629e-06, "loss": 0.0714, "step": 3060 }, { "epoch": 1.6687083163195862, "grad_norm": 0.14218097925186157, "learning_rate": 4.383297795378061e-06, "loss": 0.0675, "step": 3065 }, { "epoch": 1.6714305158568123, "grad_norm": 0.1373002678155899, "learning_rate": 4.3135265156403756e-06, "loss": 0.0791, "step": 3070 }, { "epoch": 1.6741527153940385, "grad_norm": 0.16303496062755585, "learning_rate": 4.244262529020041e-06, "loss": 0.0757, "step": 3075 }, { "epoch": 1.6768749149312645, "grad_norm": 0.1444956660270691, "learning_rate": 4.175507534073533e-06, "loss": 0.0732, "step": 3080 }, { "epoch": 1.6795971144684905, "grad_norm": 0.1505775898694992, "learning_rate": 4.107263216875362e-06, "loss": 0.0751, "step": 3085 }, { "epoch": 1.6823193140057167, "grad_norm": 0.14044207334518433, "learning_rate": 4.039531250976735e-06, "loss": 0.069, "step": 3090 }, { "epoch": 1.6850415135429428, "grad_norm": 0.14999620616436005, "learning_rate": 3.972313297364494e-06, "loss": 0.0708, "step": 3095 }, { "epoch": 1.6877637130801688, "grad_norm": 0.1265767216682434, "learning_rate": 3.90561100442036e-06, "loss": 0.0655, "step": 3100 }, { "epoch": 1.6904859126173948, "grad_norm": 0.16221390664577484, "learning_rate": 3.8394260078805755e-06, "loss": 0.0736, "step": 3105 }, { "epoch": 1.693208112154621, "grad_norm": 0.15090176463127136, "learning_rate": 3.7737599307957365e-06, "loss": 0.0669, "step": 3110 }, { "epoch": 1.6959303116918472, "grad_norm": 0.14406718313694, "learning_rate": 3.708614383491016e-06, "loss": 0.0737, "step": 3115 }, { "epoch": 1.6986525112290731, "grad_norm": 0.12115464359521866, "learning_rate": 3.643990963526672e-06, "loss": 0.0627, "step": 3120 }, { "epoch": 1.701374710766299, "grad_norm": 0.16066431999206543, "learning_rate": 3.579891255658846e-06, "loss": 0.0757, "step": 3125 }, { "epoch": 1.7040969103035253, "grad_norm": 0.13997657597064972, "learning_rate": 3.5163168318007365e-06, "loss": 0.1398, "step": 3130 }, { "epoch": 1.7068191098407515, "grad_norm": 0.13934873044490814, "learning_rate": 3.4532692509840313e-06, "loss": 0.0723, "step": 3135 }, { "epoch": 1.7095413093779774, "grad_norm": 0.13022823631763458, "learning_rate": 3.390750059320688e-06, "loss": 0.0789, "step": 3140 }, { "epoch": 1.7122635089152034, "grad_norm": 0.13600829243659973, "learning_rate": 3.328760789964988e-06, "loss": 0.0728, "step": 3145 }, { "epoch": 1.7149857084524296, "grad_norm": 0.1362736076116562, "learning_rate": 3.2673029630759745e-06, "loss": 0.0694, "step": 3150 }, { "epoch": 1.7177079079896558, "grad_norm": 0.14168758690357208, "learning_rate": 3.2063780857801596e-06, "loss": 0.0726, "step": 3155 }, { "epoch": 1.7204301075268817, "grad_norm": 0.14045333862304688, "learning_rate": 3.145987652134563e-06, "loss": 0.0666, "step": 3160 }, { "epoch": 1.7231523070641077, "grad_norm": 0.15181277692317963, "learning_rate": 3.0861331430900807e-06, "loss": 0.0788, "step": 3165 }, { "epoch": 1.725874506601334, "grad_norm": 0.14058373868465424, "learning_rate": 3.0268160264551484e-06, "loss": 0.0701, "step": 3170 }, { "epoch": 1.72859670613856, "grad_norm": 0.1474706381559372, "learning_rate": 2.968037756859776e-06, "loss": 0.0745, "step": 3175 }, { "epoch": 1.731318905675786, "grad_norm": 0.13882222771644592, "learning_rate": 2.9097997757198516e-06, "loss": 0.0717, "step": 3180 }, { "epoch": 1.734041105213012, "grad_norm": 0.14754945039749146, "learning_rate": 2.8521035112018063e-06, "loss": 0.0644, "step": 3185 }, { "epoch": 1.7367633047502382, "grad_norm": 0.1717384308576584, "learning_rate": 2.7949503781875912e-06, "loss": 0.0768, "step": 3190 }, { "epoch": 1.7394855042874644, "grad_norm": 0.1543627679347992, "learning_rate": 2.7383417782399693e-06, "loss": 0.0709, "step": 3195 }, { "epoch": 1.7422077038246904, "grad_norm": 0.1619829684495926, "learning_rate": 2.6822790995681577e-06, "loss": 0.0814, "step": 3200 }, { "epoch": 1.7449299033619163, "grad_norm": 0.13856656849384308, "learning_rate": 2.62676371699378e-06, "loss": 0.0701, "step": 3205 }, { "epoch": 1.7476521028991425, "grad_norm": 0.14767934381961823, "learning_rate": 2.5717969919171553e-06, "loss": 0.076, "step": 3210 }, { "epoch": 1.7503743024363687, "grad_norm": 0.1731363832950592, "learning_rate": 2.5173802722838867e-06, "loss": 0.0711, "step": 3215 }, { "epoch": 1.7530965019735947, "grad_norm": 0.14000795781612396, "learning_rate": 2.4635148925518577e-06, "loss": 0.0675, "step": 3220 }, { "epoch": 1.7558187015108206, "grad_norm": 0.14177638292312622, "learning_rate": 2.4102021736584575e-06, "loss": 0.0659, "step": 3225 }, { "epoch": 1.7585409010480468, "grad_norm": 0.1450289785861969, "learning_rate": 2.3574434229882145e-06, "loss": 0.0836, "step": 3230 }, { "epoch": 1.761263100585273, "grad_norm": 0.15300573408603668, "learning_rate": 2.3052399343407277e-06, "loss": 0.079, "step": 3235 }, { "epoch": 1.763985300122499, "grad_norm": 0.12552043795585632, "learning_rate": 2.253592987898942e-06, "loss": 0.075, "step": 3240 }, { "epoch": 1.766707499659725, "grad_norm": 0.13973963260650635, "learning_rate": 2.2025038501977486e-06, "loss": 0.075, "step": 3245 }, { "epoch": 1.7694296991969511, "grad_norm": 0.14422014355659485, "learning_rate": 2.1519737740929274e-06, "loss": 0.0847, "step": 3250 }, { "epoch": 1.7721518987341773, "grad_norm": 0.15774239599704742, "learning_rate": 2.1020039987304285e-06, "loss": 0.0852, "step": 3255 }, { "epoch": 1.7748740982714033, "grad_norm": 0.14551898837089539, "learning_rate": 2.05259574951599e-06, "loss": 0.0726, "step": 3260 }, { "epoch": 1.7775962978086293, "grad_norm": 0.1303868442773819, "learning_rate": 2.003750238085053e-06, "loss": 0.076, "step": 3265 }, { "epoch": 1.7803184973458555, "grad_norm": 0.16554243862628937, "learning_rate": 1.9554686622730993e-06, "loss": 0.0992, "step": 3270 }, { "epoch": 1.7830406968830816, "grad_norm": 0.15280041098594666, "learning_rate": 1.907752206086247e-06, "loss": 0.0819, "step": 3275 }, { "epoch": 1.7857628964203076, "grad_norm": 0.14237718284130096, "learning_rate": 1.8606020396722129e-06, "loss": 0.0726, "step": 3280 }, { "epoch": 1.7884850959575336, "grad_norm": 0.15816614031791687, "learning_rate": 1.814019319291635e-06, "loss": 0.0754, "step": 3285 }, { "epoch": 1.7912072954947598, "grad_norm": 0.13710108399391174, "learning_rate": 1.7680051872896869e-06, "loss": 0.0658, "step": 3290 }, { "epoch": 1.793929495031986, "grad_norm": 0.15852783620357513, "learning_rate": 1.7225607720681131e-06, "loss": 0.0732, "step": 3295 }, { "epoch": 1.796651694569212, "grad_norm": 0.15130798518657684, "learning_rate": 1.6776871880575084e-06, "loss": 0.0833, "step": 3300 }, { "epoch": 1.7993738941064379, "grad_norm": 0.15700189769268036, "learning_rate": 1.6333855356900185e-06, "loss": 0.1652, "step": 3305 }, { "epoch": 1.802096093643664, "grad_norm": 0.14298690855503082, "learning_rate": 1.5896569013723329e-06, "loss": 0.0797, "step": 3310 }, { "epoch": 1.8048182931808903, "grad_norm": 0.14160804450511932, "learning_rate": 1.5465023574590676e-06, "loss": 0.0757, "step": 3315 }, { "epoch": 1.8075404927181162, "grad_norm": 0.15241742134094238, "learning_rate": 1.5039229622264479e-06, "loss": 0.0781, "step": 3320 }, { "epoch": 1.8102626922553422, "grad_norm": 0.15816597640514374, "learning_rate": 1.4619197598463641e-06, "loss": 0.0764, "step": 3325 }, { "epoch": 1.8129848917925684, "grad_norm": 0.15082836151123047, "learning_rate": 1.420493780360771e-06, "loss": 0.0708, "step": 3330 }, { "epoch": 1.8157070913297946, "grad_norm": 0.13604667782783508, "learning_rate": 1.3796460396564098e-06, "loss": 0.0711, "step": 3335 }, { "epoch": 1.8184292908670205, "grad_norm": 0.14523722231388092, "learning_rate": 1.3393775394399123e-06, "loss": 0.0797, "step": 3340 }, { "epoch": 1.8211514904042465, "grad_norm": 0.14779233932495117, "learning_rate": 1.2996892672132416e-06, "loss": 0.0845, "step": 3345 }, { "epoch": 1.8238736899414727, "grad_norm": 0.13566653430461884, "learning_rate": 1.2605821962494574e-06, "loss": 0.1118, "step": 3350 }, { "epoch": 1.8265958894786989, "grad_norm": 0.15038765966892242, "learning_rate": 1.2220572855688507e-06, "loss": 0.0782, "step": 3355 }, { "epoch": 1.8293180890159249, "grad_norm": 0.14022061228752136, "learning_rate": 1.1841154799154374e-06, "loss": 0.0737, "step": 3360 }, { "epoch": 1.8320402885531508, "grad_norm": 0.14298531413078308, "learning_rate": 1.1467577097337867e-06, "loss": 0.0939, "step": 3365 }, { "epoch": 1.834762488090377, "grad_norm": 0.15060581266880035, "learning_rate": 1.1099848911462014e-06, "loss": 0.0752, "step": 3370 }, { "epoch": 1.8374846876276032, "grad_norm": 0.1449151188135147, "learning_rate": 1.0737979259302478e-06, "loss": 0.0688, "step": 3375 }, { "epoch": 1.8402068871648292, "grad_norm": 0.13491316139698029, "learning_rate": 1.0381977014966543e-06, "loss": 0.0996, "step": 3380 }, { "epoch": 1.8429290867020551, "grad_norm": 0.128363698720932, "learning_rate": 1.0031850908675283e-06, "loss": 0.1217, "step": 3385 }, { "epoch": 1.8456512862392813, "grad_norm": 0.14816632866859436, "learning_rate": 9.68760952654968e-07, "loss": 0.0653, "step": 3390 }, { "epoch": 1.8483734857765075, "grad_norm": 0.16936026513576508, "learning_rate": 9.349261310400037e-07, "loss": 0.0735, "step": 3395 }, { "epoch": 1.8510956853137335, "grad_norm": 0.14492131769657135, "learning_rate": 9.016814557518849e-07, "loss": 0.0683, "step": 3400 }, { "epoch": 1.8538178848509594, "grad_norm": 0.14610859751701355, "learning_rate": 8.690277420477372e-07, "loss": 0.079, "step": 3405 }, { "epoch": 1.8565400843881856, "grad_norm": 0.1451197862625122, "learning_rate": 8.369657906925732e-07, "loss": 0.0753, "step": 3410 }, { "epoch": 1.8592622839254118, "grad_norm": 0.14033135771751404, "learning_rate": 8.054963879396554e-07, "loss": 0.0646, "step": 3415 }, { "epoch": 1.8619844834626378, "grad_norm": 0.13456685841083527, "learning_rate": 7.746203055112145e-07, "loss": 0.1086, "step": 3420 }, { "epoch": 1.8647066829998638, "grad_norm": 0.15288980305194855, "learning_rate": 7.443383005795224e-07, "loss": 0.0946, "step": 3425 }, { "epoch": 1.86742888253709, "grad_norm": 0.14935944974422455, "learning_rate": 7.146511157483216e-07, "loss": 0.0829, "step": 3430 }, { "epoch": 1.8701510820743161, "grad_norm": 0.14166168868541718, "learning_rate": 6.85559479034617e-07, "loss": 0.0679, "step": 3435 }, { "epoch": 1.872873281611542, "grad_norm": 0.14961621165275574, "learning_rate": 6.570641038508296e-07, "loss": 0.0762, "step": 3440 }, { "epoch": 1.875595481148768, "grad_norm": 0.1400919258594513, "learning_rate": 6.291656889873016e-07, "loss": 0.0803, "step": 3445 }, { "epoch": 1.8783176806859943, "grad_norm": 0.13169647753238678, "learning_rate": 6.018649185951325e-07, "loss": 0.0707, "step": 3450 }, { "epoch": 1.8810398802232204, "grad_norm": 0.13241390883922577, "learning_rate": 5.751624621694429e-07, "loss": 0.0636, "step": 3455 }, { "epoch": 1.8837620797604464, "grad_norm": 0.16256216168403625, "learning_rate": 5.490589745329261e-07, "loss": 0.0774, "step": 3460 }, { "epoch": 1.8864842792976724, "grad_norm": 0.15392127633094788, "learning_rate": 5.235550958198083e-07, "loss": 0.073, "step": 3465 }, { "epoch": 1.8892064788348986, "grad_norm": 0.14630888402462006, "learning_rate": 4.986514514601415e-07, "loss": 0.0742, "step": 3470 }, { "epoch": 1.8919286783721248, "grad_norm": 0.16385363042354584, "learning_rate": 4.74348652164458e-07, "loss": 0.083, "step": 3475 }, { "epoch": 1.8946508779093507, "grad_norm": 0.16106607019901276, "learning_rate": 4.5064729390881246e-07, "loss": 0.0755, "step": 3480 }, { "epoch": 1.8973730774465767, "grad_norm": 0.13656839728355408, "learning_rate": 4.275479579201602e-07, "loss": 0.0703, "step": 3485 }, { "epoch": 1.9000952769838029, "grad_norm": 0.12616074085235596, "learning_rate": 4.050512106620913e-07, "loss": 0.0692, "step": 3490 }, { "epoch": 1.902817476521029, "grad_norm": 0.15891359746456146, "learning_rate": 3.8315760382096057e-07, "loss": 0.0794, "step": 3495 }, { "epoch": 1.905539676058255, "grad_norm": 0.1297856718301773, "learning_rate": 3.6186767429234323e-07, "loss": 0.0671, "step": 3500 }, { "epoch": 1.905539676058255, "eval_loss": 0.07686587423086166, "eval_runtime": 218.8338, "eval_samples_per_second": 2.742, "eval_steps_per_second": 0.685, "step": 3500 } ], "logging_steps": 5, "max_steps": 3672, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.596605492461568e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }