{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9992217898832685, "eval_steps": 500, "global_step": 963, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0010376134889753567, "grad_norm": 23.969658904297845, "learning_rate": 1.0309278350515465e-07, "loss": 1.3725, "step": 1 }, { "epoch": 0.005188067444876783, "grad_norm": 22.449071975220065, "learning_rate": 5.154639175257732e-07, "loss": 1.3719, "step": 5 }, { "epoch": 0.010376134889753566, "grad_norm": 8.548339530816044, "learning_rate": 1.0309278350515464e-06, "loss": 1.2558, "step": 10 }, { "epoch": 0.01556420233463035, "grad_norm": 8.215121326664015, "learning_rate": 1.5463917525773197e-06, "loss": 1.081, "step": 15 }, { "epoch": 0.020752269779507133, "grad_norm": 3.0841636846785283, "learning_rate": 2.061855670103093e-06, "loss": 0.9504, "step": 20 }, { "epoch": 0.02594033722438392, "grad_norm": 2.3540131831672575, "learning_rate": 2.577319587628866e-06, "loss": 0.9092, "step": 25 }, { "epoch": 0.0311284046692607, "grad_norm": 2.217217553056043, "learning_rate": 3.0927835051546395e-06, "loss": 0.8692, "step": 30 }, { "epoch": 0.03631647211413749, "grad_norm": 2.2563290398206615, "learning_rate": 3.6082474226804126e-06, "loss": 0.8445, "step": 35 }, { "epoch": 0.041504539559014265, "grad_norm": 2.2874485501473907, "learning_rate": 4.123711340206186e-06, "loss": 0.8336, "step": 40 }, { "epoch": 0.04669260700389105, "grad_norm": 2.3330889165134967, "learning_rate": 4.639175257731959e-06, "loss": 0.8218, "step": 45 }, { "epoch": 0.05188067444876784, "grad_norm": 2.16633956379982, "learning_rate": 5.154639175257732e-06, "loss": 0.8255, "step": 50 }, { "epoch": 0.057068741893644616, "grad_norm": 2.355679025636223, "learning_rate": 5.670103092783505e-06, "loss": 0.7891, "step": 55 }, { "epoch": 0.0622568093385214, "grad_norm": 2.4839432117685036, "learning_rate": 6.185567010309279e-06, "loss": 0.7814, "step": 60 }, { "epoch": 0.06744487678339818, "grad_norm": 2.480013950899919, "learning_rate": 6.701030927835052e-06, "loss": 0.7674, "step": 65 }, { "epoch": 0.07263294422827497, "grad_norm": 2.3378408946103284, "learning_rate": 7.216494845360825e-06, "loss": 0.766, "step": 70 }, { "epoch": 0.07782101167315175, "grad_norm": 2.2751461973205482, "learning_rate": 7.731958762886599e-06, "loss": 0.7433, "step": 75 }, { "epoch": 0.08300907911802853, "grad_norm": 2.428511002623931, "learning_rate": 8.247422680412371e-06, "loss": 0.7414, "step": 80 }, { "epoch": 0.08819714656290532, "grad_norm": 2.4298836600324045, "learning_rate": 8.762886597938146e-06, "loss": 0.7358, "step": 85 }, { "epoch": 0.0933852140077821, "grad_norm": 2.41313947506571, "learning_rate": 9.278350515463918e-06, "loss": 0.7319, "step": 90 }, { "epoch": 0.09857328145265888, "grad_norm": 2.4451429150679274, "learning_rate": 9.793814432989691e-06, "loss": 0.7323, "step": 95 }, { "epoch": 0.10376134889753567, "grad_norm": 2.450193589248992, "learning_rate": 9.999703897419048e-06, "loss": 0.7231, "step": 100 }, { "epoch": 0.10894941634241245, "grad_norm": 2.271786014084883, "learning_rate": 9.997894508649995e-06, "loss": 0.7149, "step": 105 }, { "epoch": 0.11413748378728923, "grad_norm": 2.354564055245926, "learning_rate": 9.99444082710777e-06, "loss": 0.708, "step": 110 }, { "epoch": 0.11932555123216602, "grad_norm": 2.220428698962425, "learning_rate": 9.989343989043563e-06, "loss": 0.7216, "step": 115 }, { "epoch": 0.1245136186770428, "grad_norm": 2.3141712328751396, "learning_rate": 9.982605671302293e-06, "loss": 0.7091, "step": 120 }, { "epoch": 0.1297016861219196, "grad_norm": 2.100396054783955, "learning_rate": 9.97422809077092e-06, "loss": 0.7066, "step": 125 }, { "epoch": 0.13488975356679636, "grad_norm": 2.2484885982675413, "learning_rate": 9.9642140036491e-06, "loss": 0.7085, "step": 130 }, { "epoch": 0.14007782101167315, "grad_norm": 2.1795476193729413, "learning_rate": 9.9525667045424e-06, "loss": 0.6889, "step": 135 }, { "epoch": 0.14526588845654995, "grad_norm": 2.1757051871338593, "learning_rate": 9.93929002537839e-06, "loss": 0.6921, "step": 140 }, { "epoch": 0.1504539559014267, "grad_norm": 2.143005235580036, "learning_rate": 9.924388334145943e-06, "loss": 0.6907, "step": 145 }, { "epoch": 0.1556420233463035, "grad_norm": 2.1989760690420157, "learning_rate": 9.90786653345818e-06, "loss": 0.6912, "step": 150 }, { "epoch": 0.1608300907911803, "grad_norm": 2.004571277860471, "learning_rate": 9.889730058939529e-06, "loss": 0.6859, "step": 155 }, { "epoch": 0.16601815823605706, "grad_norm": 2.05691987455993, "learning_rate": 9.869984877437413e-06, "loss": 0.6894, "step": 160 }, { "epoch": 0.17120622568093385, "grad_norm": 2.230053895792029, "learning_rate": 9.848637485059183e-06, "loss": 0.6814, "step": 165 }, { "epoch": 0.17639429312581065, "grad_norm": 1.9493958638517837, "learning_rate": 9.82569490503491e-06, "loss": 0.6731, "step": 170 }, { "epoch": 0.1815823605706874, "grad_norm": 2.133120594361784, "learning_rate": 9.80116468540677e-06, "loss": 0.6594, "step": 175 }, { "epoch": 0.1867704280155642, "grad_norm": 2.01624934264464, "learning_rate": 9.775054896545755e-06, "loss": 0.6751, "step": 180 }, { "epoch": 0.191958495460441, "grad_norm": 2.1502691215852527, "learning_rate": 9.747374128496541e-06, "loss": 0.6789, "step": 185 }, { "epoch": 0.19714656290531776, "grad_norm": 2.0484102083185194, "learning_rate": 9.718131488151399e-06, "loss": 0.6676, "step": 190 }, { "epoch": 0.20233463035019456, "grad_norm": 2.0715841424222337, "learning_rate": 9.687336596254045e-06, "loss": 0.6616, "step": 195 }, { "epoch": 0.20752269779507135, "grad_norm": 2.012157328183036, "learning_rate": 9.654999584234444e-06, "loss": 0.652, "step": 200 }, { "epoch": 0.2127107652399481, "grad_norm": 2.0669739212271923, "learning_rate": 9.621131090875603e-06, "loss": 0.6426, "step": 205 }, { "epoch": 0.2178988326848249, "grad_norm": 2.0105636015375143, "learning_rate": 9.585742258813447e-06, "loss": 0.6445, "step": 210 }, { "epoch": 0.2230869001297017, "grad_norm": 2.1108266544110688, "learning_rate": 9.548844730870903e-06, "loss": 0.6438, "step": 215 }, { "epoch": 0.22827496757457846, "grad_norm": 2.072355913378756, "learning_rate": 9.51045064622747e-06, "loss": 0.6565, "step": 220 }, { "epoch": 0.23346303501945526, "grad_norm": 2.166007360772802, "learning_rate": 9.470572636425451e-06, "loss": 0.647, "step": 225 }, { "epoch": 0.23865110246433205, "grad_norm": 2.022875957881762, "learning_rate": 9.429223821214213e-06, "loss": 0.6325, "step": 230 }, { "epoch": 0.2438391699092088, "grad_norm": 2.006861087987301, "learning_rate": 9.386417804233836e-06, "loss": 0.6477, "step": 235 }, { "epoch": 0.2490272373540856, "grad_norm": 2.0140489204477645, "learning_rate": 9.34216866853954e-06, "loss": 0.6391, "step": 240 }, { "epoch": 0.25421530479896237, "grad_norm": 1.9489606047213677, "learning_rate": 9.296490971968416e-06, "loss": 0.6283, "step": 245 }, { "epoch": 0.2594033722438392, "grad_norm": 2.072486707132733, "learning_rate": 9.249399742349928e-06, "loss": 0.6377, "step": 250 }, { "epoch": 0.26459143968871596, "grad_norm": 1.9650189580925839, "learning_rate": 9.20091047256181e-06, "loss": 0.6261, "step": 255 }, { "epoch": 0.2697795071335927, "grad_norm": 1.9241991797476943, "learning_rate": 9.151039115432946e-06, "loss": 0.6184, "step": 260 }, { "epoch": 0.27496757457846954, "grad_norm": 1.9743470888532664, "learning_rate": 9.099802078494947e-06, "loss": 0.6142, "step": 265 }, { "epoch": 0.2801556420233463, "grad_norm": 2.160988187935936, "learning_rate": 9.047216218584105e-06, "loss": 0.6094, "step": 270 }, { "epoch": 0.2853437094682231, "grad_norm": 1.9697508480614465, "learning_rate": 8.993298836295556e-06, "loss": 0.6196, "step": 275 }, { "epoch": 0.2905317769130999, "grad_norm": 1.8771524751425768, "learning_rate": 8.93806767029143e-06, "loss": 0.6163, "step": 280 }, { "epoch": 0.29571984435797666, "grad_norm": 2.125863779805947, "learning_rate": 8.88154089146488e-06, "loss": 0.6167, "step": 285 }, { "epoch": 0.3009079118028534, "grad_norm": 2.1188493077731514, "learning_rate": 8.823737096961916e-06, "loss": 0.5992, "step": 290 }, { "epoch": 0.30609597924773024, "grad_norm": 2.1335267497592807, "learning_rate": 8.764675304062992e-06, "loss": 0.6071, "step": 295 }, { "epoch": 0.311284046692607, "grad_norm": 2.036189297244598, "learning_rate": 8.704374943926386e-06, "loss": 0.609, "step": 300 }, { "epoch": 0.3164721141374838, "grad_norm": 1.915927299304865, "learning_rate": 8.642855855195394e-06, "loss": 0.5945, "step": 305 }, { "epoch": 0.3216601815823606, "grad_norm": 2.005194485630929, "learning_rate": 8.580138277471476e-06, "loss": 0.5959, "step": 310 }, { "epoch": 0.32684824902723736, "grad_norm": 2.1368034472887527, "learning_rate": 8.516242844655498e-06, "loss": 0.5941, "step": 315 }, { "epoch": 0.3320363164721141, "grad_norm": 1.9360804934529585, "learning_rate": 8.45119057815922e-06, "loss": 0.5915, "step": 320 }, { "epoch": 0.33722438391699094, "grad_norm": 1.9356101875463727, "learning_rate": 8.385002879989328e-06, "loss": 0.5838, "step": 325 }, { "epoch": 0.3424124513618677, "grad_norm": 2.4311425501079023, "learning_rate": 8.317701525706226e-06, "loss": 0.5946, "step": 330 }, { "epoch": 0.3476005188067445, "grad_norm": 2.356263841306792, "learning_rate": 8.249308657259943e-06, "loss": 0.567, "step": 335 }, { "epoch": 0.3527885862516213, "grad_norm": 2.048334150791661, "learning_rate": 8.179846775705504e-06, "loss": 0.5795, "step": 340 }, { "epoch": 0.35797665369649806, "grad_norm": 1.9977511587812506, "learning_rate": 8.109338733800132e-06, "loss": 0.5751, "step": 345 }, { "epoch": 0.3631647211413748, "grad_norm": 1.8688618314869894, "learning_rate": 8.03780772848477e-06, "loss": 0.568, "step": 350 }, { "epoch": 0.36835278858625164, "grad_norm": 1.93022130905715, "learning_rate": 7.965277293252354e-06, "loss": 0.5682, "step": 355 }, { "epoch": 0.3735408560311284, "grad_norm": 2.0382225242835528, "learning_rate": 7.891771290405351e-06, "loss": 0.5617, "step": 360 }, { "epoch": 0.3787289234760052, "grad_norm": 1.9924209327442368, "learning_rate": 7.817313903205148e-06, "loss": 0.5577, "step": 365 }, { "epoch": 0.383916990920882, "grad_norm": 1.9678458173326334, "learning_rate": 7.741929627915814e-06, "loss": 0.56, "step": 370 }, { "epoch": 0.38910505836575876, "grad_norm": 2.2405618654805215, "learning_rate": 7.66564326574491e-06, "loss": 0.5513, "step": 375 }, { "epoch": 0.3942931258106355, "grad_norm": 1.9971872990885233, "learning_rate": 7.588479914683954e-06, "loss": 0.5445, "step": 380 }, { "epoch": 0.39948119325551235, "grad_norm": 2.06807252227761, "learning_rate": 7.510464961251271e-06, "loss": 0.5674, "step": 385 }, { "epoch": 0.4046692607003891, "grad_norm": 1.9627368535332135, "learning_rate": 7.431624072139884e-06, "loss": 0.5435, "step": 390 }, { "epoch": 0.4098573281452659, "grad_norm": 1.9716804464407136, "learning_rate": 7.351983185773259e-06, "loss": 0.5552, "step": 395 }, { "epoch": 0.4150453955901427, "grad_norm": 1.9693396583392846, "learning_rate": 7.271568503771632e-06, "loss": 0.5343, "step": 400 }, { "epoch": 0.42023346303501946, "grad_norm": 1.9432949161104107, "learning_rate": 7.190406482331757e-06, "loss": 0.5475, "step": 405 }, { "epoch": 0.4254215304798962, "grad_norm": 2.0194917717314045, "learning_rate": 7.108523823522891e-06, "loss": 0.5477, "step": 410 }, { "epoch": 0.43060959792477305, "grad_norm": 2.206404974952941, "learning_rate": 7.0259474665018915e-06, "loss": 0.5425, "step": 415 }, { "epoch": 0.4357976653696498, "grad_norm": 1.9526533277899327, "learning_rate": 6.942704578650312e-06, "loss": 0.5161, "step": 420 }, { "epoch": 0.4409857328145266, "grad_norm": 2.0097466124913117, "learning_rate": 6.858822546636417e-06, "loss": 0.5331, "step": 425 }, { "epoch": 0.4461738002594034, "grad_norm": 1.8348649689633039, "learning_rate": 6.774328967405035e-06, "loss": 0.523, "step": 430 }, { "epoch": 0.45136186770428016, "grad_norm": 2.139084532722164, "learning_rate": 6.689251639098261e-06, "loss": 0.5251, "step": 435 }, { "epoch": 0.4565499351491569, "grad_norm": 1.9708479629081865, "learning_rate": 6.603618551909935e-06, "loss": 0.5232, "step": 440 }, { "epoch": 0.46173800259403375, "grad_norm": 1.9331722289768318, "learning_rate": 6.517457878876958e-06, "loss": 0.5305, "step": 445 }, { "epoch": 0.4669260700389105, "grad_norm": 1.859009250403284, "learning_rate": 6.430797966610436e-06, "loss": 0.5159, "step": 450 }, { "epoch": 0.4721141374837873, "grad_norm": 1.986527066309499, "learning_rate": 6.343667325969736e-06, "loss": 0.5367, "step": 455 }, { "epoch": 0.4773022049286641, "grad_norm": 1.9771277544299588, "learning_rate": 6.256094622682493e-06, "loss": 0.5123, "step": 460 }, { "epoch": 0.48249027237354086, "grad_norm": 2.0022259730400904, "learning_rate": 6.168108667913666e-06, "loss": 0.5166, "step": 465 }, { "epoch": 0.4876783398184176, "grad_norm": 1.9991961519932744, "learning_rate": 6.079738408786753e-06, "loss": 0.5161, "step": 470 }, { "epoch": 0.49286640726329445, "grad_norm": 2.0805595238898307, "learning_rate": 5.9910129188602665e-06, "loss": 0.5179, "step": 475 }, { "epoch": 0.4980544747081712, "grad_norm": 1.929253006230254, "learning_rate": 5.9019613885626235e-06, "loss": 0.5097, "step": 480 }, { "epoch": 0.503242542153048, "grad_norm": 2.25129632838715, "learning_rate": 5.812613115588575e-06, "loss": 0.4971, "step": 485 }, { "epoch": 0.5084306095979247, "grad_norm": 1.9119339241166262, "learning_rate": 5.722997495260348e-06, "loss": 0.4988, "step": 490 }, { "epoch": 0.5136186770428015, "grad_norm": 1.8300200112998326, "learning_rate": 5.6331440108566735e-06, "loss": 0.4941, "step": 495 }, { "epoch": 0.5188067444876784, "grad_norm": 1.9591247994452368, "learning_rate": 5.543082223912875e-06, "loss": 0.492, "step": 500 }, { "epoch": 0.5239948119325551, "grad_norm": 1.99136453982626, "learning_rate": 5.452841764495203e-06, "loss": 0.5002, "step": 505 }, { "epoch": 0.5291828793774319, "grad_norm": 1.9961024804052654, "learning_rate": 5.362452321452636e-06, "loss": 0.4772, "step": 510 }, { "epoch": 0.5343709468223087, "grad_norm": 1.9607124098040063, "learning_rate": 5.2719436326493255e-06, "loss": 0.4908, "step": 515 }, { "epoch": 0.5395590142671854, "grad_norm": 1.9303906010446525, "learning_rate": 5.181345475180941e-06, "loss": 0.4866, "step": 520 }, { "epoch": 0.5447470817120622, "grad_norm": 2.0420688559734503, "learning_rate": 5.090687655578078e-06, "loss": 0.4769, "step": 525 }, { "epoch": 0.5499351491569391, "grad_norm": 1.9908642175713687, "learning_rate": 5e-06, "loss": 0.4742, "step": 530 }, { "epoch": 0.5551232166018158, "grad_norm": 1.9960779934532675, "learning_rate": 4.909312344421923e-06, "loss": 0.4666, "step": 535 }, { "epoch": 0.5603112840466926, "grad_norm": 1.9274839933909422, "learning_rate": 4.8186545248190604e-06, "loss": 0.4866, "step": 540 }, { "epoch": 0.5654993514915694, "grad_norm": 1.9162466337096817, "learning_rate": 4.7280563673506745e-06, "loss": 0.4692, "step": 545 }, { "epoch": 0.5706874189364461, "grad_norm": 2.07386431606307, "learning_rate": 4.637547678547366e-06, "loss": 0.4859, "step": 550 }, { "epoch": 0.5758754863813229, "grad_norm": 2.0201984812958385, "learning_rate": 4.547158235504797e-06, "loss": 0.4718, "step": 555 }, { "epoch": 0.5810635538261998, "grad_norm": 1.95015272613481, "learning_rate": 4.4569177760871255e-06, "loss": 0.475, "step": 560 }, { "epoch": 0.5862516212710766, "grad_norm": 1.944586565605588, "learning_rate": 4.366855989143326e-06, "loss": 0.4551, "step": 565 }, { "epoch": 0.5914396887159533, "grad_norm": 1.9208589567145171, "learning_rate": 4.277002504739653e-06, "loss": 0.4686, "step": 570 }, { "epoch": 0.5966277561608301, "grad_norm": 1.8639671285460482, "learning_rate": 4.187386884411426e-06, "loss": 0.4557, "step": 575 }, { "epoch": 0.6018158236057068, "grad_norm": 1.9975578797091653, "learning_rate": 4.098038611437377e-06, "loss": 0.4651, "step": 580 }, { "epoch": 0.6070038910505836, "grad_norm": 1.961651938542185, "learning_rate": 4.008987081139734e-06, "loss": 0.4643, "step": 585 }, { "epoch": 0.6121919584954605, "grad_norm": 1.9374158302120401, "learning_rate": 3.920261591213249e-06, "loss": 0.4556, "step": 590 }, { "epoch": 0.6173800259403373, "grad_norm": 1.9090835435895448, "learning_rate": 3.8318913320863355e-06, "loss": 0.4536, "step": 595 }, { "epoch": 0.622568093385214, "grad_norm": 1.8975263865890188, "learning_rate": 3.7439053773175092e-06, "loss": 0.4615, "step": 600 }, { "epoch": 0.6277561608300908, "grad_norm": 1.9060390294655216, "learning_rate": 3.6563326740302664e-06, "loss": 0.4459, "step": 605 }, { "epoch": 0.6329442282749675, "grad_norm": 1.9725006931962796, "learning_rate": 3.569202033389565e-06, "loss": 0.4451, "step": 610 }, { "epoch": 0.6381322957198443, "grad_norm": 1.9621067476956515, "learning_rate": 3.4825421211230437e-06, "loss": 0.4419, "step": 615 }, { "epoch": 0.6433203631647212, "grad_norm": 2.098443239659209, "learning_rate": 3.3963814480900665e-06, "loss": 0.4415, "step": 620 }, { "epoch": 0.648508430609598, "grad_norm": 1.8981208726840302, "learning_rate": 3.310748360901741e-06, "loss": 0.4456, "step": 625 }, { "epoch": 0.6536964980544747, "grad_norm": 1.8947168989269416, "learning_rate": 3.225671032594966e-06, "loss": 0.4229, "step": 630 }, { "epoch": 0.6588845654993515, "grad_norm": 2.0138652650509288, "learning_rate": 3.1411774533635854e-06, "loss": 0.437, "step": 635 }, { "epoch": 0.6640726329442282, "grad_norm": 1.8903378440015823, "learning_rate": 3.0572954213496897e-06, "loss": 0.4454, "step": 640 }, { "epoch": 0.669260700389105, "grad_norm": 1.8448484960177367, "learning_rate": 2.9740525334981105e-06, "loss": 0.4398, "step": 645 }, { "epoch": 0.6744487678339819, "grad_norm": 1.9976530631786225, "learning_rate": 2.8914761764771093e-06, "loss": 0.429, "step": 650 }, { "epoch": 0.6796368352788587, "grad_norm": 1.9155018572353837, "learning_rate": 2.809593517668243e-06, "loss": 0.4309, "step": 655 }, { "epoch": 0.6848249027237354, "grad_norm": 1.942714148946629, "learning_rate": 2.728431496228369e-06, "loss": 0.4248, "step": 660 }, { "epoch": 0.6900129701686122, "grad_norm": 2.013023734418392, "learning_rate": 2.648016814226742e-06, "loss": 0.4304, "step": 665 }, { "epoch": 0.695201037613489, "grad_norm": 1.9023117871214554, "learning_rate": 2.5683759278601174e-06, "loss": 0.4338, "step": 670 }, { "epoch": 0.7003891050583657, "grad_norm": 1.8911448184302957, "learning_rate": 2.4895350387487304e-06, "loss": 0.4245, "step": 675 }, { "epoch": 0.7055771725032426, "grad_norm": 2.0358392917626813, "learning_rate": 2.4115200853160475e-06, "loss": 0.4194, "step": 680 }, { "epoch": 0.7107652399481194, "grad_norm": 1.9510576677492195, "learning_rate": 2.3343567342550933e-06, "loss": 0.4267, "step": 685 }, { "epoch": 0.7159533073929961, "grad_norm": 1.8690267408594539, "learning_rate": 2.258070372084188e-06, "loss": 0.4312, "step": 690 }, { "epoch": 0.7211413748378729, "grad_norm": 1.8322122073891454, "learning_rate": 2.182686096794852e-06, "loss": 0.4207, "step": 695 }, { "epoch": 0.7263294422827496, "grad_norm": 2.0311002524177253, "learning_rate": 2.108228709594649e-06, "loss": 0.4227, "step": 700 }, { "epoch": 0.7315175097276264, "grad_norm": 1.8678394687630775, "learning_rate": 2.0347227067476478e-06, "loss": 0.4149, "step": 705 }, { "epoch": 0.7367055771725033, "grad_norm": 1.8521301731665931, "learning_rate": 1.962192271515232e-06, "loss": 0.4192, "step": 710 }, { "epoch": 0.74189364461738, "grad_norm": 1.9291143236144128, "learning_rate": 1.8906612661998698e-06, "loss": 0.4128, "step": 715 }, { "epoch": 0.7470817120622568, "grad_norm": 1.991134829662921, "learning_rate": 1.820153224294498e-06, "loss": 0.4102, "step": 720 }, { "epoch": 0.7522697795071336, "grad_norm": 1.8597303553848081, "learning_rate": 1.750691342740058e-06, "loss": 0.4104, "step": 725 }, { "epoch": 0.7574578469520103, "grad_norm": 1.8334844899907363, "learning_rate": 1.6822984742937764e-06, "loss": 0.4049, "step": 730 }, { "epoch": 0.7626459143968871, "grad_norm": 1.839241095874111, "learning_rate": 1.6149971200106723e-06, "loss": 0.4009, "step": 735 }, { "epoch": 0.767833981841764, "grad_norm": 1.8614288971061537, "learning_rate": 1.548809421840779e-06, "loss": 0.4029, "step": 740 }, { "epoch": 0.7730220492866408, "grad_norm": 1.9712640153496117, "learning_rate": 1.483757155344503e-06, "loss": 0.4056, "step": 745 }, { "epoch": 0.7782101167315175, "grad_norm": 1.7973975593361922, "learning_rate": 1.4198617225285244e-06, "loss": 0.409, "step": 750 }, { "epoch": 0.7833981841763943, "grad_norm": 1.8870139707940816, "learning_rate": 1.3571441448046086e-06, "loss": 0.4117, "step": 755 }, { "epoch": 0.788586251621271, "grad_norm": 1.9255920717839368, "learning_rate": 1.2956250560736143e-06, "loss": 0.4097, "step": 760 }, { "epoch": 0.7937743190661478, "grad_norm": 1.8605570734597534, "learning_rate": 1.2353246959370086e-06, "loss": 0.3885, "step": 765 }, { "epoch": 0.7989623865110247, "grad_norm": 1.8678208345700735, "learning_rate": 1.1762629030380867e-06, "loss": 0.4044, "step": 770 }, { "epoch": 0.8041504539559015, "grad_norm": 1.8406724398818959, "learning_rate": 1.118459108535122e-06, "loss": 0.3991, "step": 775 }, { "epoch": 0.8093385214007782, "grad_norm": 1.9447178497450672, "learning_rate": 1.061932329708572e-06, "loss": 0.3878, "step": 780 }, { "epoch": 0.814526588845655, "grad_norm": 1.903347183666585, "learning_rate": 1.006701163704445e-06, "loss": 0.3994, "step": 785 }, { "epoch": 0.8197146562905318, "grad_norm": 1.7817648380438804, "learning_rate": 9.527837814158963e-07, "loss": 0.3943, "step": 790 }, { "epoch": 0.8249027237354085, "grad_norm": 1.89718875917406, "learning_rate": 9.001979215050544e-07, "loss": 0.3929, "step": 795 }, { "epoch": 0.8300907911802854, "grad_norm": 1.8825895138353903, "learning_rate": 8.489608845670527e-07, "loss": 0.3924, "step": 800 }, { "epoch": 0.8352788586251622, "grad_norm": 1.7908515887362904, "learning_rate": 7.99089527438191e-07, "loss": 0.3919, "step": 805 }, { "epoch": 0.8404669260700389, "grad_norm": 2.0314129121613034, "learning_rate": 7.506002576500732e-07, "loss": 0.3941, "step": 810 }, { "epoch": 0.8456549935149157, "grad_norm": 1.838737045068825, "learning_rate": 7.035090280315854e-07, "loss": 0.398, "step": 815 }, { "epoch": 0.8508430609597925, "grad_norm": 1.8701463175206698, "learning_rate": 6.578313314604612e-07, "loss": 0.395, "step": 820 }, { "epoch": 0.8560311284046692, "grad_norm": 1.8734176088672492, "learning_rate": 6.135821957661658e-07, "loss": 0.3945, "step": 825 }, { "epoch": 0.8612191958495461, "grad_norm": 1.8454123160341045, "learning_rate": 5.707761787857879e-07, "loss": 0.3855, "step": 830 }, { "epoch": 0.8664072632944229, "grad_norm": 1.7755466173110739, "learning_rate": 5.294273635745517e-07, "loss": 0.3971, "step": 835 }, { "epoch": 0.8715953307392996, "grad_norm": 1.8394864397787671, "learning_rate": 4.895493537725326e-07, "loss": 0.3966, "step": 840 }, { "epoch": 0.8767833981841764, "grad_norm": 1.7915928948304078, "learning_rate": 4.511552691290988e-07, "loss": 0.3979, "step": 845 }, { "epoch": 0.8819714656290532, "grad_norm": 1.7863787006400424, "learning_rate": 4.1425774118655505e-07, "loss": 0.3826, "step": 850 }, { "epoch": 0.8871595330739299, "grad_norm": 1.8024141112662704, "learning_rate": 3.7886890912439633e-07, "loss": 0.3862, "step": 855 }, { "epoch": 0.8923476005188068, "grad_norm": 1.8180972720099156, "learning_rate": 3.4500041576555733e-07, "loss": 0.3859, "step": 860 }, { "epoch": 0.8975356679636836, "grad_norm": 1.7622051589037506, "learning_rate": 3.1266340374595693e-07, "loss": 0.3831, "step": 865 }, { "epoch": 0.9027237354085603, "grad_norm": 1.7926800043760007, "learning_rate": 2.818685118486025e-07, "loss": 0.3927, "step": 870 }, { "epoch": 0.9079118028534371, "grad_norm": 1.8515815235983688, "learning_rate": 2.526258715034602e-07, "loss": 0.3787, "step": 875 }, { "epoch": 0.9130998702983139, "grad_norm": 1.761786866550431, "learning_rate": 2.2494510345424657e-07, "loss": 0.3881, "step": 880 }, { "epoch": 0.9182879377431906, "grad_norm": 1.731506703869926, "learning_rate": 1.988353145932298e-07, "loss": 0.3762, "step": 885 }, { "epoch": 0.9234760051880675, "grad_norm": 1.8427166106595052, "learning_rate": 1.7430509496508985e-07, "loss": 0.3975, "step": 890 }, { "epoch": 0.9286640726329443, "grad_norm": 1.761769698023775, "learning_rate": 1.5136251494081822e-07, "loss": 0.3842, "step": 895 }, { "epoch": 0.933852140077821, "grad_norm": 1.8297504100937483, "learning_rate": 1.3001512256258841e-07, "loss": 0.3916, "step": 900 }, { "epoch": 0.9390402075226978, "grad_norm": 1.8143369848190358, "learning_rate": 1.1026994106047296e-07, "loss": 0.3911, "step": 905 }, { "epoch": 0.9442282749675746, "grad_norm": 1.7462314691918333, "learning_rate": 9.213346654182054e-08, "loss": 0.3888, "step": 910 }, { "epoch": 0.9494163424124513, "grad_norm": 1.842285372864709, "learning_rate": 7.561166585405789e-08, "loss": 0.3823, "step": 915 }, { "epoch": 0.9546044098573282, "grad_norm": 1.798454935332072, "learning_rate": 6.070997462161055e-08, "loss": 0.4032, "step": 920 }, { "epoch": 0.959792477302205, "grad_norm": 1.8579672164577692, "learning_rate": 4.743329545760122e-08, "loss": 0.3811, "step": 925 }, { "epoch": 0.9649805447470817, "grad_norm": 1.764976690651984, "learning_rate": 3.578599635090163e-08, "loss": 0.3806, "step": 930 }, { "epoch": 0.9701686121919585, "grad_norm": 1.7085373084916373, "learning_rate": 2.577190922908035e-08, "loss": 0.3888, "step": 935 }, { "epoch": 0.9753566796368353, "grad_norm": 1.7431684765639506, "learning_rate": 1.7394328697707407e-08, "loss": 0.3901, "step": 940 }, { "epoch": 0.980544747081712, "grad_norm": 1.8495600056895127, "learning_rate": 1.0656010956437979e-08, "loss": 0.3918, "step": 945 }, { "epoch": 0.9857328145265889, "grad_norm": 1.8616847493274582, "learning_rate": 5.5591728922316235e-09, "loss": 0.3895, "step": 950 }, { "epoch": 0.9909208819714657, "grad_norm": 1.8274058784400706, "learning_rate": 2.1054913500051512e-09, "loss": 0.3831, "step": 955 }, { "epoch": 0.9961089494163424, "grad_norm": 1.7888916632814764, "learning_rate": 2.9610258095169596e-10, "loss": 0.3863, "step": 960 }, { "epoch": 0.9992217898832685, "eval_loss": 0.35284245014190674, "eval_runtime": 0.9437, "eval_samples_per_second": 2.119, "eval_steps_per_second": 1.06, "step": 963 }, { "epoch": 0.9992217898832685, "step": 963, "total_flos": 201580263505920.0, "train_loss": 0.5411187405403034, "train_runtime": 23935.6127, "train_samples_per_second": 1.288, "train_steps_per_second": 0.04 } ], "logging_steps": 5, "max_steps": 963, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 201580263505920.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }