{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.10481081647626035, "eval_steps": 42, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002096216329525207, "eval_loss": 1.4208108186721802, "eval_runtime": 163.5996, "eval_samples_per_second": 49.114, "eval_steps_per_second": 6.143, "step": 1 }, { "epoch": 0.0006288648988575621, "grad_norm": 0.39683395624160767, "learning_rate": 3e-05, "loss": 1.4354, "step": 3 }, { "epoch": 0.0012577297977151242, "grad_norm": 0.4367702305316925, "learning_rate": 6e-05, "loss": 1.4342, "step": 6 }, { "epoch": 0.0018865946965726864, "grad_norm": 0.37110769748687744, "learning_rate": 9e-05, "loss": 1.3479, "step": 9 }, { "epoch": 0.0025154595954302483, "grad_norm": 0.36730247735977173, "learning_rate": 9.999588943391597e-05, "loss": 1.361, "step": 12 }, { "epoch": 0.0031443244942878103, "grad_norm": 0.35089993476867676, "learning_rate": 9.99743108100344e-05, "loss": 1.2814, "step": 15 }, { "epoch": 0.0037731893931453727, "grad_norm": 0.38672298192977905, "learning_rate": 9.993424445916923e-05, "loss": 1.3016, "step": 18 }, { "epoch": 0.004402054292002935, "grad_norm": 0.321205198764801, "learning_rate": 9.987570520365104e-05, "loss": 1.2163, "step": 21 }, { "epoch": 0.005030919190860497, "grad_norm": 0.3282659649848938, "learning_rate": 9.979871469976196e-05, "loss": 1.2312, "step": 24 }, { "epoch": 0.005659784089718059, "grad_norm": 0.3265891373157501, "learning_rate": 9.970330142972401e-05, "loss": 1.2505, "step": 27 }, { "epoch": 0.006288648988575621, "grad_norm": 0.3450949192047119, "learning_rate": 9.95895006911623e-05, "loss": 1.1982, "step": 30 }, { "epoch": 0.0069175138874331835, "grad_norm": 0.3153296411037445, "learning_rate": 9.945735458404681e-05, "loss": 1.1669, "step": 33 }, { "epoch": 0.0075463787862907455, "grad_norm": 0.3419332802295685, "learning_rate": 9.930691199511775e-05, "loss": 1.2492, "step": 36 }, { "epoch": 0.008175243685148307, "grad_norm": 0.2867058515548706, "learning_rate": 9.91382285798002e-05, "loss": 1.1652, "step": 39 }, { "epoch": 0.00880410858400587, "grad_norm": 0.3220820426940918, "learning_rate": 9.895136674161465e-05, "loss": 1.1996, "step": 42 }, { "epoch": 0.00880410858400587, "eval_loss": 1.1905547380447388, "eval_runtime": 165.0006, "eval_samples_per_second": 48.697, "eval_steps_per_second": 6.091, "step": 42 }, { "epoch": 0.009432973482863432, "grad_norm": 0.30218949913978577, "learning_rate": 9.874639560909117e-05, "loss": 1.1916, "step": 45 }, { "epoch": 0.010061838381720993, "grad_norm": 0.3343711197376251, "learning_rate": 9.852339101019574e-05, "loss": 1.1487, "step": 48 }, { "epoch": 0.010690703280578556, "grad_norm": 0.3268805146217346, "learning_rate": 9.828243544427796e-05, "loss": 1.1781, "step": 51 }, { "epoch": 0.011319568179436117, "grad_norm": 0.34138286113739014, "learning_rate": 9.802361805155097e-05, "loss": 1.1215, "step": 54 }, { "epoch": 0.01194843307829368, "grad_norm": 0.3412030339241028, "learning_rate": 9.774703458011453e-05, "loss": 1.2044, "step": 57 }, { "epoch": 0.012577297977151241, "grad_norm": 0.3168179392814636, "learning_rate": 9.745278735053343e-05, "loss": 1.1087, "step": 60 }, { "epoch": 0.013206162876008804, "grad_norm": 0.3409186005592346, "learning_rate": 9.714098521798465e-05, "loss": 1.2062, "step": 63 }, { "epoch": 0.013835027774866367, "grad_norm": 0.29721856117248535, "learning_rate": 9.681174353198687e-05, "loss": 1.1158, "step": 66 }, { "epoch": 0.014463892673723928, "grad_norm": 0.3369769752025604, "learning_rate": 9.64651840937276e-05, "loss": 1.1834, "step": 69 }, { "epoch": 0.015092757572581491, "grad_norm": 0.31268638372421265, "learning_rate": 9.610143511100354e-05, "loss": 1.1367, "step": 72 }, { "epoch": 0.015721622471439052, "grad_norm": 0.333360493183136, "learning_rate": 9.572063115079063e-05, "loss": 1.1805, "step": 75 }, { "epoch": 0.016350487370296613, "grad_norm": 0.2951726019382477, "learning_rate": 9.53229130894619e-05, "loss": 1.1459, "step": 78 }, { "epoch": 0.016979352269154178, "grad_norm": 0.36082521080970764, "learning_rate": 9.490842806067095e-05, "loss": 1.17, "step": 81 }, { "epoch": 0.01760821716801174, "grad_norm": 0.33050239086151123, "learning_rate": 9.44773294009206e-05, "loss": 1.1611, "step": 84 }, { "epoch": 0.01760821716801174, "eval_loss": 1.147385835647583, "eval_runtime": 165.0588, "eval_samples_per_second": 48.68, "eval_steps_per_second": 6.089, "step": 84 }, { "epoch": 0.0182370820668693, "grad_norm": 0.3466218411922455, "learning_rate": 9.40297765928369e-05, "loss": 1.181, "step": 87 }, { "epoch": 0.018865946965726865, "grad_norm": 0.3219515085220337, "learning_rate": 9.356593520616948e-05, "loss": 1.1236, "step": 90 }, { "epoch": 0.019494811864584426, "grad_norm": 0.35251516103744507, "learning_rate": 9.308597683653975e-05, "loss": 1.1749, "step": 93 }, { "epoch": 0.020123676763441987, "grad_norm": 0.3366820514202118, "learning_rate": 9.259007904196023e-05, "loss": 1.1411, "step": 96 }, { "epoch": 0.020752541662299548, "grad_norm": 0.3501774072647095, "learning_rate": 9.207842527714767e-05, "loss": 1.122, "step": 99 }, { "epoch": 0.021381406561157112, "grad_norm": 0.3260366916656494, "learning_rate": 9.155120482565521e-05, "loss": 1.1032, "step": 102 }, { "epoch": 0.022010271460014674, "grad_norm": 0.3130464255809784, "learning_rate": 9.10086127298478e-05, "loss": 1.1025, "step": 105 }, { "epoch": 0.022639136358872235, "grad_norm": 0.3319336473941803, "learning_rate": 9.045084971874738e-05, "loss": 1.1279, "step": 108 }, { "epoch": 0.0232680012577298, "grad_norm": 0.3683924973011017, "learning_rate": 8.987812213377424e-05, "loss": 1.1384, "step": 111 }, { "epoch": 0.02389686615658736, "grad_norm": 0.3043256998062134, "learning_rate": 8.929064185241213e-05, "loss": 1.106, "step": 114 }, { "epoch": 0.02452573105544492, "grad_norm": 0.3852984607219696, "learning_rate": 8.868862620982534e-05, "loss": 1.123, "step": 117 }, { "epoch": 0.025154595954302483, "grad_norm": 0.36246711015701294, "learning_rate": 8.807229791845673e-05, "loss": 1.1248, "step": 120 }, { "epoch": 0.025783460853160047, "grad_norm": 0.3592873215675354, "learning_rate": 8.744188498563641e-05, "loss": 1.0983, "step": 123 }, { "epoch": 0.02641232575201761, "grad_norm": 0.3461982309818268, "learning_rate": 8.679762062923175e-05, "loss": 1.1211, "step": 126 }, { "epoch": 0.02641232575201761, "eval_loss": 1.1246975660324097, "eval_runtime": 165.051, "eval_samples_per_second": 48.682, "eval_steps_per_second": 6.089, "step": 126 }, { "epoch": 0.02704119065087517, "grad_norm": 0.3487012982368469, "learning_rate": 8.613974319136958e-05, "loss": 1.1212, "step": 129 }, { "epoch": 0.027670055549732734, "grad_norm": 0.37176796793937683, "learning_rate": 8.54684960502629e-05, "loss": 1.0707, "step": 132 }, { "epoch": 0.028298920448590295, "grad_norm": 0.3854920268058777, "learning_rate": 8.478412753017433e-05, "loss": 1.1204, "step": 135 }, { "epoch": 0.028927785347447856, "grad_norm": 0.3561486303806305, "learning_rate": 8.408689080954998e-05, "loss": 1.1238, "step": 138 }, { "epoch": 0.029556650246305417, "grad_norm": 0.35637167096138, "learning_rate": 8.33770438273574e-05, "loss": 1.1415, "step": 141 }, { "epoch": 0.030185515145162982, "grad_norm": 0.37608301639556885, "learning_rate": 8.265484918766243e-05, "loss": 1.1374, "step": 144 }, { "epoch": 0.030814380044020543, "grad_norm": 0.3702179789543152, "learning_rate": 8.192057406248028e-05, "loss": 1.1001, "step": 147 }, { "epoch": 0.031443244942878104, "grad_norm": 0.35054582357406616, "learning_rate": 8.117449009293668e-05, "loss": 1.122, "step": 150 }, { "epoch": 0.03207210984173567, "grad_norm": 0.35157763957977295, "learning_rate": 8.041687328877567e-05, "loss": 1.1126, "step": 153 }, { "epoch": 0.032700974740593226, "grad_norm": 0.3567587435245514, "learning_rate": 7.964800392625129e-05, "loss": 1.1196, "step": 156 }, { "epoch": 0.03332983963945079, "grad_norm": 0.33376169204711914, "learning_rate": 7.886816644444098e-05, "loss": 1.1269, "step": 159 }, { "epoch": 0.033958704538308356, "grad_norm": 0.33005964756011963, "learning_rate": 7.807764934001874e-05, "loss": 1.0881, "step": 162 }, { "epoch": 0.03458756943716591, "grad_norm": 0.3465365171432495, "learning_rate": 7.727674506052743e-05, "loss": 1.0912, "step": 165 }, { "epoch": 0.03521643433602348, "grad_norm": 0.3523496985435486, "learning_rate": 7.646574989618938e-05, "loss": 1.0962, "step": 168 }, { "epoch": 0.03521643433602348, "eval_loss": 1.1077728271484375, "eval_runtime": 165.0063, "eval_samples_per_second": 48.695, "eval_steps_per_second": 6.091, "step": 168 }, { "epoch": 0.03584529923488104, "grad_norm": 0.3241373300552368, "learning_rate": 7.564496387029532e-05, "loss": 1.1243, "step": 171 }, { "epoch": 0.0364741641337386, "grad_norm": 0.33574438095092773, "learning_rate": 7.481469062821252e-05, "loss": 1.1031, "step": 174 }, { "epoch": 0.037103029032596165, "grad_norm": 0.38013267517089844, "learning_rate": 7.39752373250527e-05, "loss": 1.1191, "step": 177 }, { "epoch": 0.03773189393145373, "grad_norm": 0.35769525170326233, "learning_rate": 7.312691451204178e-05, "loss": 1.0992, "step": 180 }, { "epoch": 0.03836075883031129, "grad_norm": 0.32820338010787964, "learning_rate": 7.227003602163295e-05, "loss": 1.0621, "step": 183 }, { "epoch": 0.03898962372916885, "grad_norm": 0.32962238788604736, "learning_rate": 7.14049188514063e-05, "loss": 1.0828, "step": 186 }, { "epoch": 0.03961848862802641, "grad_norm": 0.34197545051574707, "learning_rate": 7.05318830467969e-05, "loss": 1.1168, "step": 189 }, { "epoch": 0.040247353526883974, "grad_norm": 0.37021368741989136, "learning_rate": 6.965125158269619e-05, "loss": 1.1374, "step": 192 }, { "epoch": 0.04087621842574154, "grad_norm": 0.3301369547843933, "learning_rate": 6.876335024396872e-05, "loss": 1.054, "step": 195 }, { "epoch": 0.041505083324599096, "grad_norm": 0.35479190945625305, "learning_rate": 6.786850750493006e-05, "loss": 1.0923, "step": 198 }, { "epoch": 0.04213394822345666, "grad_norm": 0.35211730003356934, "learning_rate": 6.696705440782938e-05, "loss": 1.1124, "step": 201 }, { "epoch": 0.042762813122314225, "grad_norm": 0.4026525616645813, "learning_rate": 6.605932444038229e-05, "loss": 1.1229, "step": 204 }, { "epoch": 0.04339167802117178, "grad_norm": 0.40407177805900574, "learning_rate": 6.514565341239861e-05, "loss": 1.066, "step": 207 }, { "epoch": 0.04402054292002935, "grad_norm": 0.3534255027770996, "learning_rate": 6.422637933155162e-05, "loss": 1.0462, "step": 210 }, { "epoch": 0.04402054292002935, "eval_loss": 1.0946003198623657, "eval_runtime": 165.0602, "eval_samples_per_second": 48.679, "eval_steps_per_second": 6.089, "step": 210 }, { "epoch": 0.04464940781888691, "grad_norm": 0.3846144676208496, "learning_rate": 6.330184227833376e-05, "loss": 1.0922, "step": 213 }, { "epoch": 0.04527827271774447, "grad_norm": 0.37346020340919495, "learning_rate": 6.237238428024572e-05, "loss": 1.0727, "step": 216 }, { "epoch": 0.045907137616602034, "grad_norm": 0.3563274145126343, "learning_rate": 6.143834918526527e-05, "loss": 1.144, "step": 219 }, { "epoch": 0.0465360025154596, "grad_norm": 0.3684130907058716, "learning_rate": 6.0500082534642464e-05, "loss": 1.0526, "step": 222 }, { "epoch": 0.047164867414317156, "grad_norm": 0.37623313069343567, "learning_rate": 5.955793143506863e-05, "loss": 1.0694, "step": 225 }, { "epoch": 0.04779373231317472, "grad_norm": 0.335705429315567, "learning_rate": 5.861224443026595e-05, "loss": 1.0766, "step": 228 }, { "epoch": 0.04842259721203228, "grad_norm": 0.35654446482658386, "learning_rate": 5.766337137204579e-05, "loss": 1.091, "step": 231 }, { "epoch": 0.04905146211088984, "grad_norm": 0.33855140209198, "learning_rate": 5.6711663290882776e-05, "loss": 1.0715, "step": 234 }, { "epoch": 0.04968032700974741, "grad_norm": 0.3658079206943512, "learning_rate": 5.575747226605298e-05, "loss": 1.0915, "step": 237 }, { "epoch": 0.050309191908604965, "grad_norm": 0.3918761610984802, "learning_rate": 5.480115129538409e-05, "loss": 1.073, "step": 240 }, { "epoch": 0.05093805680746253, "grad_norm": 0.37080520391464233, "learning_rate": 5.384305416466584e-05, "loss": 1.0234, "step": 243 }, { "epoch": 0.051566921706320094, "grad_norm": 0.315390020608902, "learning_rate": 5.288353531676873e-05, "loss": 1.0456, "step": 246 }, { "epoch": 0.05219578660517765, "grad_norm": 0.34985366463661194, "learning_rate": 5.192294972051992e-05, "loss": 1.0505, "step": 249 }, { "epoch": 0.05282465150403522, "grad_norm": 0.35391372442245483, "learning_rate": 5.0961652739384356e-05, "loss": 1.0728, "step": 252 }, { "epoch": 0.05282465150403522, "eval_loss": 1.0837104320526123, "eval_runtime": 164.9887, "eval_samples_per_second": 48.7, "eval_steps_per_second": 6.091, "step": 252 }, { "epoch": 0.05345351640289278, "grad_norm": 0.36929386854171753, "learning_rate": 5e-05, "loss": 1.0279, "step": 255 }, { "epoch": 0.05408238130175034, "grad_norm": 0.34936413168907166, "learning_rate": 4.903834726061565e-05, "loss": 1.0564, "step": 258 }, { "epoch": 0.0547112462006079, "grad_norm": 0.33937400579452515, "learning_rate": 4.807705027948008e-05, "loss": 1.0349, "step": 261 }, { "epoch": 0.05534011109946547, "grad_norm": 0.3877185583114624, "learning_rate": 4.711646468323129e-05, "loss": 1.0518, "step": 264 }, { "epoch": 0.055968975998323026, "grad_norm": 0.36920610070228577, "learning_rate": 4.6156945835334184e-05, "loss": 1.0907, "step": 267 }, { "epoch": 0.05659784089718059, "grad_norm": 0.3572535216808319, "learning_rate": 4.5198848704615914e-05, "loss": 1.0811, "step": 270 }, { "epoch": 0.05722670579603815, "grad_norm": 0.3565980792045593, "learning_rate": 4.424252773394704e-05, "loss": 1.0598, "step": 273 }, { "epoch": 0.05785557069489571, "grad_norm": 0.3740125298500061, "learning_rate": 4.328833670911724e-05, "loss": 1.0403, "step": 276 }, { "epoch": 0.05848443559375328, "grad_norm": 0.382012277841568, "learning_rate": 4.23366286279542e-05, "loss": 1.0877, "step": 279 }, { "epoch": 0.059113300492610835, "grad_norm": 0.359355092048645, "learning_rate": 4.138775556973406e-05, "loss": 1.0644, "step": 282 }, { "epoch": 0.0597421653914684, "grad_norm": 0.35905787348747253, "learning_rate": 4.04420685649314e-05, "loss": 1.0783, "step": 285 }, { "epoch": 0.060371030290325964, "grad_norm": 0.4150475263595581, "learning_rate": 3.9499917465357534e-05, "loss": 1.0996, "step": 288 }, { "epoch": 0.06099989518918352, "grad_norm": 0.3718733489513397, "learning_rate": 3.856165081473474e-05, "loss": 1.0191, "step": 291 }, { "epoch": 0.061628760088041086, "grad_norm": 0.36830800771713257, "learning_rate": 3.762761571975429e-05, "loss": 1.1456, "step": 294 }, { "epoch": 0.061628760088041086, "eval_loss": 1.0759488344192505, "eval_runtime": 164.9861, "eval_samples_per_second": 48.701, "eval_steps_per_second": 6.091, "step": 294 }, { "epoch": 0.06225762498689865, "grad_norm": 0.36433523893356323, "learning_rate": 3.6698157721666246e-05, "loss": 1.0619, "step": 297 }, { "epoch": 0.06288648988575621, "grad_norm": 0.382252961397171, "learning_rate": 3.5773620668448384e-05, "loss": 1.0544, "step": 300 }, { "epoch": 0.06351535478461377, "grad_norm": 0.4051714539527893, "learning_rate": 3.48543465876014e-05, "loss": 0.9961, "step": 303 }, { "epoch": 0.06414421968347134, "grad_norm": 0.34751641750335693, "learning_rate": 3.3940675559617724e-05, "loss": 1.0396, "step": 306 }, { "epoch": 0.0647730845823289, "grad_norm": 0.38340121507644653, "learning_rate": 3.303294559217063e-05, "loss": 1.0916, "step": 309 }, { "epoch": 0.06540194948118645, "grad_norm": 0.41020599007606506, "learning_rate": 3.213149249506997e-05, "loss": 1.0738, "step": 312 }, { "epoch": 0.06603081438004402, "grad_norm": 0.3864920735359192, "learning_rate": 3.12366497560313e-05, "loss": 1.0936, "step": 315 }, { "epoch": 0.06665967927890158, "grad_norm": 0.35983094573020935, "learning_rate": 3.0348748417303823e-05, "loss": 1.1001, "step": 318 }, { "epoch": 0.06728854417775915, "grad_norm": 0.39053815603256226, "learning_rate": 2.9468116953203107e-05, "loss": 1.0366, "step": 321 }, { "epoch": 0.06791740907661671, "grad_norm": 0.3519478142261505, "learning_rate": 2.8595081148593738e-05, "loss": 1.044, "step": 324 }, { "epoch": 0.06854627397547428, "grad_norm": 0.37353023886680603, "learning_rate": 2.772996397836704e-05, "loss": 1.0449, "step": 327 }, { "epoch": 0.06917513887433183, "grad_norm": 0.357658714056015, "learning_rate": 2.687308548795825e-05, "loss": 1.0519, "step": 330 }, { "epoch": 0.06980400377318939, "grad_norm": 0.40195825695991516, "learning_rate": 2.6024762674947313e-05, "loss": 1.0915, "step": 333 }, { "epoch": 0.07043286867204696, "grad_norm": 0.3929975628852844, "learning_rate": 2.5185309371787513e-05, "loss": 1.0375, "step": 336 }, { "epoch": 0.07043286867204696, "eval_loss": 1.0703336000442505, "eval_runtime": 165.0152, "eval_samples_per_second": 48.692, "eval_steps_per_second": 6.09, "step": 336 }, { "epoch": 0.07106173357090452, "grad_norm": 0.4022517502307892, "learning_rate": 2.43550361297047e-05, "loss": 1.0749, "step": 339 }, { "epoch": 0.07169059846976208, "grad_norm": 0.35164332389831543, "learning_rate": 2.353425010381063e-05, "loss": 1.0352, "step": 342 }, { "epoch": 0.07231946336861964, "grad_norm": 0.40566059947013855, "learning_rate": 2.272325493947257e-05, "loss": 1.0625, "step": 345 }, { "epoch": 0.0729483282674772, "grad_norm": 0.37188711762428284, "learning_rate": 2.192235065998126e-05, "loss": 1.072, "step": 348 }, { "epoch": 0.07357719316633476, "grad_norm": 0.3737729489803314, "learning_rate": 2.1131833555559037e-05, "loss": 1.0546, "step": 351 }, { "epoch": 0.07420605806519233, "grad_norm": 0.35681530833244324, "learning_rate": 2.0351996073748713e-05, "loss": 1.0669, "step": 354 }, { "epoch": 0.0748349229640499, "grad_norm": 0.3745366334915161, "learning_rate": 1.9583126711224343e-05, "loss": 1.0731, "step": 357 }, { "epoch": 0.07546378786290746, "grad_norm": 0.3998556137084961, "learning_rate": 1.8825509907063327e-05, "loss": 1.1069, "step": 360 }, { "epoch": 0.07609265276176501, "grad_norm": 0.4358106255531311, "learning_rate": 1.807942593751973e-05, "loss": 1.0876, "step": 363 }, { "epoch": 0.07672151766062257, "grad_norm": 0.3841058313846588, "learning_rate": 1.7345150812337564e-05, "loss": 1.0822, "step": 366 }, { "epoch": 0.07735038255948014, "grad_norm": 0.4276648759841919, "learning_rate": 1.66229561726426e-05, "loss": 1.0894, "step": 369 }, { "epoch": 0.0779792474583377, "grad_norm": 0.40756258368492126, "learning_rate": 1.5913109190450032e-05, "loss": 1.0673, "step": 372 }, { "epoch": 0.07860811235719527, "grad_norm": 0.34232285618782043, "learning_rate": 1.5215872469825682e-05, "loss": 1.0659, "step": 375 }, { "epoch": 0.07923697725605282, "grad_norm": 0.35964226722717285, "learning_rate": 1.4531503949737108e-05, "loss": 1.0067, "step": 378 }, { "epoch": 0.07923697725605282, "eval_loss": 1.0661753416061401, "eval_runtime": 165.0357, "eval_samples_per_second": 48.686, "eval_steps_per_second": 6.09, "step": 378 }, { "epoch": 0.07986584215491038, "grad_norm": 0.3825712203979492, "learning_rate": 1.3860256808630428e-05, "loss": 1.1019, "step": 381 }, { "epoch": 0.08049470705376795, "grad_norm": 0.39544931054115295, "learning_rate": 1.3202379370768252e-05, "loss": 1.0646, "step": 384 }, { "epoch": 0.08112357195262551, "grad_norm": 0.3967124819755554, "learning_rate": 1.2558115014363592e-05, "loss": 1.0394, "step": 387 }, { "epoch": 0.08175243685148308, "grad_norm": 0.39682498574256897, "learning_rate": 1.1927702081543279e-05, "loss": 1.0968, "step": 390 }, { "epoch": 0.08238130175034064, "grad_norm": 0.3636738359928131, "learning_rate": 1.1311373790174657e-05, "loss": 1.031, "step": 393 }, { "epoch": 0.08301016664919819, "grad_norm": 0.32566869258880615, "learning_rate": 1.0709358147587884e-05, "loss": 1.094, "step": 396 }, { "epoch": 0.08363903154805576, "grad_norm": 0.38560715317726135, "learning_rate": 1.0121877866225781e-05, "loss": 1.1144, "step": 399 }, { "epoch": 0.08426789644691332, "grad_norm": 0.32941296696662903, "learning_rate": 9.549150281252633e-06, "loss": 1.0141, "step": 402 }, { "epoch": 0.08489676134577089, "grad_norm": 0.36348846554756165, "learning_rate": 8.991387270152201e-06, "loss": 1.043, "step": 405 }, { "epoch": 0.08552562624462845, "grad_norm": 0.37229597568511963, "learning_rate": 8.448795174344804e-06, "loss": 1.077, "step": 408 }, { "epoch": 0.08615449114348601, "grad_norm": 0.3720276653766632, "learning_rate": 7.921574722852343e-06, "loss": 0.9869, "step": 411 }, { "epoch": 0.08678335604234357, "grad_norm": 0.3702375590801239, "learning_rate": 7.409920958039795e-06, "loss": 0.9772, "step": 414 }, { "epoch": 0.08741222094120113, "grad_norm": 0.37576282024383545, "learning_rate": 6.9140231634602485e-06, "loss": 1.0302, "step": 417 }, { "epoch": 0.0880410858400587, "grad_norm": 0.39544618129730225, "learning_rate": 6.43406479383053e-06, "loss": 1.0528, "step": 420 }, { "epoch": 0.0880410858400587, "eval_loss": 1.0637660026550293, "eval_runtime": 165.1277, "eval_samples_per_second": 48.659, "eval_steps_per_second": 6.086, "step": 420 }, { "epoch": 0.08866995073891626, "grad_norm": 0.4187450110912323, "learning_rate": 5.9702234071631e-06, "loss": 1.0578, "step": 423 }, { "epoch": 0.08929881563777382, "grad_norm": 0.3692411780357361, "learning_rate": 5.5226705990794155e-06, "loss": 1.0522, "step": 426 }, { "epoch": 0.08992768053663137, "grad_norm": 0.4078660309314728, "learning_rate": 5.091571939329048e-06, "loss": 1.0434, "step": 429 }, { "epoch": 0.09055654543548894, "grad_norm": 0.4134742319583893, "learning_rate": 4.677086910538092e-06, "loss": 1.074, "step": 432 }, { "epoch": 0.0911854103343465, "grad_norm": 0.36238303780555725, "learning_rate": 4.279368849209381e-06, "loss": 1.0455, "step": 435 }, { "epoch": 0.09181427523320407, "grad_norm": 0.4054090976715088, "learning_rate": 3.898564888996476e-06, "loss": 1.0706, "step": 438 }, { "epoch": 0.09244314013206163, "grad_norm": 0.3747381567955017, "learning_rate": 3.534815906272404e-06, "loss": 1.0354, "step": 441 }, { "epoch": 0.0930720050309192, "grad_norm": 0.3801961839199066, "learning_rate": 3.18825646801314e-06, "loss": 1.0818, "step": 444 }, { "epoch": 0.09370086992977675, "grad_norm": 0.3710772395133972, "learning_rate": 2.8590147820153513e-06, "loss": 1.0605, "step": 447 }, { "epoch": 0.09432973482863431, "grad_norm": 0.3717288672924042, "learning_rate": 2.547212649466568e-06, "loss": 1.0153, "step": 450 }, { "epoch": 0.09495859972749188, "grad_norm": 0.370351105928421, "learning_rate": 2.2529654198854835e-06, "loss": 1.037, "step": 453 }, { "epoch": 0.09558746462634944, "grad_norm": 0.38800954818725586, "learning_rate": 1.9763819484490355e-06, "loss": 1.0668, "step": 456 }, { "epoch": 0.096216329525207, "grad_norm": 0.39085131883621216, "learning_rate": 1.7175645557220566e-06, "loss": 1.0601, "step": 459 }, { "epoch": 0.09684519442406456, "grad_norm": 0.38424739241600037, "learning_rate": 1.4766089898042678e-06, "loss": 1.0742, "step": 462 }, { "epoch": 0.09684519442406456, "eval_loss": 1.0629161596298218, "eval_runtime": 165.0993, "eval_samples_per_second": 48.668, "eval_steps_per_second": 6.087, "step": 462 }, { "epoch": 0.09747405932292212, "grad_norm": 0.34824374318122864, "learning_rate": 1.2536043909088191e-06, "loss": 1.0597, "step": 465 }, { "epoch": 0.09810292422177969, "grad_norm": 0.3800143599510193, "learning_rate": 1.0486332583853563e-06, "loss": 1.0733, "step": 468 }, { "epoch": 0.09873178912063725, "grad_norm": 0.39402127265930176, "learning_rate": 8.617714201998084e-07, "loss": 1.0444, "step": 471 }, { "epoch": 0.09936065401949482, "grad_norm": 0.3665209114551544, "learning_rate": 6.93088004882253e-07, "loss": 1.0652, "step": 474 }, { "epoch": 0.09998951891835238, "grad_norm": 0.3840101659297943, "learning_rate": 5.426454159531913e-07, "loss": 1.0667, "step": 477 }, { "epoch": 0.10061838381720993, "grad_norm": 0.40634945034980774, "learning_rate": 4.104993088376974e-07, "loss": 1.024, "step": 480 }, { "epoch": 0.1012472487160675, "grad_norm": 0.4399307370185852, "learning_rate": 2.966985702759828e-07, "loss": 1.0763, "step": 483 }, { "epoch": 0.10187611361492506, "grad_norm": 0.3584943115711212, "learning_rate": 2.012853002380466e-07, "loss": 1.0044, "step": 486 }, { "epoch": 0.10250497851378262, "grad_norm": 0.3854651153087616, "learning_rate": 1.2429479634897267e-07, "loss": 1.0373, "step": 489 }, { "epoch": 0.10313384341264019, "grad_norm": 0.3579770028591156, "learning_rate": 6.575554083078084e-08, "loss": 1.0518, "step": 492 }, { "epoch": 0.10376270831149775, "grad_norm": 0.38003799319267273, "learning_rate": 2.568918996560532e-08, "loss": 1.1002, "step": 495 }, { "epoch": 0.1043915732103553, "grad_norm": 0.35773521661758423, "learning_rate": 4.110566084036816e-09, "loss": 1.062, "step": 498 } ], "logging_steps": 3, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 42, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.06657392623616e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }