{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.11373329542223486, "eval_steps": 34, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002843332385555871, "eval_loss": 1.5018354654312134, "eval_runtime": 166.295, "eval_samples_per_second": 35.623, "eval_steps_per_second": 4.456, "step": 1 }, { "epoch": 0.0008529997156667614, "grad_norm": 0.7210565209388733, "learning_rate": 1.5e-05, "loss": 1.5223, "step": 3 }, { "epoch": 0.0017059994313335229, "grad_norm": 0.6832892298698425, "learning_rate": 3e-05, "loss": 1.5221, "step": 6 }, { "epoch": 0.0025589991470002845, "grad_norm": 0.7228020429611206, "learning_rate": 4.5e-05, "loss": 1.408, "step": 9 }, { "epoch": 0.0034119988626670457, "grad_norm": 0.6384875774383545, "learning_rate": 4.999675562428437e-05, "loss": 1.4404, "step": 12 }, { "epoch": 0.004264998578333807, "grad_norm": 0.3351251482963562, "learning_rate": 4.9979724954289244e-05, "loss": 1.3736, "step": 15 }, { "epoch": 0.005117998294000569, "grad_norm": 0.182565838098526, "learning_rate": 4.994810682835951e-05, "loss": 1.3079, "step": 18 }, { "epoch": 0.00597099800966733, "grad_norm": 0.19113591313362122, "learning_rate": 4.990191971059033e-05, "loss": 1.316, "step": 21 }, { "epoch": 0.0068239977253340914, "grad_norm": 0.1948336660861969, "learning_rate": 4.984119057295783e-05, "loss": 1.2426, "step": 24 }, { "epoch": 0.007676997441000853, "grad_norm": 0.18684643507003784, "learning_rate": 4.976595487956823e-05, "loss": 1.2503, "step": 27 }, { "epoch": 0.008529997156667614, "grad_norm": 0.18414735794067383, "learning_rate": 4.967625656594782e-05, "loss": 1.2323, "step": 30 }, { "epoch": 0.009382996872334376, "grad_norm": 0.17294025421142578, "learning_rate": 4.957214801338581e-05, "loss": 1.1942, "step": 33 }, { "epoch": 0.009667330110889964, "eval_loss": 1.1813108921051025, "eval_runtime": 167.7438, "eval_samples_per_second": 35.316, "eval_steps_per_second": 4.417, "step": 34 }, { "epoch": 0.010235996588001138, "grad_norm": 0.16702412068843842, "learning_rate": 4.9453690018345144e-05, "loss": 1.1981, "step": 36 }, { "epoch": 0.011088996303667898, "grad_norm": 0.1968175172805786, "learning_rate": 4.932095175695911e-05, "loss": 1.1675, "step": 39 }, { "epoch": 0.01194199601933466, "grad_norm": 0.18244469165802002, "learning_rate": 4.917401074463441e-05, "loss": 1.1584, "step": 42 }, { "epoch": 0.01279499573500142, "grad_norm": 0.16749081015586853, "learning_rate": 4.901295279078431e-05, "loss": 1.1134, "step": 45 }, { "epoch": 0.013647995450668183, "grad_norm": 0.17398597300052643, "learning_rate": 4.883787194871841e-05, "loss": 1.1139, "step": 48 }, { "epoch": 0.014500995166334945, "grad_norm": 0.17164087295532227, "learning_rate": 4.864887046071813e-05, "loss": 1.079, "step": 51 }, { "epoch": 0.015353994882001705, "grad_norm": 0.1644001007080078, "learning_rate": 4.8446058698330115e-05, "loss": 1.0646, "step": 54 }, { "epoch": 0.01620699459766847, "grad_norm": 0.16490623354911804, "learning_rate": 4.822955509791233e-05, "loss": 1.0739, "step": 57 }, { "epoch": 0.017059994313335228, "grad_norm": 0.17708458006381989, "learning_rate": 4.799948609147061e-05, "loss": 1.0897, "step": 60 }, { "epoch": 0.01791299402900199, "grad_norm": 0.15597032010555267, "learning_rate": 4.7755986032825864e-05, "loss": 1.0566, "step": 63 }, { "epoch": 0.018765993744668752, "grad_norm": 0.17728550732135773, "learning_rate": 4.74991971191553e-05, "loss": 1.0275, "step": 66 }, { "epoch": 0.019334660221779928, "eval_loss": 1.0011852979660034, "eval_runtime": 168.1174, "eval_samples_per_second": 35.237, "eval_steps_per_second": 4.408, "step": 68 }, { "epoch": 0.019618993460335514, "grad_norm": 0.16994404792785645, "learning_rate": 4.7229269307953235e-05, "loss": 0.9841, "step": 69 }, { "epoch": 0.020471993176002276, "grad_norm": 0.17181497812271118, "learning_rate": 4.694636022946012e-05, "loss": 0.9944, "step": 72 }, { "epoch": 0.021324992891669035, "grad_norm": 0.21411621570587158, "learning_rate": 4.665063509461097e-05, "loss": 1.0013, "step": 75 }, { "epoch": 0.022177992607335797, "grad_norm": 0.19594340026378632, "learning_rate": 4.6342266598556814e-05, "loss": 0.9969, "step": 78 }, { "epoch": 0.02303099232300256, "grad_norm": 0.2057276964187622, "learning_rate": 4.6021434819815555e-05, "loss": 0.9808, "step": 81 }, { "epoch": 0.02388399203866932, "grad_norm": 0.21778051555156708, "learning_rate": 4.568832711511125e-05, "loss": 0.9456, "step": 84 }, { "epoch": 0.024736991754336083, "grad_norm": 0.21349306404590607, "learning_rate": 4.534313800996299e-05, "loss": 0.953, "step": 87 }, { "epoch": 0.02558999147000284, "grad_norm": 0.21818219125270844, "learning_rate": 4.498606908508754e-05, "loss": 0.9133, "step": 90 }, { "epoch": 0.026442991185669604, "grad_norm": 0.2510223388671875, "learning_rate": 4.46173288586818e-05, "loss": 0.9125, "step": 93 }, { "epoch": 0.027295990901336366, "grad_norm": 0.24031595885753632, "learning_rate": 4.4237132664654154e-05, "loss": 0.8784, "step": 96 }, { "epoch": 0.028148990617003128, "grad_norm": 0.2543729543685913, "learning_rate": 4.384570252687542e-05, "loss": 0.8984, "step": 99 }, { "epoch": 0.02900199033266989, "grad_norm": 0.27020716667175293, "learning_rate": 4.344326702952326e-05, "loss": 0.8719, "step": 102 }, { "epoch": 0.02900199033266989, "eval_loss": 0.8574855923652649, "eval_runtime": 168.0651, "eval_samples_per_second": 35.248, "eval_steps_per_second": 4.409, "step": 102 }, { "epoch": 0.029854990048336652, "grad_norm": 0.2670081853866577, "learning_rate": 4.303006118359537e-05, "loss": 0.8789, "step": 105 }, { "epoch": 0.03070798976400341, "grad_norm": 0.3037591874599457, "learning_rate": 4.260632628966974e-05, "loss": 0.8526, "step": 108 }, { "epoch": 0.031560989479670176, "grad_norm": 0.28440290689468384, "learning_rate": 4.217230979699188e-05, "loss": 0.8274, "step": 111 }, { "epoch": 0.03241398919533694, "grad_norm": 0.49433842301368713, "learning_rate": 4.172826515897146e-05, "loss": 0.8104, "step": 114 }, { "epoch": 0.03326698891100369, "grad_norm": 0.2911880314350128, "learning_rate": 4.12744516851726e-05, "loss": 0.8392, "step": 117 }, { "epoch": 0.034119988626670456, "grad_norm": 0.3204285502433777, "learning_rate": 4.0811134389884433e-05, "loss": 0.788, "step": 120 }, { "epoch": 0.03497298834233722, "grad_norm": 0.35650599002838135, "learning_rate": 4.0338583837360225e-05, "loss": 0.8004, "step": 123 }, { "epoch": 0.03582598805800398, "grad_norm": 0.310162216424942, "learning_rate": 3.985707598381544e-05, "loss": 0.8085, "step": 126 }, { "epoch": 0.03667898777367074, "grad_norm": 0.3370768129825592, "learning_rate": 3.9366892016277096e-05, "loss": 0.7317, "step": 129 }, { "epoch": 0.037531987489337504, "grad_norm": 0.4600171446800232, "learning_rate": 3.886831818837847e-05, "loss": 0.7626, "step": 132 }, { "epoch": 0.038384987205004266, "grad_norm": 0.43600377440452576, "learning_rate": 3.8361645653195026e-05, "loss": 0.7679, "step": 135 }, { "epoch": 0.038669320443559856, "eval_loss": 0.7138456106185913, "eval_runtime": 168.0686, "eval_samples_per_second": 35.248, "eval_steps_per_second": 4.409, "step": 136 }, { "epoch": 0.03923798692067103, "grad_norm": 0.4538515508174896, "learning_rate": 3.784717029321922e-05, "loss": 0.7585, "step": 138 }, { "epoch": 0.04009098663633779, "grad_norm": 0.5923985838890076, "learning_rate": 3.732519254757344e-05, "loss": 0.7211, "step": 141 }, { "epoch": 0.04094398635200455, "grad_norm": 0.36592337489128113, "learning_rate": 3.679601723656205e-05, "loss": 0.6604, "step": 144 }, { "epoch": 0.04179698606767131, "grad_norm": 0.44677990674972534, "learning_rate": 3.625995338366492e-05, "loss": 0.7846, "step": 147 }, { "epoch": 0.04264998578333807, "grad_norm": 0.4921860098838806, "learning_rate": 3.5717314035076355e-05, "loss": 0.6709, "step": 150 }, { "epoch": 0.04350298549900483, "grad_norm": 0.5457685589790344, "learning_rate": 3.516841607689501e-05, "loss": 0.6974, "step": 153 }, { "epoch": 0.044355985214671594, "grad_norm": 0.46299344301223755, "learning_rate": 3.461358005007128e-05, "loss": 0.6643, "step": 156 }, { "epoch": 0.045208984930338356, "grad_norm": 0.44877251982688904, "learning_rate": 3.405312996322042e-05, "loss": 0.6035, "step": 159 }, { "epoch": 0.04606198464600512, "grad_norm": 0.3635808527469635, "learning_rate": 3.348739310341068e-05, "loss": 0.6781, "step": 162 }, { "epoch": 0.04691498436167188, "grad_norm": 0.4172018766403198, "learning_rate": 3.2916699845036816e-05, "loss": 0.6195, "step": 165 }, { "epoch": 0.04776798407733864, "grad_norm": 0.36372795701026917, "learning_rate": 3.234138345689077e-05, "loss": 0.6599, "step": 168 }, { "epoch": 0.048336650554449814, "eval_loss": 0.606797456741333, "eval_runtime": 168.0542, "eval_samples_per_second": 35.251, "eval_steps_per_second": 4.409, "step": 170 }, { "epoch": 0.048620983793005404, "grad_norm": 0.4114479720592499, "learning_rate": 3.17617799075421e-05, "loss": 0.6287, "step": 171 }, { "epoch": 0.049473983508672166, "grad_norm": 0.36733385920524597, "learning_rate": 3.1178227669141744e-05, "loss": 0.621, "step": 174 }, { "epoch": 0.05032698322433893, "grad_norm": 0.3716995418071747, "learning_rate": 3.0591067519763895e-05, "loss": 0.5532, "step": 177 }, { "epoch": 0.05117998294000568, "grad_norm": 0.4997945725917816, "learning_rate": 3.0000642344401113e-05, "loss": 0.6024, "step": 180 }, { "epoch": 0.052032982655672445, "grad_norm": 0.43555817008018494, "learning_rate": 2.9407296934729227e-05, "loss": 0.5092, "step": 183 }, { "epoch": 0.05288598237133921, "grad_norm": 0.49766504764556885, "learning_rate": 2.8811377787758636e-05, "loss": 0.6556, "step": 186 }, { "epoch": 0.05373898208700597, "grad_norm": 0.6171467304229736, "learning_rate": 2.8213232903489865e-05, "loss": 0.5736, "step": 189 }, { "epoch": 0.05459198180267273, "grad_norm": 0.8651450276374817, "learning_rate": 2.761321158169134e-05, "loss": 0.5568, "step": 192 }, { "epoch": 0.055444981518339494, "grad_norm": 0.4486936926841736, "learning_rate": 2.7011664217918154e-05, "loss": 0.6145, "step": 195 }, { "epoch": 0.056297981234006256, "grad_norm": 0.5348999500274658, "learning_rate": 2.6408942098890936e-05, "loss": 0.591, "step": 198 }, { "epoch": 0.05715098094967302, "grad_norm": 0.4498997628688812, "learning_rate": 2.580539719735433e-05, "loss": 0.5572, "step": 201 }, { "epoch": 0.05800398066533978, "grad_norm": 0.5493082404136658, "learning_rate": 2.5201381966534748e-05, "loss": 0.5316, "step": 204 }, { "epoch": 0.05800398066533978, "eval_loss": 0.5341343879699707, "eval_runtime": 167.9533, "eval_samples_per_second": 35.272, "eval_steps_per_second": 4.412, "step": 204 }, { "epoch": 0.05885698038100654, "grad_norm": 0.47230827808380127, "learning_rate": 2.459724913431772e-05, "loss": 0.5152, "step": 207 }, { "epoch": 0.059709980096673304, "grad_norm": 0.5665333867073059, "learning_rate": 2.399335149726463e-05, "loss": 0.5736, "step": 210 }, { "epoch": 0.06056297981234006, "grad_norm": 0.47652244567871094, "learning_rate": 2.3390041714589514e-05, "loss": 0.5006, "step": 213 }, { "epoch": 0.06141597952800682, "grad_norm": 0.7172293663024902, "learning_rate": 2.2787672102216042e-05, "loss": 0.4619, "step": 216 }, { "epoch": 0.06226897924367358, "grad_norm": 0.5251888036727905, "learning_rate": 2.2186594427034864e-05, "loss": 0.4852, "step": 219 }, { "epoch": 0.06312197895934035, "grad_norm": 0.45427563786506653, "learning_rate": 2.1587159701481716e-05, "loss": 0.4936, "step": 222 }, { "epoch": 0.06397497867500711, "grad_norm": 0.49813351035118103, "learning_rate": 2.098971797855599e-05, "loss": 0.5062, "step": 225 }, { "epoch": 0.06482797839067388, "grad_norm": 0.510427713394165, "learning_rate": 2.0394618147399713e-05, "loss": 0.4742, "step": 228 }, { "epoch": 0.06568097810634063, "grad_norm": 0.5860615372657776, "learning_rate": 1.980220772955602e-05, "loss": 0.5472, "step": 231 }, { "epoch": 0.06653397782200739, "grad_norm": 0.45956236124038696, "learning_rate": 1.921283267602643e-05, "loss": 0.5065, "step": 234 }, { "epoch": 0.06738697753767416, "grad_norm": 0.46069616079330444, "learning_rate": 1.8626837165245165e-05, "loss": 0.4309, "step": 237 }, { "epoch": 0.06767131077622975, "eval_loss": 0.4809924364089966, "eval_runtime": 167.8669, "eval_samples_per_second": 35.29, "eval_steps_per_second": 4.414, "step": 238 }, { "epoch": 0.06823997725334091, "grad_norm": 0.4818466603755951, "learning_rate": 1.8044563402088684e-05, "loss": 0.5623, "step": 240 }, { "epoch": 0.06909297696900768, "grad_norm": 0.45440998673439026, "learning_rate": 1.746635141803761e-05, "loss": 0.5626, "step": 243 }, { "epoch": 0.06994597668467444, "grad_norm": 0.5940297842025757, "learning_rate": 1.6892538872607937e-05, "loss": 0.5264, "step": 246 }, { "epoch": 0.0707989764003412, "grad_norm": 0.41404712200164795, "learning_rate": 1.6323460856167426e-05, "loss": 0.4484, "step": 249 }, { "epoch": 0.07165197611600796, "grad_norm": 0.44981393218040466, "learning_rate": 1.5759449694252226e-05, "loss": 0.473, "step": 252 }, { "epoch": 0.07250497583167473, "grad_norm": 0.6551511883735657, "learning_rate": 1.5200834753498128e-05, "loss": 0.4956, "step": 255 }, { "epoch": 0.07335797554734148, "grad_norm": 0.5032558441162109, "learning_rate": 1.4647942249299707e-05, "loss": 0.5262, "step": 258 }, { "epoch": 0.07421097526300825, "grad_norm": 0.5430291891098022, "learning_rate": 1.4101095055309746e-05, "loss": 0.4899, "step": 261 }, { "epoch": 0.07506397497867501, "grad_norm": 0.5020308494567871, "learning_rate": 1.356061251489012e-05, "loss": 0.494, "step": 264 }, { "epoch": 0.07591697469434176, "grad_norm": 0.5282646417617798, "learning_rate": 1.302681025462424e-05, "loss": 0.5339, "step": 267 }, { "epoch": 0.07676997441000853, "grad_norm": 0.48478028178215027, "learning_rate": 1.2500000000000006e-05, "loss": 0.5133, "step": 270 }, { "epoch": 0.07733864088711971, "eval_loss": 0.4457505941390991, "eval_runtime": 168.037, "eval_samples_per_second": 35.254, "eval_steps_per_second": 4.41, "step": 272 }, { "epoch": 0.07762297412567529, "grad_norm": 0.6064507961273193, "learning_rate": 1.1980489393370938e-05, "loss": 0.4708, "step": 273 }, { "epoch": 0.07847597384134206, "grad_norm": 0.7535691261291504, "learning_rate": 1.1468581814301717e-05, "loss": 0.426, "step": 276 }, { "epoch": 0.07932897355700881, "grad_norm": 0.5106021165847778, "learning_rate": 1.096457620240298e-05, "loss": 0.457, "step": 279 }, { "epoch": 0.08018197327267558, "grad_norm": 0.7610855102539062, "learning_rate": 1.0468766882759094e-05, "loss": 0.3867, "step": 282 }, { "epoch": 0.08103497298834234, "grad_norm": 0.57286137342453, "learning_rate": 9.981443394050525e-06, "loss": 0.4744, "step": 285 }, { "epoch": 0.0818879727040091, "grad_norm": 0.6350270509719849, "learning_rate": 9.502890319471491e-06, "loss": 0.4382, "step": 288 }, { "epoch": 0.08274097241967586, "grad_norm": 0.5878217816352844, "learning_rate": 9.033387120541306e-06, "loss": 0.4471, "step": 291 }, { "epoch": 0.08359397213534261, "grad_norm": 0.4584065079689026, "learning_rate": 8.573207973906735e-06, "loss": 0.4223, "step": 294 }, { "epoch": 0.08444697185100938, "grad_norm": 0.514761745929718, "learning_rate": 8.1226216112306e-06, "loss": 0.4439, "step": 297 }, { "epoch": 0.08529997156667614, "grad_norm": 0.4604704678058624, "learning_rate": 7.681891162260015e-06, "loss": 0.4769, "step": 300 }, { "epoch": 0.08615297128234291, "grad_norm": 0.5754848718643188, "learning_rate": 7.251274001166044e-06, "loss": 0.4719, "step": 303 }, { "epoch": 0.08700597099800966, "grad_norm": 0.6474661827087402, "learning_rate": 6.831021596244424e-06, "loss": 0.3843, "step": 306 }, { "epoch": 0.08700597099800966, "eval_loss": 0.4252224862575531, "eval_runtime": 168.0333, "eval_samples_per_second": 35.255, "eval_steps_per_second": 4.41, "step": 306 }, { "epoch": 0.08785897071367643, "grad_norm": 0.5052304863929749, "learning_rate": 6.421379363065142e-06, "loss": 0.3779, "step": 309 }, { "epoch": 0.08871197042934319, "grad_norm": 0.5623005628585815, "learning_rate": 6.022586521156715e-06, "loss": 0.5177, "step": 312 }, { "epoch": 0.08956497014500996, "grad_norm": 0.615669846534729, "learning_rate": 5.634875954308638e-06, "loss": 0.3902, "step": 315 }, { "epoch": 0.09041796986067671, "grad_norm": 0.5363687872886658, "learning_rate": 5.258474074573877e-06, "loss": 0.3245, "step": 318 }, { "epoch": 0.09127096957634348, "grad_norm": 0.45147329568862915, "learning_rate": 4.893600690050579e-06, "loss": 0.4372, "step": 321 }, { "epoch": 0.09212396929201024, "grad_norm": 0.42790791392326355, "learning_rate": 4.540468876520323e-06, "loss": 0.4359, "step": 324 }, { "epoch": 0.09297696900767699, "grad_norm": 0.512692928314209, "learning_rate": 4.199284853017896e-06, "loss": 0.41, "step": 327 }, { "epoch": 0.09382996872334376, "grad_norm": 0.5399787425994873, "learning_rate": 3.8702478614051355e-06, "loss": 0.4455, "step": 330 }, { "epoch": 0.09468296843901051, "grad_norm": 0.5106746554374695, "learning_rate": 3.5535500500193357e-06, "loss": 0.4301, "step": 333 }, { "epoch": 0.09553596815467728, "grad_norm": 0.6958709955215454, "learning_rate": 3.249376361464021e-06, "loss": 0.3602, "step": 336 }, { "epoch": 0.09638896787034404, "grad_norm": 0.5049157738685608, "learning_rate": 2.957904424607652e-06, "loss": 0.35, "step": 339 }, { "epoch": 0.09667330110889963, "eval_loss": 0.4159228503704071, "eval_runtime": 167.9747, "eval_samples_per_second": 35.267, "eval_steps_per_second": 4.411, "step": 340 }, { "epoch": 0.09724196758601081, "grad_norm": 0.5438792705535889, "learning_rate": 2.679304450853401e-06, "loss": 0.4406, "step": 342 }, { "epoch": 0.09809496730167756, "grad_norm": 0.6174736022949219, "learning_rate": 2.4137391347404476e-06, "loss": 0.4503, "step": 345 }, { "epoch": 0.09894796701734433, "grad_norm": 0.5348644256591797, "learning_rate": 2.1613635589349756e-06, "loss": 0.5056, "step": 348 }, { "epoch": 0.09980096673301109, "grad_norm": 0.44019100069999695, "learning_rate": 1.922325103666281e-06, "loss": 0.3926, "step": 351 }, { "epoch": 0.10065396644867786, "grad_norm": 0.5055895447731018, "learning_rate": 1.696763360660808e-06, "loss": 0.5037, "step": 354 }, { "epoch": 0.10150696616434461, "grad_norm": 0.5478758215904236, "learning_rate": 1.4848100516245717e-06, "loss": 0.3948, "step": 357 }, { "epoch": 0.10235996588001137, "grad_norm": 0.5242781639099121, "learning_rate": 1.286588951321363e-06, "loss": 0.4522, "step": 360 }, { "epoch": 0.10321296559567814, "grad_norm": 0.5367030501365662, "learning_rate": 1.102215815291774e-06, "loss": 0.4246, "step": 363 }, { "epoch": 0.10406596531134489, "grad_norm": 0.5049583315849304, "learning_rate": 9.317983122552332e-07, "loss": 0.3515, "step": 366 }, { "epoch": 0.10491896502701166, "grad_norm": 0.47986966371536255, "learning_rate": 7.754359612344859e-07, "loss": 0.3667, "step": 369 }, { "epoch": 0.10577196474267841, "grad_norm": 0.5948217511177063, "learning_rate": 6.332200734393057e-07, "loss": 0.4115, "step": 372 }, { "epoch": 0.1063406312197896, "eval_loss": 0.41257256269454956, "eval_runtime": 168.1203, "eval_samples_per_second": 35.237, "eval_steps_per_second": 4.408, "step": 374 }, { "epoch": 0.10662496445834518, "grad_norm": 0.46795177459716797, "learning_rate": 5.052336989433082e-07, "loss": 0.4216, "step": 375 }, { "epoch": 0.10747796417401194, "grad_norm": 0.560074508190155, "learning_rate": 3.915515781850565e-07, "loss": 0.4089, "step": 378 }, { "epoch": 0.10833096388967871, "grad_norm": 0.5042891502380371, "learning_rate": 2.922400983217416e-07, "loss": 0.4484, "step": 381 }, { "epoch": 0.10918396360534546, "grad_norm": 0.4858133792877197, "learning_rate": 2.0735725446094923e-07, "loss": 0.4354, "step": 384 }, { "epoch": 0.11003696332101223, "grad_norm": 0.5573092103004456, "learning_rate": 1.3695261579316777e-07, "loss": 0.3903, "step": 387 }, { "epoch": 0.11088996303667899, "grad_norm": 0.4502072334289551, "learning_rate": 8.106729664475176e-08, "loss": 0.416, "step": 390 }, { "epoch": 0.11174296275234574, "grad_norm": 0.5383118987083435, "learning_rate": 3.9733932468333234e-08, "loss": 0.4726, "step": 393 }, { "epoch": 0.11259596246801251, "grad_norm": 0.6252465844154358, "learning_rate": 1.297666078462767e-08, "loss": 0.4435, "step": 396 }, { "epoch": 0.11344896218367927, "grad_norm": 0.5062450170516968, "learning_rate": 8.111070868010995e-10, "loss": 0.373, "step": 399 } ], "logging_steps": 3, "max_steps": 400, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 34, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.280941246906368e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }