{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.05116398055768739, "eval_steps": 9, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005116398055768739, "grad_norm": 1.6488349437713623, "learning_rate": 1e-05, "loss": 2.1352, "step": 1 }, { "epoch": 0.0005116398055768739, "eval_loss": 1.228289008140564, "eval_runtime": 274.3879, "eval_samples_per_second": 5.999, "eval_steps_per_second": 0.751, "step": 1 }, { "epoch": 0.0010232796111537478, "grad_norm": 1.8014962673187256, "learning_rate": 2e-05, "loss": 2.439, "step": 2 }, { "epoch": 0.0015349194167306216, "grad_norm": 1.226792573928833, "learning_rate": 3e-05, "loss": 2.2876, "step": 3 }, { "epoch": 0.0020465592223074956, "grad_norm": 1.5848373174667358, "learning_rate": 4e-05, "loss": 2.1205, "step": 4 }, { "epoch": 0.0025581990278843694, "grad_norm": 3.6875414848327637, "learning_rate": 5e-05, "loss": 2.973, "step": 5 }, { "epoch": 0.0030698388334612432, "grad_norm": 1.861879587173462, "learning_rate": 6e-05, "loss": 2.2063, "step": 6 }, { "epoch": 0.003581478639038117, "grad_norm": 2.6092193126678467, "learning_rate": 7e-05, "loss": 2.5909, "step": 7 }, { "epoch": 0.004093118444614991, "grad_norm": 2.884692668914795, "learning_rate": 8e-05, "loss": 2.3501, "step": 8 }, { "epoch": 0.004604758250191865, "grad_norm": 1.9096957445144653, "learning_rate": 9e-05, "loss": 1.7905, "step": 9 }, { "epoch": 0.004604758250191865, "eval_loss": 1.1009039878845215, "eval_runtime": 274.3445, "eval_samples_per_second": 6.0, "eval_steps_per_second": 0.751, "step": 9 }, { "epoch": 0.005116398055768739, "grad_norm": 2.686815023422241, "learning_rate": 0.0001, "loss": 2.0176, "step": 10 }, { "epoch": 0.005628037861345613, "grad_norm": 1.8589001893997192, "learning_rate": 9.99695413509548e-05, "loss": 1.8315, "step": 11 }, { "epoch": 0.0061396776669224865, "grad_norm": 1.6375495195388794, "learning_rate": 9.987820251299122e-05, "loss": 1.8289, "step": 12 }, { "epoch": 0.00665131747249936, "grad_norm": 1.9399513006210327, "learning_rate": 9.972609476841367e-05, "loss": 1.5838, "step": 13 }, { "epoch": 0.007162957278076234, "grad_norm": 2.0972728729248047, "learning_rate": 9.951340343707852e-05, "loss": 1.514, "step": 14 }, { "epoch": 0.007674597083653108, "grad_norm": 2.4550724029541016, "learning_rate": 9.924038765061042e-05, "loss": 1.8727, "step": 15 }, { "epoch": 0.008186236889229983, "grad_norm": 2.246565580368042, "learning_rate": 9.890738003669029e-05, "loss": 1.5098, "step": 16 }, { "epoch": 0.008697876694806856, "grad_norm": 2.1880674362182617, "learning_rate": 9.851478631379982e-05, "loss": 1.5074, "step": 17 }, { "epoch": 0.00920951650038373, "grad_norm": 2.00897216796875, "learning_rate": 9.806308479691595e-05, "loss": 1.7004, "step": 18 }, { "epoch": 0.00920951650038373, "eval_loss": 0.8399195671081543, "eval_runtime": 274.0678, "eval_samples_per_second": 6.006, "eval_steps_per_second": 0.752, "step": 18 }, { "epoch": 0.009721156305960604, "grad_norm": 1.782171368598938, "learning_rate": 9.755282581475769e-05, "loss": 1.3817, "step": 19 }, { "epoch": 0.010232796111537478, "grad_norm": 1.8624811172485352, "learning_rate": 9.698463103929542e-05, "loss": 1.4014, "step": 20 }, { "epoch": 0.010744435917114352, "grad_norm": 2.667515516281128, "learning_rate": 9.635919272833938e-05, "loss": 1.771, "step": 21 }, { "epoch": 0.011256075722691225, "grad_norm": 2.9106290340423584, "learning_rate": 9.567727288213005e-05, "loss": 1.4126, "step": 22 }, { "epoch": 0.0117677155282681, "grad_norm": 2.704364776611328, "learning_rate": 9.493970231495835e-05, "loss": 1.3825, "step": 23 }, { "epoch": 0.012279355333844973, "grad_norm": 2.78434419631958, "learning_rate": 9.414737964294636e-05, "loss": 1.5669, "step": 24 }, { "epoch": 0.012790995139421847, "grad_norm": 2.526869058609009, "learning_rate": 9.330127018922194e-05, "loss": 1.4226, "step": 25 }, { "epoch": 0.01330263494499872, "grad_norm": 1.8918830156326294, "learning_rate": 9.24024048078213e-05, "loss": 1.3571, "step": 26 }, { "epoch": 0.013814274750575594, "grad_norm": 2.5752553939819336, "learning_rate": 9.145187862775209e-05, "loss": 1.2146, "step": 27 }, { "epoch": 0.013814274750575594, "eval_loss": 0.7836452126502991, "eval_runtime": 274.2003, "eval_samples_per_second": 6.003, "eval_steps_per_second": 0.751, "step": 27 }, { "epoch": 0.014325914556152468, "grad_norm": 2.8034920692443848, "learning_rate": 9.045084971874738e-05, "loss": 1.6562, "step": 28 }, { "epoch": 0.014837554361729342, "grad_norm": 2.0168240070343018, "learning_rate": 8.940053768033609e-05, "loss": 1.3663, "step": 29 }, { "epoch": 0.015349194167306216, "grad_norm": 3.294177770614624, "learning_rate": 8.83022221559489e-05, "loss": 1.5281, "step": 30 }, { "epoch": 0.01586083397288309, "grad_norm": 2.886345624923706, "learning_rate": 8.715724127386972e-05, "loss": 1.5451, "step": 31 }, { "epoch": 0.016372473778459965, "grad_norm": 1.7247331142425537, "learning_rate": 8.596699001693255e-05, "loss": 1.0179, "step": 32 }, { "epoch": 0.01688411358403684, "grad_norm": 1.886189579963684, "learning_rate": 8.473291852294987e-05, "loss": 1.3923, "step": 33 }, { "epoch": 0.017395753389613713, "grad_norm": 2.080334424972534, "learning_rate": 8.345653031794292e-05, "loss": 1.2178, "step": 34 }, { "epoch": 0.017907393195190587, "grad_norm": 2.708214282989502, "learning_rate": 8.213938048432697e-05, "loss": 1.2374, "step": 35 }, { "epoch": 0.01841903300076746, "grad_norm": 2.5027191638946533, "learning_rate": 8.07830737662829e-05, "loss": 1.0007, "step": 36 }, { "epoch": 0.01841903300076746, "eval_loss": 0.7541875243186951, "eval_runtime": 274.1248, "eval_samples_per_second": 6.005, "eval_steps_per_second": 0.751, "step": 36 }, { "epoch": 0.018930672806344334, "grad_norm": 1.938878059387207, "learning_rate": 7.938926261462366e-05, "loss": 1.2976, "step": 37 }, { "epoch": 0.019442312611921208, "grad_norm": 2.1836838722229004, "learning_rate": 7.795964517353735e-05, "loss": 1.5049, "step": 38 }, { "epoch": 0.019953952417498082, "grad_norm": 2.365196704864502, "learning_rate": 7.649596321166024e-05, "loss": 1.2362, "step": 39 }, { "epoch": 0.020465592223074956, "grad_norm": 1.9262018203735352, "learning_rate": 7.500000000000001e-05, "loss": 1.7173, "step": 40 }, { "epoch": 0.02097723202865183, "grad_norm": 2.8925254344940186, "learning_rate": 7.347357813929454e-05, "loss": 1.3671, "step": 41 }, { "epoch": 0.021488871834228703, "grad_norm": 1.5692651271820068, "learning_rate": 7.191855733945387e-05, "loss": 1.1358, "step": 42 }, { "epoch": 0.022000511639805577, "grad_norm": 2.8360183238983154, "learning_rate": 7.033683215379002e-05, "loss": 1.3804, "step": 43 }, { "epoch": 0.02251215144538245, "grad_norm": 1.7869185209274292, "learning_rate": 6.873032967079561e-05, "loss": 1.2711, "step": 44 }, { "epoch": 0.023023791250959325, "grad_norm": 2.3221404552459717, "learning_rate": 6.710100716628344e-05, "loss": 1.7126, "step": 45 }, { "epoch": 0.023023791250959325, "eval_loss": 0.7342292070388794, "eval_runtime": 274.1224, "eval_samples_per_second": 6.005, "eval_steps_per_second": 0.751, "step": 45 }, { "epoch": 0.0235354310565362, "grad_norm": 4.136687278747559, "learning_rate": 6.545084971874738e-05, "loss": 1.7686, "step": 46 }, { "epoch": 0.024047070862113072, "grad_norm": 2.2173714637756348, "learning_rate": 6.378186779084995e-05, "loss": 1.2285, "step": 47 }, { "epoch": 0.024558710667689946, "grad_norm": 2.4053614139556885, "learning_rate": 6.209609477998338e-05, "loss": 1.4621, "step": 48 }, { "epoch": 0.02507035047326682, "grad_norm": 2.109034299850464, "learning_rate": 6.0395584540887963e-05, "loss": 1.4664, "step": 49 }, { "epoch": 0.025581990278843694, "grad_norm": 1.8373016119003296, "learning_rate": 5.868240888334653e-05, "loss": 1.2307, "step": 50 }, { "epoch": 0.026093630084420567, "grad_norm": 1.6459299325942993, "learning_rate": 5.695865504800327e-05, "loss": 1.1249, "step": 51 }, { "epoch": 0.02660526988999744, "grad_norm": 1.93509042263031, "learning_rate": 5.522642316338268e-05, "loss": 1.4258, "step": 52 }, { "epoch": 0.027116909695574315, "grad_norm": 3.230286121368408, "learning_rate": 5.348782368720626e-05, "loss": 1.4488, "step": 53 }, { "epoch": 0.02762854950115119, "grad_norm": 1.5479521751403809, "learning_rate": 5.174497483512506e-05, "loss": 1.0165, "step": 54 }, { "epoch": 0.02762854950115119, "eval_loss": 0.7164188623428345, "eval_runtime": 274.108, "eval_samples_per_second": 6.005, "eval_steps_per_second": 0.752, "step": 54 }, { "epoch": 0.028140189306728063, "grad_norm": 2.426544666290283, "learning_rate": 5e-05, "loss": 1.5535, "step": 55 }, { "epoch": 0.028651829112304936, "grad_norm": 1.6888998746871948, "learning_rate": 4.825502516487497e-05, "loss": 1.2925, "step": 56 }, { "epoch": 0.02916346891788181, "grad_norm": 1.6494107246398926, "learning_rate": 4.6512176312793736e-05, "loss": 1.0268, "step": 57 }, { "epoch": 0.029675108723458684, "grad_norm": 2.0526039600372314, "learning_rate": 4.477357683661734e-05, "loss": 1.3885, "step": 58 }, { "epoch": 0.030186748529035558, "grad_norm": 2.0388901233673096, "learning_rate": 4.3041344951996746e-05, "loss": 1.418, "step": 59 }, { "epoch": 0.03069838833461243, "grad_norm": 2.1916987895965576, "learning_rate": 4.131759111665349e-05, "loss": 1.2828, "step": 60 }, { "epoch": 0.031210028140189305, "grad_norm": 2.3835113048553467, "learning_rate": 3.960441545911204e-05, "loss": 1.5845, "step": 61 }, { "epoch": 0.03172166794576618, "grad_norm": 2.7145097255706787, "learning_rate": 3.790390522001662e-05, "loss": 1.8305, "step": 62 }, { "epoch": 0.03223330775134305, "grad_norm": 1.8284953832626343, "learning_rate": 3.6218132209150045e-05, "loss": 1.3383, "step": 63 }, { "epoch": 0.03223330775134305, "eval_loss": 0.708600640296936, "eval_runtime": 274.0602, "eval_samples_per_second": 6.006, "eval_steps_per_second": 0.752, "step": 63 }, { "epoch": 0.03274494755691993, "grad_norm": 2.4546236991882324, "learning_rate": 3.4549150281252636e-05, "loss": 1.4384, "step": 64 }, { "epoch": 0.0332565873624968, "grad_norm": 1.848671317100525, "learning_rate": 3.289899283371657e-05, "loss": 1.3614, "step": 65 }, { "epoch": 0.03376822716807368, "grad_norm": 1.7728267908096313, "learning_rate": 3.12696703292044e-05, "loss": 1.3155, "step": 66 }, { "epoch": 0.03427986697365055, "grad_norm": 2.5115482807159424, "learning_rate": 2.9663167846209998e-05, "loss": 1.5332, "step": 67 }, { "epoch": 0.034791506779227425, "grad_norm": 2.179997682571411, "learning_rate": 2.8081442660546125e-05, "loss": 1.1482, "step": 68 }, { "epoch": 0.035303146584804296, "grad_norm": 2.3714139461517334, "learning_rate": 2.6526421860705473e-05, "loss": 1.2697, "step": 69 }, { "epoch": 0.03581478639038117, "grad_norm": 2.672269344329834, "learning_rate": 2.500000000000001e-05, "loss": 1.6706, "step": 70 }, { "epoch": 0.03632642619595804, "grad_norm": 1.9255658388137817, "learning_rate": 2.350403678833976e-05, "loss": 1.1635, "step": 71 }, { "epoch": 0.03683806600153492, "grad_norm": 1.1116821765899658, "learning_rate": 2.2040354826462668e-05, "loss": 0.8958, "step": 72 }, { "epoch": 0.03683806600153492, "eval_loss": 0.7018402814865112, "eval_runtime": 274.1606, "eval_samples_per_second": 6.004, "eval_steps_per_second": 0.751, "step": 72 }, { "epoch": 0.03734970580711179, "grad_norm": 1.3924734592437744, "learning_rate": 2.061073738537635e-05, "loss": 1.0895, "step": 73 }, { "epoch": 0.03786134561268867, "grad_norm": 3.168287992477417, "learning_rate": 1.9216926233717085e-05, "loss": 1.6212, "step": 74 }, { "epoch": 0.03837298541826554, "grad_norm": 2.118021249771118, "learning_rate": 1.7860619515673033e-05, "loss": 1.2363, "step": 75 }, { "epoch": 0.038884625223842416, "grad_norm": 2.252734422683716, "learning_rate": 1.6543469682057106e-05, "loss": 0.9702, "step": 76 }, { "epoch": 0.039396265029419286, "grad_norm": 1.5854482650756836, "learning_rate": 1.526708147705013e-05, "loss": 1.4033, "step": 77 }, { "epoch": 0.039907904834996163, "grad_norm": 2.2333011627197266, "learning_rate": 1.4033009983067452e-05, "loss": 1.3619, "step": 78 }, { "epoch": 0.040419544640573034, "grad_norm": 2.2637572288513184, "learning_rate": 1.2842758726130283e-05, "loss": 1.3294, "step": 79 }, { "epoch": 0.04093118444614991, "grad_norm": 2.184448003768921, "learning_rate": 1.1697777844051105e-05, "loss": 1.3759, "step": 80 }, { "epoch": 0.04144282425172678, "grad_norm": 2.1181156635284424, "learning_rate": 1.0599462319663905e-05, "loss": 1.4547, "step": 81 }, { "epoch": 0.04144282425172678, "eval_loss": 0.6984646916389465, "eval_runtime": 274.0143, "eval_samples_per_second": 6.007, "eval_steps_per_second": 0.752, "step": 81 }, { "epoch": 0.04195446405730366, "grad_norm": 2.124769926071167, "learning_rate": 9.549150281252633e-06, "loss": 1.711, "step": 82 }, { "epoch": 0.04246610386288053, "grad_norm": 2.768644332885742, "learning_rate": 8.548121372247918e-06, "loss": 1.4417, "step": 83 }, { "epoch": 0.042977743668457406, "grad_norm": 1.6614168882369995, "learning_rate": 7.597595192178702e-06, "loss": 1.1462, "step": 84 }, { "epoch": 0.04348938347403428, "grad_norm": 1.6592438220977783, "learning_rate": 6.698729810778065e-06, "loss": 1.1821, "step": 85 }, { "epoch": 0.044001023279611154, "grad_norm": 2.4975907802581787, "learning_rate": 5.852620357053651e-06, "loss": 1.5, "step": 86 }, { "epoch": 0.044512663085188024, "grad_norm": 1.9415825605392456, "learning_rate": 5.060297685041659e-06, "loss": 1.3181, "step": 87 }, { "epoch": 0.0450243028907649, "grad_norm": 1.8416827917099, "learning_rate": 4.322727117869951e-06, "loss": 1.1157, "step": 88 }, { "epoch": 0.04553594269634178, "grad_norm": 2.9629881381988525, "learning_rate": 3.6408072716606346e-06, "loss": 1.6228, "step": 89 }, { "epoch": 0.04604758250191865, "grad_norm": 1.538602590560913, "learning_rate": 3.0153689607045845e-06, "loss": 1.2217, "step": 90 }, { "epoch": 0.04604758250191865, "eval_loss": 0.6968083381652832, "eval_runtime": 274.0049, "eval_samples_per_second": 6.007, "eval_steps_per_second": 0.752, "step": 90 }, { "epoch": 0.046559222307495526, "grad_norm": 3.1427130699157715, "learning_rate": 2.4471741852423237e-06, "loss": 1.4783, "step": 91 }, { "epoch": 0.0470708621130724, "grad_norm": 1.9339940547943115, "learning_rate": 1.9369152030840556e-06, "loss": 1.3989, "step": 92 }, { "epoch": 0.047582501918649274, "grad_norm": 1.509247899055481, "learning_rate": 1.4852136862001764e-06, "loss": 1.277, "step": 93 }, { "epoch": 0.048094141724226144, "grad_norm": 2.101160764694214, "learning_rate": 1.0926199633097157e-06, "loss": 1.129, "step": 94 }, { "epoch": 0.04860578152980302, "grad_norm": 1.8684135675430298, "learning_rate": 7.596123493895991e-07, "loss": 1.397, "step": 95 }, { "epoch": 0.04911742133537989, "grad_norm": 2.9883639812469482, "learning_rate": 4.865965629214819e-07, "loss": 1.4069, "step": 96 }, { "epoch": 0.04962906114095677, "grad_norm": 2.2074315547943115, "learning_rate": 2.7390523158633554e-07, "loss": 1.4236, "step": 97 }, { "epoch": 0.05014070094653364, "grad_norm": 2.8094723224639893, "learning_rate": 1.2179748700879012e-07, "loss": 1.4329, "step": 98 }, { "epoch": 0.05065234075211052, "grad_norm": 1.6626664400100708, "learning_rate": 3.04586490452119e-08, "loss": 1.1333, "step": 99 }, { "epoch": 0.05065234075211052, "eval_loss": 0.6963222026824951, "eval_runtime": 274.0109, "eval_samples_per_second": 6.007, "eval_steps_per_second": 0.752, "step": 99 }, { "epoch": 0.05116398055768739, "grad_norm": 2.4238545894622803, "learning_rate": 0.0, "loss": 1.3018, "step": 100 } ], "logging_steps": 1, "max_steps": 100, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.313992467283968e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }