{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.0025556710340245005, "eval_steps": 25, "global_step": 75, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 3.407561378699334e-05, "grad_norm": 2.5001614093780518, "learning_rate": 3.3333333333333335e-05, "loss": 1.825, "step": 1 }, { "epoch": 3.407561378699334e-05, "eval_loss": 5.655045509338379, "eval_runtime": 5197.3091, "eval_samples_per_second": 2.378, "eval_steps_per_second": 1.189, "step": 1 }, { "epoch": 6.815122757398668e-05, "grad_norm": 2.1226208209991455, "learning_rate": 6.666666666666667e-05, "loss": 1.6844, "step": 2 }, { "epoch": 0.00010222684136098001, "grad_norm": 2.5686793327331543, "learning_rate": 0.0001, "loss": 2.1539, "step": 3 }, { "epoch": 0.00013630245514797336, "grad_norm": 4.603909015655518, "learning_rate": 9.99571699711836e-05, "loss": 4.0091, "step": 4 }, { "epoch": 0.0001703780689349667, "grad_norm": 2.2434098720550537, "learning_rate": 9.982876141412856e-05, "loss": 1.482, "step": 5 }, { "epoch": 0.00020445368272196002, "grad_norm": 2.6793839931488037, "learning_rate": 9.961501876182148e-05, "loss": 0.8997, "step": 6 }, { "epoch": 0.00023852929650895337, "grad_norm": 3.525517463684082, "learning_rate": 9.931634888554937e-05, "loss": 1.328, "step": 7 }, { "epoch": 0.0002726049102959467, "grad_norm": 4.246636867523193, "learning_rate": 9.893332032039701e-05, "loss": 1.3833, "step": 8 }, { "epoch": 0.00030668052408294006, "grad_norm": 5.746326923370361, "learning_rate": 9.846666218300807e-05, "loss": 1.1267, "step": 9 }, { "epoch": 0.0003407561378699334, "grad_norm": 5.541106700897217, "learning_rate": 9.791726278367022e-05, "loss": 0.8371, "step": 10 }, { "epoch": 0.0003748317516569267, "grad_norm": 4.492342472076416, "learning_rate": 9.728616793536588e-05, "loss": 0.8644, "step": 11 }, { "epoch": 0.00040890736544392005, "grad_norm": 5.070585250854492, "learning_rate": 9.657457896300791e-05, "loss": 0.5652, "step": 12 }, { "epoch": 0.0004429829792309134, "grad_norm": 3.305798292160034, "learning_rate": 9.578385041664925e-05, "loss": 0.5209, "step": 13 }, { "epoch": 0.00047705859301790674, "grad_norm": 1.7791708707809448, "learning_rate": 9.491548749301997e-05, "loss": 0.1586, "step": 14 }, { "epoch": 0.0005111342068049, "grad_norm": 4.680875301361084, "learning_rate": 9.397114317029975e-05, "loss": 0.5646, "step": 15 }, { "epoch": 0.0005452098205918934, "grad_norm": 3.8708016872406006, "learning_rate": 9.295261506157986e-05, "loss": 0.431, "step": 16 }, { "epoch": 0.0005792854343788867, "grad_norm": 4.305510520935059, "learning_rate": 9.186184199300464e-05, "loss": 0.4888, "step": 17 }, { "epoch": 0.0006133610481658801, "grad_norm": 4.049569606781006, "learning_rate": 9.070090031310558e-05, "loss": 0.2277, "step": 18 }, { "epoch": 0.0006474366619528734, "grad_norm": 3.324610948562622, "learning_rate": 8.947199994035401e-05, "loss": 0.2978, "step": 19 }, { "epoch": 0.0006815122757398668, "grad_norm": 3.7074697017669678, "learning_rate": 8.817748015645558e-05, "loss": 0.3028, "step": 20 }, { "epoch": 0.0007155878895268601, "grad_norm": 1.7295836210250854, "learning_rate": 8.681980515339464e-05, "loss": 0.0916, "step": 21 }, { "epoch": 0.0007496635033138534, "grad_norm": 2.569653272628784, "learning_rate": 8.540155934270471e-05, "loss": 0.3138, "step": 22 }, { "epoch": 0.0007837391171008468, "grad_norm": 2.379580020904541, "learning_rate": 8.392544243589427e-05, "loss": 0.2004, "step": 23 }, { "epoch": 0.0008178147308878401, "grad_norm": 2.292313814163208, "learning_rate": 8.239426430539243e-05, "loss": 0.1673, "step": 24 }, { "epoch": 0.0008518903446748335, "grad_norm": 2.342252731323242, "learning_rate": 8.081093963579707e-05, "loss": 0.2094, "step": 25 }, { "epoch": 0.0008518903446748335, "eval_loss": 0.17272759974002838, "eval_runtime": 5196.4661, "eval_samples_per_second": 2.378, "eval_steps_per_second": 1.189, "step": 25 }, { "epoch": 0.0008859659584618268, "grad_norm": 3.6202399730682373, "learning_rate": 7.917848237560709e-05, "loss": 0.127, "step": 26 }, { "epoch": 0.0009200415722488202, "grad_norm": 1.361546516418457, "learning_rate": 7.75e-05, "loss": 0.0881, "step": 27 }, { "epoch": 0.0009541171860358135, "grad_norm": 3.2597994804382324, "learning_rate": 7.577868759557654e-05, "loss": 0.209, "step": 28 }, { "epoch": 0.0009881927998228069, "grad_norm": 2.3739514350891113, "learning_rate": 7.401782177833148e-05, "loss": 0.0655, "step": 29 }, { "epoch": 0.0010222684136098, "grad_norm": 0.39956721663475037, "learning_rate": 7.222075445642904e-05, "loss": 0.013, "step": 30 }, { "epoch": 0.0010563440273967935, "grad_norm": 2.310549736022949, "learning_rate": 7.03909064496551e-05, "loss": 0.0997, "step": 31 }, { "epoch": 0.0010904196411837869, "grad_norm": 3.8629372119903564, "learning_rate": 6.853176097769229e-05, "loss": 0.1503, "step": 32 }, { "epoch": 0.0011244952549707803, "grad_norm": 1.4826256036758423, "learning_rate": 6.664685702961344e-05, "loss": 0.0285, "step": 33 }, { "epoch": 0.0011585708687577734, "grad_norm": 1.613270878791809, "learning_rate": 6.473978262721463e-05, "loss": 0.0595, "step": 34 }, { "epoch": 0.0011926464825447668, "grad_norm": 0.6385797262191772, "learning_rate": 6.281416799501188e-05, "loss": 0.0196, "step": 35 }, { "epoch": 0.0012267220963317602, "grad_norm": 1.283711552619934, "learning_rate": 6.087367864990233e-05, "loss": 0.0276, "step": 36 }, { "epoch": 0.0012607977101187534, "grad_norm": 5.805143356323242, "learning_rate": 5.8922008423644624e-05, "loss": 0.1078, "step": 37 }, { "epoch": 0.0012948733239057468, "grad_norm": 0.6078380346298218, "learning_rate": 5.696287243144013e-05, "loss": 0.0143, "step": 38 }, { "epoch": 0.0013289489376927402, "grad_norm": 1.3938196897506714, "learning_rate": 5.500000000000001e-05, "loss": 0.0196, "step": 39 }, { "epoch": 0.0013630245514797336, "grad_norm": 4.788840293884277, "learning_rate": 5.303712756855988e-05, "loss": 0.1522, "step": 40 }, { "epoch": 0.0013971001652667268, "grad_norm": 7.833195686340332, "learning_rate": 5.107799157635538e-05, "loss": 0.0832, "step": 41 }, { "epoch": 0.0014311757790537202, "grad_norm": 6.889073371887207, "learning_rate": 4.912632135009769e-05, "loss": 0.1242, "step": 42 }, { "epoch": 0.0014652513928407136, "grad_norm": 6.13986349105835, "learning_rate": 4.718583200498814e-05, "loss": 0.0414, "step": 43 }, { "epoch": 0.0014993270066277068, "grad_norm": 7.624245643615723, "learning_rate": 4.526021737278538e-05, "loss": 0.1478, "step": 44 }, { "epoch": 0.0015334026204147002, "grad_norm": 0.07408539205789566, "learning_rate": 4.3353142970386564e-05, "loss": 0.0022, "step": 45 }, { "epoch": 0.0015674782342016936, "grad_norm": 1.7765361070632935, "learning_rate": 4.146823902230772e-05, "loss": 0.1055, "step": 46 }, { "epoch": 0.001601553847988687, "grad_norm": 1.6538923978805542, "learning_rate": 3.960909355034491e-05, "loss": 0.0226, "step": 47 }, { "epoch": 0.0016356294617756802, "grad_norm": 13.931591987609863, "learning_rate": 3.777924554357096e-05, "loss": 0.2841, "step": 48 }, { "epoch": 0.0016697050755626736, "grad_norm": 11.1533203125, "learning_rate": 3.598217822166854e-05, "loss": 0.1394, "step": 49 }, { "epoch": 0.001703780689349667, "grad_norm": 4.493430137634277, "learning_rate": 3.422131240442349e-05, "loss": 0.0744, "step": 50 }, { "epoch": 0.001703780689349667, "eval_loss": 0.0442812517285347, "eval_runtime": 5199.0583, "eval_samples_per_second": 2.377, "eval_steps_per_second": 1.188, "step": 50 }, { "epoch": 0.0017378563031366602, "grad_norm": 4.004736423492432, "learning_rate": 3.250000000000001e-05, "loss": 0.4785, "step": 51 }, { "epoch": 0.0017719319169236536, "grad_norm": 0.8578753471374512, "learning_rate": 3.082151762439293e-05, "loss": 0.0086, "step": 52 }, { "epoch": 0.001806007530710647, "grad_norm": 1.8137702941894531, "learning_rate": 2.9189060364202943e-05, "loss": 0.0333, "step": 53 }, { "epoch": 0.0018400831444976404, "grad_norm": 0.4614853262901306, "learning_rate": 2.760573569460757e-05, "loss": 0.0078, "step": 54 }, { "epoch": 0.0018741587582846336, "grad_norm": 0.09576446563005447, "learning_rate": 2.6074557564105727e-05, "loss": 0.0015, "step": 55 }, { "epoch": 0.001908234372071627, "grad_norm": 0.09370886534452438, "learning_rate": 2.459844065729529e-05, "loss": 0.0019, "step": 56 }, { "epoch": 0.0019423099858586204, "grad_norm": 0.4553550183773041, "learning_rate": 2.3180194846605367e-05, "loss": 0.0037, "step": 57 }, { "epoch": 0.0019763855996456138, "grad_norm": 0.11098048835992813, "learning_rate": 2.1822519843544424e-05, "loss": 0.0025, "step": 58 }, { "epoch": 0.002010461213432607, "grad_norm": 0.11278193444013596, "learning_rate": 2.0528000059645997e-05, "loss": 0.0022, "step": 59 }, { "epoch": 0.0020445368272196, "grad_norm": 1.5145128965377808, "learning_rate": 1.9299099686894423e-05, "loss": 0.042, "step": 60 }, { "epoch": 0.0020786124410065937, "grad_norm": 0.15734955668449402, "learning_rate": 1.8138158006995364e-05, "loss": 0.0031, "step": 61 }, { "epoch": 0.002112688054793587, "grad_norm": 0.16131563484668732, "learning_rate": 1.7047384938420154e-05, "loss": 0.0041, "step": 62 }, { "epoch": 0.00214676366858058, "grad_norm": 1.3162636756896973, "learning_rate": 1.602885682970026e-05, "loss": 0.017, "step": 63 }, { "epoch": 0.0021808392823675737, "grad_norm": 0.5748438239097595, "learning_rate": 1.5084512506980026e-05, "loss": 0.0109, "step": 64 }, { "epoch": 0.002214914896154567, "grad_norm": 1.6675885915756226, "learning_rate": 1.4216149583350754e-05, "loss": 0.0192, "step": 65 }, { "epoch": 0.0022489905099415605, "grad_norm": 0.22785191237926483, "learning_rate": 1.3425421036992098e-05, "loss": 0.0039, "step": 66 }, { "epoch": 0.0022830661237285537, "grad_norm": 0.2156006395816803, "learning_rate": 1.2713832064634126e-05, "loss": 0.0043, "step": 67 }, { "epoch": 0.002317141737515547, "grad_norm": 1.7005534172058105, "learning_rate": 1.2082737216329794e-05, "loss": 0.0998, "step": 68 }, { "epoch": 0.0023512173513025405, "grad_norm": 0.23563118278980255, "learning_rate": 1.1533337816991932e-05, "loss": 0.0034, "step": 69 }, { "epoch": 0.0023852929650895337, "grad_norm": 0.18278473615646362, "learning_rate": 1.1066679679603e-05, "loss": 0.0037, "step": 70 }, { "epoch": 0.002419368578876527, "grad_norm": 2.191817045211792, "learning_rate": 1.0683651114450641e-05, "loss": 0.0575, "step": 71 }, { "epoch": 0.0024534441926635205, "grad_norm": 0.29528528451919556, "learning_rate": 1.0384981238178534e-05, "loss": 0.0046, "step": 72 }, { "epoch": 0.0024875198064505137, "grad_norm": 0.05887974426150322, "learning_rate": 1.017123858587145e-05, "loss": 0.0012, "step": 73 }, { "epoch": 0.002521595420237507, "grad_norm": 0.8336066603660583, "learning_rate": 1.00428300288164e-05, "loss": 0.0193, "step": 74 }, { "epoch": 0.0025556710340245005, "grad_norm": 0.5219399333000183, "learning_rate": 1e-05, "loss": 0.0082, "step": 75 }, { "epoch": 0.0025556710340245005, "eval_loss": 0.019740475341677666, "eval_runtime": 5200.5985, "eval_samples_per_second": 2.376, "eval_steps_per_second": 1.188, "step": 75 } ], "logging_steps": 1, "max_steps": 75, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.1111982529917747e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }