{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9990732159406858, "eval_steps": 500, "global_step": 539, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0018535681186283596, "grad_norm": 26.35410018433926, "learning_rate": 5.555555555555555e-06, "loss": 1.6809, "step": 1 }, { "epoch": 0.009267840593141797, "grad_norm": 39.94533926724377, "learning_rate": 2.7777777777777772e-05, "loss": 1.343, "step": 5 }, { "epoch": 0.018535681186283594, "grad_norm": 4.5543620999927485, "learning_rate": 5.5555555555555545e-05, "loss": 0.8749, "step": 10 }, { "epoch": 0.027803521779425393, "grad_norm": 4.24495034141875, "learning_rate": 8.333333333333333e-05, "loss": 0.8704, "step": 15 }, { "epoch": 0.03707136237256719, "grad_norm": 39.42543160126464, "learning_rate": 0.00011111111111111109, "loss": 0.8733, "step": 20 }, { "epoch": 0.04633920296570899, "grad_norm": 11.962483325702362, "learning_rate": 0.0001388888888888889, "loss": 1.0553, "step": 25 }, { "epoch": 0.05560704355885079, "grad_norm": 14.195688905715766, "learning_rate": 0.00016666666666666666, "loss": 1.0623, "step": 30 }, { "epoch": 0.06487488415199258, "grad_norm": 4.164105779251046, "learning_rate": 0.00019444444444444443, "loss": 0.8718, "step": 35 }, { "epoch": 0.07414272474513438, "grad_norm": 6.991113945747281, "learning_rate": 0.00022222222222222218, "loss": 0.8619, "step": 40 }, { "epoch": 0.08341056533827618, "grad_norm": 631.4969668103296, "learning_rate": 0.00025, "loss": 1.8292, "step": 45 }, { "epoch": 0.09267840593141798, "grad_norm": 51.99035558508526, "learning_rate": 0.0002777777777777778, "loss": 1.6128, "step": 50 }, { "epoch": 0.10194624652455977, "grad_norm": 5.615188207831111, "learning_rate": 0.0002999968531502098, "loss": 1.1458, "step": 55 }, { "epoch": 0.11121408711770157, "grad_norm": 162.57564664314097, "learning_rate": 0.0002998867272706619, "loss": 1.7169, "step": 60 }, { "epoch": 0.12048192771084337, "grad_norm": 27.819505339583905, "learning_rate": 0.0002996193909122197, "loss": 2.1196, "step": 65 }, { "epoch": 0.12974976830398516, "grad_norm": 12.017979123334465, "learning_rate": 0.00029919512447380625, "loss": 1.3348, "step": 70 }, { "epoch": 0.13901760889712697, "grad_norm": 529.5307829885717, "learning_rate": 0.0002986143729523282, "loss": 1.0105, "step": 75 }, { "epoch": 0.14828544949026876, "grad_norm": 2.511805320667406, "learning_rate": 0.000297877745475935, "loss": 1.0586, "step": 80 }, { "epoch": 0.15755329008341057, "grad_norm": 4.022429995067504, "learning_rate": 0.0002969860146651276, "loss": 0.9055, "step": 85 }, { "epoch": 0.16682113067655235, "grad_norm": 512.3998660297882, "learning_rate": 0.0002959401158223867, "loss": 5.3655, "step": 90 }, { "epoch": 0.17608897126969417, "grad_norm": 131.99417407337796, "learning_rate": 0.00029474114595116896, "loss": 3.1238, "step": 95 }, { "epoch": 0.18535681186283595, "grad_norm": 19.953838142892188, "learning_rate": 0.0002933903626053024, "loss": 1.9603, "step": 100 }, { "epoch": 0.19462465245597776, "grad_norm": 7.472797510255509, "learning_rate": 0.00029188918256998564, "loss": 1.4801, "step": 105 }, { "epoch": 0.20389249304911955, "grad_norm": 9.70552960980865, "learning_rate": 0.00029023918037577635, "loss": 1.3544, "step": 110 }, { "epoch": 0.21316033364226136, "grad_norm": 17.11916105312361, "learning_rate": 0.00028844208664712575, "loss": 1.0681, "step": 115 }, { "epoch": 0.22242817423540315, "grad_norm": 22.28855318123011, "learning_rate": 0.00028649978628719254, "loss": 1.2611, "step": 120 }, { "epoch": 0.23169601482854496, "grad_norm": 34.76579789562846, "learning_rate": 0.00028441431650084016, "loss": 1.6181, "step": 125 }, { "epoch": 0.24096385542168675, "grad_norm": 15.11803328233237, "learning_rate": 0.0002821878646578898, "loss": 1.1601, "step": 130 }, { "epoch": 0.25023169601482853, "grad_norm": 10.594239787348334, "learning_rate": 0.0002798227659988717, "loss": 1.0309, "step": 135 }, { "epoch": 0.2594995366079703, "grad_norm": 3.881883182518721, "learning_rate": 0.00027732150118568017, "loss": 1.5651, "step": 140 }, { "epoch": 0.26876737720111216, "grad_norm": 3.036719624432077, "learning_rate": 0.00027468669369970207, "loss": 1.1445, "step": 145 }, { "epoch": 0.27803521779425394, "grad_norm": 14.352629655607771, "learning_rate": 0.00027192110709014697, "loss": 0.9305, "step": 150 }, { "epoch": 0.2873030583873957, "grad_norm": 2.933168446331413, "learning_rate": 0.0002690276420754655, "loss": 0.9324, "step": 155 }, { "epoch": 0.2965708989805375, "grad_norm": 4.661928504358525, "learning_rate": 0.00026600933350089654, "loss": 0.9491, "step": 160 }, { "epoch": 0.30583873957367935, "grad_norm": 5.782107803222577, "learning_rate": 0.0002628693471553335, "loss": 0.8689, "step": 165 }, { "epoch": 0.31510658016682114, "grad_norm": 4.287118434199453, "learning_rate": 0.00025961097645084885, "loss": 1.112, "step": 170 }, { "epoch": 0.3243744207599629, "grad_norm": 170.8246279898043, "learning_rate": 0.0002562376389683599, "loss": 2.2669, "step": 175 }, { "epoch": 0.3336422613531047, "grad_norm": 243.40178580373365, "learning_rate": 0.00025275287287305814, "loss": 3.2917, "step": 180 }, { "epoch": 0.34291010194624655, "grad_norm": 160.38147120987801, "learning_rate": 0.00024916033320336264, "loss": 3.067, "step": 185 }, { "epoch": 0.35217794253938833, "grad_norm": 30.070071532273786, "learning_rate": 0.0002454637880372892, "loss": 3.1301, "step": 190 }, { "epoch": 0.3614457831325301, "grad_norm": 48.47525499191394, "learning_rate": 0.0002416671145402575, "loss": 2.6178, "step": 195 }, { "epoch": 0.3707136237256719, "grad_norm": 48.37664262960068, "learning_rate": 0.00023777429489847934, "loss": 1.441, "step": 200 }, { "epoch": 0.3799814643188137, "grad_norm": 3.8665655676681245, "learning_rate": 0.0002337894121421954, "loss": 1.2978, "step": 205 }, { "epoch": 0.38924930491195553, "grad_norm": 2.9273559971935055, "learning_rate": 0.00022971664586314054, "loss": 1.0238, "step": 210 }, { "epoch": 0.3985171455050973, "grad_norm": 20.165413444524287, "learning_rate": 0.00022556026783072895, "loss": 0.8787, "step": 215 }, { "epoch": 0.4077849860982391, "grad_norm": 2.3870488681045203, "learning_rate": 0.00022132463751155812, "loss": 0.9467, "step": 220 }, { "epoch": 0.4170528266913809, "grad_norm": 4.038604860534557, "learning_rate": 0.00021701419749693034, "loss": 0.9708, "step": 225 }, { "epoch": 0.4263206672845227, "grad_norm": 2.7240106071190535, "learning_rate": 0.00021263346884318777, "loss": 1.0564, "step": 230 }, { "epoch": 0.4355885078776645, "grad_norm": 1.6377519083501402, "learning_rate": 0.00020818704632974896, "loss": 0.7724, "step": 235 }, { "epoch": 0.4448563484708063, "grad_norm": 1.0433662853439323, "learning_rate": 0.00020367959363981936, "loss": 0.8052, "step": 240 }, { "epoch": 0.4541241890639481, "grad_norm": 1.5351301360333338, "learning_rate": 0.00019911583846883197, "loss": 0.8407, "step": 245 }, { "epoch": 0.4633920296570899, "grad_norm": 1.2741936551633426, "learning_rate": 0.0001945005675657475, "loss": 0.8344, "step": 250 }, { "epoch": 0.4726598702502317, "grad_norm": 2.4652187839401467, "learning_rate": 0.00018983862171241577, "loss": 0.9688, "step": 255 }, { "epoch": 0.4819277108433735, "grad_norm": 1.1552631433359035, "learning_rate": 0.00018513489064626398, "loss": 0.8647, "step": 260 }, { "epoch": 0.4911955514365153, "grad_norm": 1.113328634852908, "learning_rate": 0.00018039430793163753, "loss": 0.8514, "step": 265 }, { "epoch": 0.5004633920296571, "grad_norm": 0.995245914359209, "learning_rate": 0.00017562184578517203, "loss": 0.8845, "step": 270 }, { "epoch": 0.5097312326227988, "grad_norm": 1.5488835001137233, "learning_rate": 0.00017082250986062502, "loss": 0.7809, "step": 275 }, { "epoch": 0.5189990732159406, "grad_norm": 1.5756484423065427, "learning_rate": 0.00016600133399863594, "loss": 0.8326, "step": 280 }, { "epoch": 0.5282669138090825, "grad_norm": 206.03580929049917, "learning_rate": 0.0001611633749469231, "loss": 2.2034, "step": 285 }, { "epoch": 0.5375347544022243, "grad_norm": 4.169210184616496, "learning_rate": 0.0001563137070564528, "loss": 1.9396, "step": 290 }, { "epoch": 0.5468025949953661, "grad_norm": 7.018651154352153, "learning_rate": 0.0001514574169591466, "loss": 1.4702, "step": 295 }, { "epoch": 0.5560704355885079, "grad_norm": 1.7304196471580864, "learning_rate": 0.0001465995982327065, "loss": 0.9425, "step": 300 }, { "epoch": 0.5653382761816497, "grad_norm": 2.241775914452003, "learning_rate": 0.00014174534605815525, "loss": 0.7961, "step": 305 }, { "epoch": 0.5746061167747915, "grad_norm": 3.456242970621782, "learning_rate": 0.0001368997518756954, "loss": 0.834, "step": 310 }, { "epoch": 0.5838739573679332, "grad_norm": 5.072902047355265, "learning_rate": 0.00013206789804449116, "loss": 0.99, "step": 315 }, { "epoch": 0.593141797961075, "grad_norm": 6.231760889035276, "learning_rate": 0.0001272548525119758, "loss": 0.8985, "step": 320 }, { "epoch": 0.6024096385542169, "grad_norm": 0.8710278075890822, "learning_rate": 0.0001224656634982746, "loss": 0.7725, "step": 325 }, { "epoch": 0.6116774791473587, "grad_norm": 1.799368795567863, "learning_rate": 0.00011770535420131876, "loss": 0.7621, "step": 330 }, { "epoch": 0.6209453197405005, "grad_norm": 0.9742992413928552, "learning_rate": 0.00011297891752820484, "loss": 0.8327, "step": 335 }, { "epoch": 0.6302131603336423, "grad_norm": 0.85009149768552, "learning_rate": 0.0001082913108583245, "loss": 0.6863, "step": 340 }, { "epoch": 0.6394810009267841, "grad_norm": 0.8181844995131707, "learning_rate": 0.0001036474508437579, "loss": 0.6779, "step": 345 }, { "epoch": 0.6487488415199258, "grad_norm": 0.9546399988989202, "learning_rate": 9.905220825238491e-05, "loss": 0.6872, "step": 350 }, { "epoch": 0.6580166821130676, "grad_norm": 0.6995327936772554, "learning_rate": 9.45104028591222e-05, "loss": 0.7025, "step": 355 }, { "epoch": 0.6672845227062094, "grad_norm": 0.7195227820211112, "learning_rate": 9.002679839064463e-05, "loss": 0.6807, "step": 360 }, { "epoch": 0.6765523632993512, "grad_norm": 0.6283846451216366, "learning_rate": 8.560609752889412e-05, "loss": 0.6888, "step": 365 }, { "epoch": 0.6858202038924931, "grad_norm": 0.7394400936445648, "learning_rate": 8.125293697861548e-05, "loss": 0.6542, "step": 370 }, { "epoch": 0.6950880444856349, "grad_norm": 0.6824620929581083, "learning_rate": 7.697188260409356e-05, "loss": 0.671, "step": 375 }, { "epoch": 0.7043558850787767, "grad_norm": 0.6116075298871171, "learning_rate": 7.276742464019198e-05, "loss": 0.6729, "step": 380 }, { "epoch": 0.7136237256719185, "grad_norm": 0.6453928151826652, "learning_rate": 6.864397298271699e-05, "loss": 0.6626, "step": 385 }, { "epoch": 0.7228915662650602, "grad_norm": 0.6716078881868212, "learning_rate": 6.460585256304559e-05, "loss": 0.6851, "step": 390 }, { "epoch": 0.732159406858202, "grad_norm": 0.6801638199214254, "learning_rate": 6.065729881186982e-05, "loss": 0.6168, "step": 395 }, { "epoch": 0.7414272474513438, "grad_norm": 0.4794662188577268, "learning_rate": 5.680245321681471e-05, "loss": 0.6531, "step": 400 }, { "epoch": 0.7506950880444856, "grad_norm": 1.029961404045439, "learning_rate": 5.304535897858999e-05, "loss": 0.6295, "step": 405 }, { "epoch": 0.7599629286376274, "grad_norm": 0.5284338458030478, "learning_rate": 4.938995677023054e-05, "loss": 0.6201, "step": 410 }, { "epoch": 0.7692307692307693, "grad_norm": 0.4516339145236783, "learning_rate": 4.584008060387455e-05, "loss": 0.6083, "step": 415 }, { "epoch": 0.7784986098239111, "grad_norm": 0.5727589086366411, "learning_rate": 4.239945380941461e-05, "loss": 0.6021, "step": 420 }, { "epoch": 0.7877664504170528, "grad_norm": 0.4804742186453503, "learning_rate": 3.907168512923842e-05, "loss": 0.5933, "step": 425 }, { "epoch": 0.7970342910101946, "grad_norm": 0.45776606255626884, "learning_rate": 3.5860264933156324e-05, "loss": 0.5774, "step": 430 }, { "epoch": 0.8063021316033364, "grad_norm": 1.2302657626291176, "learning_rate": 3.276856155748584e-05, "loss": 0.5908, "step": 435 }, { "epoch": 0.8155699721964782, "grad_norm": 0.9348743887035583, "learning_rate": 2.9799817772131516e-05, "loss": 0.592, "step": 440 }, { "epoch": 0.82483781278962, "grad_norm": 0.4027186043623421, "learning_rate": 2.6957147379367217e-05, "loss": 0.5798, "step": 445 }, { "epoch": 0.8341056533827618, "grad_norm": 0.5240634600695104, "learning_rate": 2.4243531947887802e-05, "loss": 0.5805, "step": 450 }, { "epoch": 0.8433734939759037, "grad_norm": 0.5416112282286268, "learning_rate": 2.1661817685554833e-05, "loss": 0.6067, "step": 455 }, { "epoch": 0.8526413345690455, "grad_norm": 0.47266676591358336, "learning_rate": 1.921471245411794e-05, "loss": 0.5962, "step": 460 }, { "epoch": 0.8619091751621872, "grad_norm": 0.6382596120183166, "learning_rate": 1.6904782929041693e-05, "loss": 0.5791, "step": 465 }, { "epoch": 0.871177015755329, "grad_norm": 0.49306751144827193, "learning_rate": 1.4734451907417255e-05, "loss": 0.6182, "step": 470 }, { "epoch": 0.8804448563484708, "grad_norm": 0.49635162924553927, "learning_rate": 1.2705995766783079e-05, "loss": 0.5521, "step": 475 }, { "epoch": 0.8897126969416126, "grad_norm": 0.49868563464480065, "learning_rate": 1.0821542077519169e-05, "loss": 0.5579, "step": 480 }, { "epoch": 0.8989805375347544, "grad_norm": 0.384294096915848, "learning_rate": 9.083067371319324e-06, "loss": 0.5532, "step": 485 }, { "epoch": 0.9082483781278962, "grad_norm": 0.6584770004045724, "learning_rate": 7.492395068082619e-06, "loss": 0.544, "step": 490 }, { "epoch": 0.917516218721038, "grad_norm": 0.5198655418147264, "learning_rate": 6.051193563397599e-06, "loss": 0.555, "step": 495 }, { "epoch": 0.9267840593141798, "grad_norm": 0.46652880015863485, "learning_rate": 4.760974478625634e-06, "loss": 0.5437, "step": 500 }, { "epoch": 0.9360518999073216, "grad_norm": 1.0569529661255628, "learning_rate": 3.623091075418977e-06, "loss": 0.5573, "step": 505 }, { "epoch": 0.9453197405004634, "grad_norm": 0.4373874071284105, "learning_rate": 2.638736836336158e-06, "loss": 0.5312, "step": 510 }, { "epoch": 0.9545875810936052, "grad_norm": 0.46826046078794514, "learning_rate": 1.8089442130434061e-06, "loss": 0.5648, "step": 515 }, { "epoch": 0.963855421686747, "grad_norm": 0.40638149868456475, "learning_rate": 1.1345835434156736e-06, "loss": 0.5417, "step": 520 }, { "epoch": 0.9731232622798888, "grad_norm": 0.4274920327067933, "learning_rate": 6.163621386722218e-07, "loss": 0.5528, "step": 525 }, { "epoch": 0.9823911028730306, "grad_norm": 0.41954954402917344, "learning_rate": 2.5482354150493935e-07, "loss": 0.549, "step": 530 }, { "epoch": 0.9916589434661723, "grad_norm": 0.46407841167410513, "learning_rate": 5.0346955976976467e-08, "loss": 0.5499, "step": 535 }, { "epoch": 0.9990732159406858, "eval_loss": 2.2535195350646973, "eval_runtime": 2.3617, "eval_samples_per_second": 1.694, "eval_steps_per_second": 0.423, "step": 539 }, { "epoch": 0.9990732159406858, "step": 539, "total_flos": 28187736145920.0, "train_loss": 1.0674916249259283, "train_runtime": 10422.8572, "train_samples_per_second": 1.655, "train_steps_per_second": 0.052 } ], "logging_steps": 5, "max_steps": 539, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 28187736145920.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }