llama3.1-8b-gpt4o_100k_coding-fft / trainer_state.json
chansung's picture
Model save
6f3318e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9990732159406858,
"eval_steps": 500,
"global_step": 539,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0018535681186283596,
"grad_norm": 26.35410018433926,
"learning_rate": 5.555555555555555e-06,
"loss": 1.6809,
"step": 1
},
{
"epoch": 0.009267840593141797,
"grad_norm": 39.94533926724377,
"learning_rate": 2.7777777777777772e-05,
"loss": 1.343,
"step": 5
},
{
"epoch": 0.018535681186283594,
"grad_norm": 4.5543620999927485,
"learning_rate": 5.5555555555555545e-05,
"loss": 0.8749,
"step": 10
},
{
"epoch": 0.027803521779425393,
"grad_norm": 4.24495034141875,
"learning_rate": 8.333333333333333e-05,
"loss": 0.8704,
"step": 15
},
{
"epoch": 0.03707136237256719,
"grad_norm": 39.42543160126464,
"learning_rate": 0.00011111111111111109,
"loss": 0.8733,
"step": 20
},
{
"epoch": 0.04633920296570899,
"grad_norm": 11.962483325702362,
"learning_rate": 0.0001388888888888889,
"loss": 1.0553,
"step": 25
},
{
"epoch": 0.05560704355885079,
"grad_norm": 14.195688905715766,
"learning_rate": 0.00016666666666666666,
"loss": 1.0623,
"step": 30
},
{
"epoch": 0.06487488415199258,
"grad_norm": 4.164105779251046,
"learning_rate": 0.00019444444444444443,
"loss": 0.8718,
"step": 35
},
{
"epoch": 0.07414272474513438,
"grad_norm": 6.991113945747281,
"learning_rate": 0.00022222222222222218,
"loss": 0.8619,
"step": 40
},
{
"epoch": 0.08341056533827618,
"grad_norm": 631.4969668103296,
"learning_rate": 0.00025,
"loss": 1.8292,
"step": 45
},
{
"epoch": 0.09267840593141798,
"grad_norm": 51.99035558508526,
"learning_rate": 0.0002777777777777778,
"loss": 1.6128,
"step": 50
},
{
"epoch": 0.10194624652455977,
"grad_norm": 5.615188207831111,
"learning_rate": 0.0002999968531502098,
"loss": 1.1458,
"step": 55
},
{
"epoch": 0.11121408711770157,
"grad_norm": 162.57564664314097,
"learning_rate": 0.0002998867272706619,
"loss": 1.7169,
"step": 60
},
{
"epoch": 0.12048192771084337,
"grad_norm": 27.819505339583905,
"learning_rate": 0.0002996193909122197,
"loss": 2.1196,
"step": 65
},
{
"epoch": 0.12974976830398516,
"grad_norm": 12.017979123334465,
"learning_rate": 0.00029919512447380625,
"loss": 1.3348,
"step": 70
},
{
"epoch": 0.13901760889712697,
"grad_norm": 529.5307829885717,
"learning_rate": 0.0002986143729523282,
"loss": 1.0105,
"step": 75
},
{
"epoch": 0.14828544949026876,
"grad_norm": 2.511805320667406,
"learning_rate": 0.000297877745475935,
"loss": 1.0586,
"step": 80
},
{
"epoch": 0.15755329008341057,
"grad_norm": 4.022429995067504,
"learning_rate": 0.0002969860146651276,
"loss": 0.9055,
"step": 85
},
{
"epoch": 0.16682113067655235,
"grad_norm": 512.3998660297882,
"learning_rate": 0.0002959401158223867,
"loss": 5.3655,
"step": 90
},
{
"epoch": 0.17608897126969417,
"grad_norm": 131.99417407337796,
"learning_rate": 0.00029474114595116896,
"loss": 3.1238,
"step": 95
},
{
"epoch": 0.18535681186283595,
"grad_norm": 19.953838142892188,
"learning_rate": 0.0002933903626053024,
"loss": 1.9603,
"step": 100
},
{
"epoch": 0.19462465245597776,
"grad_norm": 7.472797510255509,
"learning_rate": 0.00029188918256998564,
"loss": 1.4801,
"step": 105
},
{
"epoch": 0.20389249304911955,
"grad_norm": 9.70552960980865,
"learning_rate": 0.00029023918037577635,
"loss": 1.3544,
"step": 110
},
{
"epoch": 0.21316033364226136,
"grad_norm": 17.11916105312361,
"learning_rate": 0.00028844208664712575,
"loss": 1.0681,
"step": 115
},
{
"epoch": 0.22242817423540315,
"grad_norm": 22.28855318123011,
"learning_rate": 0.00028649978628719254,
"loss": 1.2611,
"step": 120
},
{
"epoch": 0.23169601482854496,
"grad_norm": 34.76579789562846,
"learning_rate": 0.00028441431650084016,
"loss": 1.6181,
"step": 125
},
{
"epoch": 0.24096385542168675,
"grad_norm": 15.11803328233237,
"learning_rate": 0.0002821878646578898,
"loss": 1.1601,
"step": 130
},
{
"epoch": 0.25023169601482853,
"grad_norm": 10.594239787348334,
"learning_rate": 0.0002798227659988717,
"loss": 1.0309,
"step": 135
},
{
"epoch": 0.2594995366079703,
"grad_norm": 3.881883182518721,
"learning_rate": 0.00027732150118568017,
"loss": 1.5651,
"step": 140
},
{
"epoch": 0.26876737720111216,
"grad_norm": 3.036719624432077,
"learning_rate": 0.00027468669369970207,
"loss": 1.1445,
"step": 145
},
{
"epoch": 0.27803521779425394,
"grad_norm": 14.352629655607771,
"learning_rate": 0.00027192110709014697,
"loss": 0.9305,
"step": 150
},
{
"epoch": 0.2873030583873957,
"grad_norm": 2.933168446331413,
"learning_rate": 0.0002690276420754655,
"loss": 0.9324,
"step": 155
},
{
"epoch": 0.2965708989805375,
"grad_norm": 4.661928504358525,
"learning_rate": 0.00026600933350089654,
"loss": 0.9491,
"step": 160
},
{
"epoch": 0.30583873957367935,
"grad_norm": 5.782107803222577,
"learning_rate": 0.0002628693471553335,
"loss": 0.8689,
"step": 165
},
{
"epoch": 0.31510658016682114,
"grad_norm": 4.287118434199453,
"learning_rate": 0.00025961097645084885,
"loss": 1.112,
"step": 170
},
{
"epoch": 0.3243744207599629,
"grad_norm": 170.8246279898043,
"learning_rate": 0.0002562376389683599,
"loss": 2.2669,
"step": 175
},
{
"epoch": 0.3336422613531047,
"grad_norm": 243.40178580373365,
"learning_rate": 0.00025275287287305814,
"loss": 3.2917,
"step": 180
},
{
"epoch": 0.34291010194624655,
"grad_norm": 160.38147120987801,
"learning_rate": 0.00024916033320336264,
"loss": 3.067,
"step": 185
},
{
"epoch": 0.35217794253938833,
"grad_norm": 30.070071532273786,
"learning_rate": 0.0002454637880372892,
"loss": 3.1301,
"step": 190
},
{
"epoch": 0.3614457831325301,
"grad_norm": 48.47525499191394,
"learning_rate": 0.0002416671145402575,
"loss": 2.6178,
"step": 195
},
{
"epoch": 0.3707136237256719,
"grad_norm": 48.37664262960068,
"learning_rate": 0.00023777429489847934,
"loss": 1.441,
"step": 200
},
{
"epoch": 0.3799814643188137,
"grad_norm": 3.8665655676681245,
"learning_rate": 0.0002337894121421954,
"loss": 1.2978,
"step": 205
},
{
"epoch": 0.38924930491195553,
"grad_norm": 2.9273559971935055,
"learning_rate": 0.00022971664586314054,
"loss": 1.0238,
"step": 210
},
{
"epoch": 0.3985171455050973,
"grad_norm": 20.165413444524287,
"learning_rate": 0.00022556026783072895,
"loss": 0.8787,
"step": 215
},
{
"epoch": 0.4077849860982391,
"grad_norm": 2.3870488681045203,
"learning_rate": 0.00022132463751155812,
"loss": 0.9467,
"step": 220
},
{
"epoch": 0.4170528266913809,
"grad_norm": 4.038604860534557,
"learning_rate": 0.00021701419749693034,
"loss": 0.9708,
"step": 225
},
{
"epoch": 0.4263206672845227,
"grad_norm": 2.7240106071190535,
"learning_rate": 0.00021263346884318777,
"loss": 1.0564,
"step": 230
},
{
"epoch": 0.4355885078776645,
"grad_norm": 1.6377519083501402,
"learning_rate": 0.00020818704632974896,
"loss": 0.7724,
"step": 235
},
{
"epoch": 0.4448563484708063,
"grad_norm": 1.0433662853439323,
"learning_rate": 0.00020367959363981936,
"loss": 0.8052,
"step": 240
},
{
"epoch": 0.4541241890639481,
"grad_norm": 1.5351301360333338,
"learning_rate": 0.00019911583846883197,
"loss": 0.8407,
"step": 245
},
{
"epoch": 0.4633920296570899,
"grad_norm": 1.2741936551633426,
"learning_rate": 0.0001945005675657475,
"loss": 0.8344,
"step": 250
},
{
"epoch": 0.4726598702502317,
"grad_norm": 2.4652187839401467,
"learning_rate": 0.00018983862171241577,
"loss": 0.9688,
"step": 255
},
{
"epoch": 0.4819277108433735,
"grad_norm": 1.1552631433359035,
"learning_rate": 0.00018513489064626398,
"loss": 0.8647,
"step": 260
},
{
"epoch": 0.4911955514365153,
"grad_norm": 1.113328634852908,
"learning_rate": 0.00018039430793163753,
"loss": 0.8514,
"step": 265
},
{
"epoch": 0.5004633920296571,
"grad_norm": 0.995245914359209,
"learning_rate": 0.00017562184578517203,
"loss": 0.8845,
"step": 270
},
{
"epoch": 0.5097312326227988,
"grad_norm": 1.5488835001137233,
"learning_rate": 0.00017082250986062502,
"loss": 0.7809,
"step": 275
},
{
"epoch": 0.5189990732159406,
"grad_norm": 1.5756484423065427,
"learning_rate": 0.00016600133399863594,
"loss": 0.8326,
"step": 280
},
{
"epoch": 0.5282669138090825,
"grad_norm": 206.03580929049917,
"learning_rate": 0.0001611633749469231,
"loss": 2.2034,
"step": 285
},
{
"epoch": 0.5375347544022243,
"grad_norm": 4.169210184616496,
"learning_rate": 0.0001563137070564528,
"loss": 1.9396,
"step": 290
},
{
"epoch": 0.5468025949953661,
"grad_norm": 7.018651154352153,
"learning_rate": 0.0001514574169591466,
"loss": 1.4702,
"step": 295
},
{
"epoch": 0.5560704355885079,
"grad_norm": 1.7304196471580864,
"learning_rate": 0.0001465995982327065,
"loss": 0.9425,
"step": 300
},
{
"epoch": 0.5653382761816497,
"grad_norm": 2.241775914452003,
"learning_rate": 0.00014174534605815525,
"loss": 0.7961,
"step": 305
},
{
"epoch": 0.5746061167747915,
"grad_norm": 3.456242970621782,
"learning_rate": 0.0001368997518756954,
"loss": 0.834,
"step": 310
},
{
"epoch": 0.5838739573679332,
"grad_norm": 5.072902047355265,
"learning_rate": 0.00013206789804449116,
"loss": 0.99,
"step": 315
},
{
"epoch": 0.593141797961075,
"grad_norm": 6.231760889035276,
"learning_rate": 0.0001272548525119758,
"loss": 0.8985,
"step": 320
},
{
"epoch": 0.6024096385542169,
"grad_norm": 0.8710278075890822,
"learning_rate": 0.0001224656634982746,
"loss": 0.7725,
"step": 325
},
{
"epoch": 0.6116774791473587,
"grad_norm": 1.799368795567863,
"learning_rate": 0.00011770535420131876,
"loss": 0.7621,
"step": 330
},
{
"epoch": 0.6209453197405005,
"grad_norm": 0.9742992413928552,
"learning_rate": 0.00011297891752820484,
"loss": 0.8327,
"step": 335
},
{
"epoch": 0.6302131603336423,
"grad_norm": 0.85009149768552,
"learning_rate": 0.0001082913108583245,
"loss": 0.6863,
"step": 340
},
{
"epoch": 0.6394810009267841,
"grad_norm": 0.8181844995131707,
"learning_rate": 0.0001036474508437579,
"loss": 0.6779,
"step": 345
},
{
"epoch": 0.6487488415199258,
"grad_norm": 0.9546399988989202,
"learning_rate": 9.905220825238491e-05,
"loss": 0.6872,
"step": 350
},
{
"epoch": 0.6580166821130676,
"grad_norm": 0.6995327936772554,
"learning_rate": 9.45104028591222e-05,
"loss": 0.7025,
"step": 355
},
{
"epoch": 0.6672845227062094,
"grad_norm": 0.7195227820211112,
"learning_rate": 9.002679839064463e-05,
"loss": 0.6807,
"step": 360
},
{
"epoch": 0.6765523632993512,
"grad_norm": 0.6283846451216366,
"learning_rate": 8.560609752889412e-05,
"loss": 0.6888,
"step": 365
},
{
"epoch": 0.6858202038924931,
"grad_norm": 0.7394400936445648,
"learning_rate": 8.125293697861548e-05,
"loss": 0.6542,
"step": 370
},
{
"epoch": 0.6950880444856349,
"grad_norm": 0.6824620929581083,
"learning_rate": 7.697188260409356e-05,
"loss": 0.671,
"step": 375
},
{
"epoch": 0.7043558850787767,
"grad_norm": 0.6116075298871171,
"learning_rate": 7.276742464019198e-05,
"loss": 0.6729,
"step": 380
},
{
"epoch": 0.7136237256719185,
"grad_norm": 0.6453928151826652,
"learning_rate": 6.864397298271699e-05,
"loss": 0.6626,
"step": 385
},
{
"epoch": 0.7228915662650602,
"grad_norm": 0.6716078881868212,
"learning_rate": 6.460585256304559e-05,
"loss": 0.6851,
"step": 390
},
{
"epoch": 0.732159406858202,
"grad_norm": 0.6801638199214254,
"learning_rate": 6.065729881186982e-05,
"loss": 0.6168,
"step": 395
},
{
"epoch": 0.7414272474513438,
"grad_norm": 0.4794662188577268,
"learning_rate": 5.680245321681471e-05,
"loss": 0.6531,
"step": 400
},
{
"epoch": 0.7506950880444856,
"grad_norm": 1.029961404045439,
"learning_rate": 5.304535897858999e-05,
"loss": 0.6295,
"step": 405
},
{
"epoch": 0.7599629286376274,
"grad_norm": 0.5284338458030478,
"learning_rate": 4.938995677023054e-05,
"loss": 0.6201,
"step": 410
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.4516339145236783,
"learning_rate": 4.584008060387455e-05,
"loss": 0.6083,
"step": 415
},
{
"epoch": 0.7784986098239111,
"grad_norm": 0.5727589086366411,
"learning_rate": 4.239945380941461e-05,
"loss": 0.6021,
"step": 420
},
{
"epoch": 0.7877664504170528,
"grad_norm": 0.4804742186453503,
"learning_rate": 3.907168512923842e-05,
"loss": 0.5933,
"step": 425
},
{
"epoch": 0.7970342910101946,
"grad_norm": 0.45776606255626884,
"learning_rate": 3.5860264933156324e-05,
"loss": 0.5774,
"step": 430
},
{
"epoch": 0.8063021316033364,
"grad_norm": 1.2302657626291176,
"learning_rate": 3.276856155748584e-05,
"loss": 0.5908,
"step": 435
},
{
"epoch": 0.8155699721964782,
"grad_norm": 0.9348743887035583,
"learning_rate": 2.9799817772131516e-05,
"loss": 0.592,
"step": 440
},
{
"epoch": 0.82483781278962,
"grad_norm": 0.4027186043623421,
"learning_rate": 2.6957147379367217e-05,
"loss": 0.5798,
"step": 445
},
{
"epoch": 0.8341056533827618,
"grad_norm": 0.5240634600695104,
"learning_rate": 2.4243531947887802e-05,
"loss": 0.5805,
"step": 450
},
{
"epoch": 0.8433734939759037,
"grad_norm": 0.5416112282286268,
"learning_rate": 2.1661817685554833e-05,
"loss": 0.6067,
"step": 455
},
{
"epoch": 0.8526413345690455,
"grad_norm": 0.47266676591358336,
"learning_rate": 1.921471245411794e-05,
"loss": 0.5962,
"step": 460
},
{
"epoch": 0.8619091751621872,
"grad_norm": 0.6382596120183166,
"learning_rate": 1.6904782929041693e-05,
"loss": 0.5791,
"step": 465
},
{
"epoch": 0.871177015755329,
"grad_norm": 0.49306751144827193,
"learning_rate": 1.4734451907417255e-05,
"loss": 0.6182,
"step": 470
},
{
"epoch": 0.8804448563484708,
"grad_norm": 0.49635162924553927,
"learning_rate": 1.2705995766783079e-05,
"loss": 0.5521,
"step": 475
},
{
"epoch": 0.8897126969416126,
"grad_norm": 0.49868563464480065,
"learning_rate": 1.0821542077519169e-05,
"loss": 0.5579,
"step": 480
},
{
"epoch": 0.8989805375347544,
"grad_norm": 0.384294096915848,
"learning_rate": 9.083067371319324e-06,
"loss": 0.5532,
"step": 485
},
{
"epoch": 0.9082483781278962,
"grad_norm": 0.6584770004045724,
"learning_rate": 7.492395068082619e-06,
"loss": 0.544,
"step": 490
},
{
"epoch": 0.917516218721038,
"grad_norm": 0.5198655418147264,
"learning_rate": 6.051193563397599e-06,
"loss": 0.555,
"step": 495
},
{
"epoch": 0.9267840593141798,
"grad_norm": 0.46652880015863485,
"learning_rate": 4.760974478625634e-06,
"loss": 0.5437,
"step": 500
},
{
"epoch": 0.9360518999073216,
"grad_norm": 1.0569529661255628,
"learning_rate": 3.623091075418977e-06,
"loss": 0.5573,
"step": 505
},
{
"epoch": 0.9453197405004634,
"grad_norm": 0.4373874071284105,
"learning_rate": 2.638736836336158e-06,
"loss": 0.5312,
"step": 510
},
{
"epoch": 0.9545875810936052,
"grad_norm": 0.46826046078794514,
"learning_rate": 1.8089442130434061e-06,
"loss": 0.5648,
"step": 515
},
{
"epoch": 0.963855421686747,
"grad_norm": 0.40638149868456475,
"learning_rate": 1.1345835434156736e-06,
"loss": 0.5417,
"step": 520
},
{
"epoch": 0.9731232622798888,
"grad_norm": 0.4274920327067933,
"learning_rate": 6.163621386722218e-07,
"loss": 0.5528,
"step": 525
},
{
"epoch": 0.9823911028730306,
"grad_norm": 0.41954954402917344,
"learning_rate": 2.5482354150493935e-07,
"loss": 0.549,
"step": 530
},
{
"epoch": 0.9916589434661723,
"grad_norm": 0.46407841167410513,
"learning_rate": 5.0346955976976467e-08,
"loss": 0.5499,
"step": 535
},
{
"epoch": 0.9990732159406858,
"eval_loss": 2.2535195350646973,
"eval_runtime": 2.3617,
"eval_samples_per_second": 1.694,
"eval_steps_per_second": 0.423,
"step": 539
},
{
"epoch": 0.9990732159406858,
"step": 539,
"total_flos": 28187736145920.0,
"train_loss": 1.0674916249259283,
"train_runtime": 10422.8572,
"train_samples_per_second": 1.655,
"train_steps_per_second": 0.052
}
],
"logging_steps": 5,
"max_steps": 539,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 28187736145920.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}