dimasik2987's picture
Training in progress, step 75, checkpoint
4daba71 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.058519457719691796,
"eval_steps": 25,
"global_step": 75,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0007802594362625573,
"grad_norm": 0.3376312553882599,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.4989,
"step": 1
},
{
"epoch": 0.0007802594362625573,
"eval_loss": 0.8446513414382935,
"eval_runtime": 321.1678,
"eval_samples_per_second": 3.363,
"eval_steps_per_second": 1.681,
"step": 1
},
{
"epoch": 0.0015605188725251145,
"grad_norm": 0.4070318341255188,
"learning_rate": 6.666666666666667e-05,
"loss": 0.6737,
"step": 2
},
{
"epoch": 0.002340778308787672,
"grad_norm": 0.4004059433937073,
"learning_rate": 0.0001,
"loss": 0.5249,
"step": 3
},
{
"epoch": 0.003121037745050229,
"grad_norm": 0.43589961528778076,
"learning_rate": 9.99524110790929e-05,
"loss": 0.6278,
"step": 4
},
{
"epoch": 0.0039012971813127866,
"grad_norm": 0.5050280690193176,
"learning_rate": 9.980973490458728e-05,
"loss": 0.6549,
"step": 5
},
{
"epoch": 0.004681556617575344,
"grad_norm": 0.40668612718582153,
"learning_rate": 9.957224306869053e-05,
"loss": 0.5572,
"step": 6
},
{
"epoch": 0.005461816053837901,
"grad_norm": 0.4934292733669281,
"learning_rate": 9.924038765061042e-05,
"loss": 0.6219,
"step": 7
},
{
"epoch": 0.006242075490100458,
"grad_norm": 0.5753898620605469,
"learning_rate": 9.881480035599667e-05,
"loss": 0.5658,
"step": 8
},
{
"epoch": 0.007022334926363016,
"grad_norm": 0.5027673840522766,
"learning_rate": 9.829629131445342e-05,
"loss": 0.6963,
"step": 9
},
{
"epoch": 0.007802594362625573,
"grad_norm": 0.47841089963912964,
"learning_rate": 9.768584753741134e-05,
"loss": 0.6117,
"step": 10
},
{
"epoch": 0.00858285379888813,
"grad_norm": 0.4685758948326111,
"learning_rate": 9.698463103929542e-05,
"loss": 0.561,
"step": 11
},
{
"epoch": 0.009363113235150688,
"grad_norm": 0.5550876259803772,
"learning_rate": 9.619397662556435e-05,
"loss": 0.7298,
"step": 12
},
{
"epoch": 0.010143372671413246,
"grad_norm": 0.48063692450523376,
"learning_rate": 9.53153893518325e-05,
"loss": 0.7098,
"step": 13
},
{
"epoch": 0.010923632107675801,
"grad_norm": 0.4520593583583832,
"learning_rate": 9.435054165891109e-05,
"loss": 0.6776,
"step": 14
},
{
"epoch": 0.011703891543938359,
"grad_norm": 0.5039393305778503,
"learning_rate": 9.330127018922194e-05,
"loss": 0.8103,
"step": 15
},
{
"epoch": 0.012484150980200916,
"grad_norm": 0.4985882043838501,
"learning_rate": 9.21695722906443e-05,
"loss": 0.702,
"step": 16
},
{
"epoch": 0.013264410416463474,
"grad_norm": 0.4833678603172302,
"learning_rate": 9.09576022144496e-05,
"loss": 0.7118,
"step": 17
},
{
"epoch": 0.014044669852726031,
"grad_norm": 0.45505037903785706,
"learning_rate": 8.966766701456177e-05,
"loss": 0.6413,
"step": 18
},
{
"epoch": 0.014824929288988589,
"grad_norm": 0.5279719829559326,
"learning_rate": 8.83022221559489e-05,
"loss": 0.7873,
"step": 19
},
{
"epoch": 0.015605188725251146,
"grad_norm": 0.5257707834243774,
"learning_rate": 8.68638668405062e-05,
"loss": 0.7481,
"step": 20
},
{
"epoch": 0.016385448161513702,
"grad_norm": 0.5363844633102417,
"learning_rate": 8.535533905932738e-05,
"loss": 0.7771,
"step": 21
},
{
"epoch": 0.01716570759777626,
"grad_norm": 0.5524177551269531,
"learning_rate": 8.377951038078302e-05,
"loss": 0.8244,
"step": 22
},
{
"epoch": 0.017945967034038817,
"grad_norm": 0.5710583329200745,
"learning_rate": 8.213938048432697e-05,
"loss": 0.7883,
"step": 23
},
{
"epoch": 0.018726226470301376,
"grad_norm": 0.5935850739479065,
"learning_rate": 8.043807145043604e-05,
"loss": 0.8224,
"step": 24
},
{
"epoch": 0.019506485906563932,
"grad_norm": 0.6716726422309875,
"learning_rate": 7.86788218175523e-05,
"loss": 0.9925,
"step": 25
},
{
"epoch": 0.019506485906563932,
"eval_loss": 0.7827931642532349,
"eval_runtime": 323.3131,
"eval_samples_per_second": 3.34,
"eval_steps_per_second": 1.67,
"step": 25
},
{
"epoch": 0.02028674534282649,
"grad_norm": 0.7108809351921082,
"learning_rate": 7.68649804173412e-05,
"loss": 0.8741,
"step": 26
},
{
"epoch": 0.021067004779089047,
"grad_norm": 0.7109420299530029,
"learning_rate": 7.500000000000001e-05,
"loss": 0.9355,
"step": 27
},
{
"epoch": 0.021847264215351603,
"grad_norm": 0.6290819048881531,
"learning_rate": 7.308743066175172e-05,
"loss": 0.8823,
"step": 28
},
{
"epoch": 0.022627523651614162,
"grad_norm": 0.6449434161186218,
"learning_rate": 7.113091308703498e-05,
"loss": 0.8747,
"step": 29
},
{
"epoch": 0.023407783087876718,
"grad_norm": 0.7190560102462769,
"learning_rate": 6.91341716182545e-05,
"loss": 0.9264,
"step": 30
},
{
"epoch": 0.024188042524139277,
"grad_norm": 0.6574559807777405,
"learning_rate": 6.710100716628344e-05,
"loss": 0.8356,
"step": 31
},
{
"epoch": 0.024968301960401833,
"grad_norm": 0.6734890937805176,
"learning_rate": 6.503528997521366e-05,
"loss": 0.9268,
"step": 32
},
{
"epoch": 0.025748561396664392,
"grad_norm": 0.7330193519592285,
"learning_rate": 6.294095225512603e-05,
"loss": 0.8603,
"step": 33
},
{
"epoch": 0.026528820832926948,
"grad_norm": 0.7113292813301086,
"learning_rate": 6.0821980696905146e-05,
"loss": 0.8205,
"step": 34
},
{
"epoch": 0.027309080269189507,
"grad_norm": 0.6743346452713013,
"learning_rate": 5.868240888334653e-05,
"loss": 0.8219,
"step": 35
},
{
"epoch": 0.028089339705452063,
"grad_norm": 0.6742866039276123,
"learning_rate": 5.6526309611002594e-05,
"loss": 0.8635,
"step": 36
},
{
"epoch": 0.02886959914171462,
"grad_norm": 0.7305226922035217,
"learning_rate": 5.435778713738292e-05,
"loss": 0.818,
"step": 37
},
{
"epoch": 0.029649858577977178,
"grad_norm": 0.6999145150184631,
"learning_rate": 5.218096936826681e-05,
"loss": 0.8282,
"step": 38
},
{
"epoch": 0.030430118014239733,
"grad_norm": 0.6912006139755249,
"learning_rate": 5e-05,
"loss": 0.7905,
"step": 39
},
{
"epoch": 0.031210377450502293,
"grad_norm": 0.7328338623046875,
"learning_rate": 4.781903063173321e-05,
"loss": 0.8069,
"step": 40
},
{
"epoch": 0.03199063688676485,
"grad_norm": 0.7178049087524414,
"learning_rate": 4.564221286261709e-05,
"loss": 0.7919,
"step": 41
},
{
"epoch": 0.032770896323027404,
"grad_norm": 0.7273955941200256,
"learning_rate": 4.347369038899744e-05,
"loss": 0.8151,
"step": 42
},
{
"epoch": 0.03355115575928996,
"grad_norm": 0.7298927307128906,
"learning_rate": 4.131759111665349e-05,
"loss": 0.8524,
"step": 43
},
{
"epoch": 0.03433141519555252,
"grad_norm": 0.740074872970581,
"learning_rate": 3.917801930309486e-05,
"loss": 0.8144,
"step": 44
},
{
"epoch": 0.03511167463181508,
"grad_norm": 0.7409592866897583,
"learning_rate": 3.705904774487396e-05,
"loss": 0.8217,
"step": 45
},
{
"epoch": 0.035891934068077634,
"grad_norm": 0.757175087928772,
"learning_rate": 3.4964710024786354e-05,
"loss": 0.7642,
"step": 46
},
{
"epoch": 0.03667219350434019,
"grad_norm": 0.8015655875205994,
"learning_rate": 3.289899283371657e-05,
"loss": 0.7827,
"step": 47
},
{
"epoch": 0.03745245294060275,
"grad_norm": 0.7460522651672363,
"learning_rate": 3.086582838174551e-05,
"loss": 0.7761,
"step": 48
},
{
"epoch": 0.038232712376865305,
"grad_norm": 0.8603819608688354,
"learning_rate": 2.886908691296504e-05,
"loss": 0.7828,
"step": 49
},
{
"epoch": 0.039012971813127864,
"grad_norm": 1.0592442750930786,
"learning_rate": 2.6912569338248315e-05,
"loss": 0.8512,
"step": 50
},
{
"epoch": 0.039012971813127864,
"eval_loss": 0.7553926706314087,
"eval_runtime": 323.1681,
"eval_samples_per_second": 3.342,
"eval_steps_per_second": 1.671,
"step": 50
},
{
"epoch": 0.03979323124939042,
"grad_norm": 0.2862579822540283,
"learning_rate": 2.500000000000001e-05,
"loss": 0.507,
"step": 51
},
{
"epoch": 0.04057349068565298,
"grad_norm": 0.3461783230304718,
"learning_rate": 2.3135019582658802e-05,
"loss": 0.5549,
"step": 52
},
{
"epoch": 0.041353750121915535,
"grad_norm": 0.4043627977371216,
"learning_rate": 2.132117818244771e-05,
"loss": 0.6321,
"step": 53
},
{
"epoch": 0.042134009558178094,
"grad_norm": 0.37553367018699646,
"learning_rate": 1.9561928549563968e-05,
"loss": 0.571,
"step": 54
},
{
"epoch": 0.04291426899444065,
"grad_norm": 0.36022651195526123,
"learning_rate": 1.7860619515673033e-05,
"loss": 0.5375,
"step": 55
},
{
"epoch": 0.043694528430703206,
"grad_norm": 0.4114338457584381,
"learning_rate": 1.622048961921699e-05,
"loss": 0.6081,
"step": 56
},
{
"epoch": 0.044474787866965765,
"grad_norm": 0.3871912956237793,
"learning_rate": 1.4644660940672627e-05,
"loss": 0.6099,
"step": 57
},
{
"epoch": 0.045255047303228324,
"grad_norm": 0.40150532126426697,
"learning_rate": 1.3136133159493802e-05,
"loss": 0.6333,
"step": 58
},
{
"epoch": 0.04603530673949088,
"grad_norm": 0.49420806765556335,
"learning_rate": 1.1697777844051105e-05,
"loss": 0.773,
"step": 59
},
{
"epoch": 0.046815566175753436,
"grad_norm": 0.42875587940216064,
"learning_rate": 1.0332332985438248e-05,
"loss": 0.5722,
"step": 60
},
{
"epoch": 0.047595825612015995,
"grad_norm": 0.4141090512275696,
"learning_rate": 9.042397785550405e-06,
"loss": 0.6371,
"step": 61
},
{
"epoch": 0.048376085048278554,
"grad_norm": 0.42042094469070435,
"learning_rate": 7.830427709355725e-06,
"loss": 0.6135,
"step": 62
},
{
"epoch": 0.04915634448454111,
"grad_norm": 0.4620630145072937,
"learning_rate": 6.698729810778065e-06,
"loss": 0.7216,
"step": 63
},
{
"epoch": 0.049936603920803666,
"grad_norm": 0.47221145033836365,
"learning_rate": 5.649458341088915e-06,
"loss": 0.6438,
"step": 64
},
{
"epoch": 0.050716863357066225,
"grad_norm": 0.4742897152900696,
"learning_rate": 4.684610648167503e-06,
"loss": 0.7366,
"step": 65
},
{
"epoch": 0.051497122793328784,
"grad_norm": 0.4719804525375366,
"learning_rate": 3.8060233744356633e-06,
"loss": 0.6938,
"step": 66
},
{
"epoch": 0.052277382229591336,
"grad_norm": 0.49254298210144043,
"learning_rate": 3.0153689607045845e-06,
"loss": 0.7749,
"step": 67
},
{
"epoch": 0.053057641665853895,
"grad_norm": 0.521112859249115,
"learning_rate": 2.314152462588659e-06,
"loss": 0.7219,
"step": 68
},
{
"epoch": 0.053837901102116455,
"grad_norm": 0.5138411521911621,
"learning_rate": 1.70370868554659e-06,
"loss": 0.832,
"step": 69
},
{
"epoch": 0.054618160538379014,
"grad_norm": 0.5415810942649841,
"learning_rate": 1.1851996440033319e-06,
"loss": 0.8144,
"step": 70
},
{
"epoch": 0.055398419974641566,
"grad_norm": 0.6076228022575378,
"learning_rate": 7.596123493895991e-07,
"loss": 0.9642,
"step": 71
},
{
"epoch": 0.056178679410904125,
"grad_norm": 0.5630006194114685,
"learning_rate": 4.277569313094809e-07,
"loss": 0.7559,
"step": 72
},
{
"epoch": 0.056958938847166685,
"grad_norm": 0.608703076839447,
"learning_rate": 1.9026509541272275e-07,
"loss": 0.8729,
"step": 73
},
{
"epoch": 0.05773919828342924,
"grad_norm": 0.569412887096405,
"learning_rate": 4.7588920907110094e-08,
"loss": 0.7729,
"step": 74
},
{
"epoch": 0.058519457719691796,
"grad_norm": 0.6440850496292114,
"learning_rate": 0.0,
"loss": 0.8444,
"step": 75
},
{
"epoch": 0.058519457719691796,
"eval_loss": 0.7578212022781372,
"eval_runtime": 323.0666,
"eval_samples_per_second": 3.343,
"eval_steps_per_second": 1.671,
"step": 75
}
],
"logging_steps": 1,
"max_steps": 75,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 25,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.163503502056489e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}