hp_ablations_gemma_bsz1024 / trainer_state.json
sedrickkeh's picture
End of training
20025c5 verified
raw
history blame
12 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9946403385049365,
"eval_steps": 500,
"global_step": 663,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.045133991537376586,
"grad_norm": 1.0502111381705257,
"learning_rate": 5e-06,
"loss": 0.7438,
"step": 10
},
{
"epoch": 0.09026798307475317,
"grad_norm": 0.7055928208506984,
"learning_rate": 5e-06,
"loss": 0.6709,
"step": 20
},
{
"epoch": 0.13540197461212977,
"grad_norm": 0.5033134209699918,
"learning_rate": 5e-06,
"loss": 0.6463,
"step": 30
},
{
"epoch": 0.18053596614950634,
"grad_norm": 0.618156041678536,
"learning_rate": 5e-06,
"loss": 0.6389,
"step": 40
},
{
"epoch": 0.22566995768688294,
"grad_norm": 0.49695878808904403,
"learning_rate": 5e-06,
"loss": 0.6349,
"step": 50
},
{
"epoch": 0.27080394922425954,
"grad_norm": 0.5844890023067665,
"learning_rate": 5e-06,
"loss": 0.6247,
"step": 60
},
{
"epoch": 0.3159379407616361,
"grad_norm": 0.6341705695305015,
"learning_rate": 5e-06,
"loss": 0.6227,
"step": 70
},
{
"epoch": 0.3610719322990127,
"grad_norm": 0.9851855606862004,
"learning_rate": 5e-06,
"loss": 0.6225,
"step": 80
},
{
"epoch": 0.40620592383638926,
"grad_norm": 0.6200620013182269,
"learning_rate": 5e-06,
"loss": 0.6161,
"step": 90
},
{
"epoch": 0.4513399153737659,
"grad_norm": 0.4866582707983752,
"learning_rate": 5e-06,
"loss": 0.6118,
"step": 100
},
{
"epoch": 0.49647390691114246,
"grad_norm": 0.4846006073335934,
"learning_rate": 5e-06,
"loss": 0.6071,
"step": 110
},
{
"epoch": 0.5416078984485191,
"grad_norm": 0.4640299470151304,
"learning_rate": 5e-06,
"loss": 0.6108,
"step": 120
},
{
"epoch": 0.5867418899858956,
"grad_norm": 0.5199888815602103,
"learning_rate": 5e-06,
"loss": 0.6056,
"step": 130
},
{
"epoch": 0.6318758815232722,
"grad_norm": 0.5702031853447341,
"learning_rate": 5e-06,
"loss": 0.6056,
"step": 140
},
{
"epoch": 0.6770098730606487,
"grad_norm": 0.8001760897422481,
"learning_rate": 5e-06,
"loss": 0.6054,
"step": 150
},
{
"epoch": 0.7221438645980254,
"grad_norm": 0.5587666551476193,
"learning_rate": 5e-06,
"loss": 0.6031,
"step": 160
},
{
"epoch": 0.767277856135402,
"grad_norm": 0.43674759924584844,
"learning_rate": 5e-06,
"loss": 0.5996,
"step": 170
},
{
"epoch": 0.8124118476727785,
"grad_norm": 0.5559456367161821,
"learning_rate": 5e-06,
"loss": 0.6003,
"step": 180
},
{
"epoch": 0.8575458392101551,
"grad_norm": 0.44306696302984344,
"learning_rate": 5e-06,
"loss": 0.6049,
"step": 190
},
{
"epoch": 0.9026798307475318,
"grad_norm": 0.5092942073450014,
"learning_rate": 5e-06,
"loss": 0.5985,
"step": 200
},
{
"epoch": 0.9478138222849083,
"grad_norm": 0.47841352538603515,
"learning_rate": 5e-06,
"loss": 0.5975,
"step": 210
},
{
"epoch": 0.9929478138222849,
"grad_norm": 0.757057117339616,
"learning_rate": 5e-06,
"loss": 0.5944,
"step": 220
},
{
"epoch": 0.9974612129760225,
"eval_loss": 0.5952667593955994,
"eval_runtime": 354.3828,
"eval_samples_per_second": 33.695,
"eval_steps_per_second": 0.528,
"step": 221
},
{
"epoch": 1.039210155148096,
"grad_norm": 0.6288422422717695,
"learning_rate": 5e-06,
"loss": 0.5957,
"step": 230
},
{
"epoch": 1.0843441466854724,
"grad_norm": 0.5368840934195099,
"learning_rate": 5e-06,
"loss": 0.5546,
"step": 240
},
{
"epoch": 1.1294781382228491,
"grad_norm": 0.5562178733269311,
"learning_rate": 5e-06,
"loss": 0.5576,
"step": 250
},
{
"epoch": 1.1746121297602257,
"grad_norm": 0.44566707316498066,
"learning_rate": 5e-06,
"loss": 0.5531,
"step": 260
},
{
"epoch": 1.2197461212976022,
"grad_norm": 0.5774903613750582,
"learning_rate": 5e-06,
"loss": 0.5549,
"step": 270
},
{
"epoch": 1.264880112834979,
"grad_norm": 0.429832871209753,
"learning_rate": 5e-06,
"loss": 0.5572,
"step": 280
},
{
"epoch": 1.3100141043723554,
"grad_norm": 0.5244854621545229,
"learning_rate": 5e-06,
"loss": 0.5565,
"step": 290
},
{
"epoch": 1.355148095909732,
"grad_norm": 0.50009636557672,
"learning_rate": 5e-06,
"loss": 0.5587,
"step": 300
},
{
"epoch": 1.4002820874471085,
"grad_norm": 0.46468247492051845,
"learning_rate": 5e-06,
"loss": 0.5557,
"step": 310
},
{
"epoch": 1.4454160789844852,
"grad_norm": 0.6185616791546158,
"learning_rate": 5e-06,
"loss": 0.555,
"step": 320
},
{
"epoch": 1.4905500705218617,
"grad_norm": 0.5522113108694092,
"learning_rate": 5e-06,
"loss": 0.5578,
"step": 330
},
{
"epoch": 1.5356840620592385,
"grad_norm": 0.5687883176073543,
"learning_rate": 5e-06,
"loss": 0.5519,
"step": 340
},
{
"epoch": 1.580818053596615,
"grad_norm": 0.4317341125326038,
"learning_rate": 5e-06,
"loss": 0.5523,
"step": 350
},
{
"epoch": 1.6259520451339915,
"grad_norm": 0.4415467428367944,
"learning_rate": 5e-06,
"loss": 0.561,
"step": 360
},
{
"epoch": 1.671086036671368,
"grad_norm": 0.5437873630019581,
"learning_rate": 5e-06,
"loss": 0.5562,
"step": 370
},
{
"epoch": 1.7162200282087448,
"grad_norm": 0.46394331744324036,
"learning_rate": 5e-06,
"loss": 0.5552,
"step": 380
},
{
"epoch": 1.7613540197461213,
"grad_norm": 0.4475505593561043,
"learning_rate": 5e-06,
"loss": 0.5528,
"step": 390
},
{
"epoch": 1.806488011283498,
"grad_norm": 0.5295023528850353,
"learning_rate": 5e-06,
"loss": 0.5517,
"step": 400
},
{
"epoch": 1.8516220028208745,
"grad_norm": 0.5025367324019494,
"learning_rate": 5e-06,
"loss": 0.5543,
"step": 410
},
{
"epoch": 1.896755994358251,
"grad_norm": 0.46331161327776976,
"learning_rate": 5e-06,
"loss": 0.5578,
"step": 420
},
{
"epoch": 1.9418899858956276,
"grad_norm": 0.5483627384227747,
"learning_rate": 5e-06,
"loss": 0.5486,
"step": 430
},
{
"epoch": 1.987023977433004,
"grad_norm": 0.4673744718040457,
"learning_rate": 5e-06,
"loss": 0.5591,
"step": 440
},
{
"epoch": 1.9960507757404795,
"eval_loss": 0.5866958498954773,
"eval_runtime": 356.554,
"eval_samples_per_second": 33.49,
"eval_steps_per_second": 0.524,
"step": 442
},
{
"epoch": 2.0332863187588153,
"grad_norm": 0.7455137586802161,
"learning_rate": 5e-06,
"loss": 0.5574,
"step": 450
},
{
"epoch": 2.078420310296192,
"grad_norm": 0.7317716271008468,
"learning_rate": 5e-06,
"loss": 0.5064,
"step": 460
},
{
"epoch": 2.1235543018335683,
"grad_norm": 0.7140712178378773,
"learning_rate": 5e-06,
"loss": 0.5109,
"step": 470
},
{
"epoch": 2.168688293370945,
"grad_norm": 0.5219136342087098,
"learning_rate": 5e-06,
"loss": 0.5119,
"step": 480
},
{
"epoch": 2.213822284908322,
"grad_norm": 0.6942766993148072,
"learning_rate": 5e-06,
"loss": 0.5045,
"step": 490
},
{
"epoch": 2.2589562764456983,
"grad_norm": 0.4751025500905717,
"learning_rate": 5e-06,
"loss": 0.5058,
"step": 500
},
{
"epoch": 2.304090267983075,
"grad_norm": 0.5526788063335546,
"learning_rate": 5e-06,
"loss": 0.5134,
"step": 510
},
{
"epoch": 2.3492242595204513,
"grad_norm": 0.5471049911581812,
"learning_rate": 5e-06,
"loss": 0.5135,
"step": 520
},
{
"epoch": 2.394358251057828,
"grad_norm": 0.5542894654716121,
"learning_rate": 5e-06,
"loss": 0.5115,
"step": 530
},
{
"epoch": 2.4394922425952044,
"grad_norm": 0.5478659351590376,
"learning_rate": 5e-06,
"loss": 0.5102,
"step": 540
},
{
"epoch": 2.4846262341325813,
"grad_norm": 0.5494840585563637,
"learning_rate": 5e-06,
"loss": 0.5163,
"step": 550
},
{
"epoch": 2.529760225669958,
"grad_norm": 0.49237680882767904,
"learning_rate": 5e-06,
"loss": 0.5115,
"step": 560
},
{
"epoch": 2.5748942172073344,
"grad_norm": 0.47486464629996683,
"learning_rate": 5e-06,
"loss": 0.5129,
"step": 570
},
{
"epoch": 2.620028208744711,
"grad_norm": 0.5165043118582577,
"learning_rate": 5e-06,
"loss": 0.5121,
"step": 580
},
{
"epoch": 2.6651622002820874,
"grad_norm": 0.5794476922681516,
"learning_rate": 5e-06,
"loss": 0.5117,
"step": 590
},
{
"epoch": 2.710296191819464,
"grad_norm": 0.45233116281434377,
"learning_rate": 5e-06,
"loss": 0.5162,
"step": 600
},
{
"epoch": 2.7554301833568404,
"grad_norm": 0.5214525369766981,
"learning_rate": 5e-06,
"loss": 0.5132,
"step": 610
},
{
"epoch": 2.800564174894217,
"grad_norm": 0.5938791442315001,
"learning_rate": 5e-06,
"loss": 0.518,
"step": 620
},
{
"epoch": 2.845698166431594,
"grad_norm": 0.5796881742506971,
"learning_rate": 5e-06,
"loss": 0.5195,
"step": 630
},
{
"epoch": 2.8908321579689704,
"grad_norm": 0.47397608879546316,
"learning_rate": 5e-06,
"loss": 0.5135,
"step": 640
},
{
"epoch": 2.935966149506347,
"grad_norm": 0.471420002781148,
"learning_rate": 5e-06,
"loss": 0.5154,
"step": 650
},
{
"epoch": 2.9811001410437235,
"grad_norm": 0.45834788024298295,
"learning_rate": 5e-06,
"loss": 0.5153,
"step": 660
},
{
"epoch": 2.9946403385049365,
"eval_loss": 0.5917297005653381,
"eval_runtime": 356.1014,
"eval_samples_per_second": 33.533,
"eval_steps_per_second": 0.525,
"step": 663
},
{
"epoch": 2.9946403385049365,
"step": 663,
"total_flos": 5052524767739904.0,
"train_loss": 0.5641303292586612,
"train_runtime": 61418.5213,
"train_samples_per_second": 11.081,
"train_steps_per_second": 0.011
}
],
"logging_steps": 10,
"max_steps": 663,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5052524767739904.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}