vdos's picture
Training in progress, step 50, checkpoint
f756030 verified
raw
history blame
10.3 kB
{
"best_metric": 1.6356873512268066,
"best_model_checkpoint": "miner_id_24/checkpoint-50",
"epoch": 0.03311806590495115,
"eval_steps": 25,
"global_step": 50,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.000662361318099023,
"grad_norm": 40.82220458984375,
"learning_rate": 5e-05,
"loss": 27.5824,
"step": 1
},
{
"epoch": 0.000662361318099023,
"eval_loss": 2.2696444988250732,
"eval_runtime": 432.347,
"eval_samples_per_second": 23.525,
"eval_steps_per_second": 2.942,
"step": 1
},
{
"epoch": 0.001324722636198046,
"grad_norm": 36.23447799682617,
"learning_rate": 0.0001,
"loss": 27.0301,
"step": 2
},
{
"epoch": 0.001987083954297069,
"grad_norm": 41.014190673828125,
"learning_rate": 9.989294616193017e-05,
"loss": 29.9936,
"step": 3
},
{
"epoch": 0.002649445272396092,
"grad_norm": 36.32714080810547,
"learning_rate": 9.957224306869053e-05,
"loss": 27.1201,
"step": 4
},
{
"epoch": 0.003311806590495115,
"grad_norm": 36.35200500488281,
"learning_rate": 9.903926402016153e-05,
"loss": 30.1767,
"step": 5
},
{
"epoch": 0.003974167908594138,
"grad_norm": 34.66734313964844,
"learning_rate": 9.829629131445342e-05,
"loss": 32.16,
"step": 6
},
{
"epoch": 0.004636529226693161,
"grad_norm": 27.75533103942871,
"learning_rate": 9.73465064747553e-05,
"loss": 28.1419,
"step": 7
},
{
"epoch": 0.005298890544792184,
"grad_norm": 39.82036209106445,
"learning_rate": 9.619397662556435e-05,
"loss": 31.4935,
"step": 8
},
{
"epoch": 0.005961251862891207,
"grad_norm": 32.18791198730469,
"learning_rate": 9.484363707663442e-05,
"loss": 29.9285,
"step": 9
},
{
"epoch": 0.00662361318099023,
"grad_norm": 33.4499626159668,
"learning_rate": 9.330127018922194e-05,
"loss": 29.7212,
"step": 10
},
{
"epoch": 0.007285974499089253,
"grad_norm": 38.62579345703125,
"learning_rate": 9.157348061512727e-05,
"loss": 33.2777,
"step": 11
},
{
"epoch": 0.007948335817188276,
"grad_norm": 36.961639404296875,
"learning_rate": 8.966766701456177e-05,
"loss": 31.1595,
"step": 12
},
{
"epoch": 0.0086106971352873,
"grad_norm": 37.11787414550781,
"learning_rate": 8.759199037394887e-05,
"loss": 30.9798,
"step": 13
},
{
"epoch": 0.009273058453386322,
"grad_norm": 35.7690315246582,
"learning_rate": 8.535533905932738e-05,
"loss": 26.4066,
"step": 14
},
{
"epoch": 0.009935419771485345,
"grad_norm": 25.68424415588379,
"learning_rate": 8.296729075500344e-05,
"loss": 26.8711,
"step": 15
},
{
"epoch": 0.010597781089584368,
"grad_norm": 20.35919189453125,
"learning_rate": 8.043807145043604e-05,
"loss": 24.8348,
"step": 16
},
{
"epoch": 0.011260142407683391,
"grad_norm": 20.65962028503418,
"learning_rate": 7.777851165098012e-05,
"loss": 24.6985,
"step": 17
},
{
"epoch": 0.011922503725782414,
"grad_norm": 20.867441177368164,
"learning_rate": 7.500000000000001e-05,
"loss": 26.0421,
"step": 18
},
{
"epoch": 0.012584865043881437,
"grad_norm": 23.38364028930664,
"learning_rate": 7.211443451095007e-05,
"loss": 25.5464,
"step": 19
},
{
"epoch": 0.01324722636198046,
"grad_norm": 24.096189498901367,
"learning_rate": 6.91341716182545e-05,
"loss": 26.3065,
"step": 20
},
{
"epoch": 0.013909587680079483,
"grad_norm": 25.358400344848633,
"learning_rate": 6.607197326515808e-05,
"loss": 27.5427,
"step": 21
},
{
"epoch": 0.014571948998178506,
"grad_norm": 25.04808235168457,
"learning_rate": 6.294095225512603e-05,
"loss": 27.4076,
"step": 22
},
{
"epoch": 0.01523431031627753,
"grad_norm": 29.721607208251953,
"learning_rate": 5.9754516100806423e-05,
"loss": 27.733,
"step": 23
},
{
"epoch": 0.015896671634376552,
"grad_norm": 34.963172912597656,
"learning_rate": 5.6526309611002594e-05,
"loss": 29.4103,
"step": 24
},
{
"epoch": 0.016559032952475575,
"grad_norm": 36.04214096069336,
"learning_rate": 5.327015646150716e-05,
"loss": 29.9496,
"step": 25
},
{
"epoch": 0.016559032952475575,
"eval_loss": 1.6913299560546875,
"eval_runtime": 435.6417,
"eval_samples_per_second": 23.347,
"eval_steps_per_second": 2.92,
"step": 25
},
{
"epoch": 0.0172213942705746,
"grad_norm": 23.54070472717285,
"learning_rate": 5e-05,
"loss": 22.5009,
"step": 26
},
{
"epoch": 0.01788375558867362,
"grad_norm": 20.027286529541016,
"learning_rate": 4.6729843538492847e-05,
"loss": 23.1153,
"step": 27
},
{
"epoch": 0.018546116906772644,
"grad_norm": 18.963481903076172,
"learning_rate": 4.347369038899744e-05,
"loss": 23.7903,
"step": 28
},
{
"epoch": 0.019208478224871667,
"grad_norm": 26.33202362060547,
"learning_rate": 4.0245483899193595e-05,
"loss": 27.938,
"step": 29
},
{
"epoch": 0.01987083954297069,
"grad_norm": 21.261442184448242,
"learning_rate": 3.705904774487396e-05,
"loss": 25.7281,
"step": 30
},
{
"epoch": 0.020533200861069714,
"grad_norm": 20.291833877563477,
"learning_rate": 3.392802673484193e-05,
"loss": 24.2849,
"step": 31
},
{
"epoch": 0.021195562179168737,
"grad_norm": 22.858013153076172,
"learning_rate": 3.086582838174551e-05,
"loss": 25.7152,
"step": 32
},
{
"epoch": 0.02185792349726776,
"grad_norm": 23.46312713623047,
"learning_rate": 2.7885565489049946e-05,
"loss": 26.8651,
"step": 33
},
{
"epoch": 0.022520284815366783,
"grad_norm": 23.533655166625977,
"learning_rate": 2.500000000000001e-05,
"loss": 26.9475,
"step": 34
},
{
"epoch": 0.023182646133465806,
"grad_norm": 23.187999725341797,
"learning_rate": 2.2221488349019903e-05,
"loss": 25.6272,
"step": 35
},
{
"epoch": 0.02384500745156483,
"grad_norm": 26.695240020751953,
"learning_rate": 1.9561928549563968e-05,
"loss": 28.5214,
"step": 36
},
{
"epoch": 0.02450736876966385,
"grad_norm": 27.083887100219727,
"learning_rate": 1.703270924499656e-05,
"loss": 31.241,
"step": 37
},
{
"epoch": 0.025169730087762875,
"grad_norm": 33.784095764160156,
"learning_rate": 1.4644660940672627e-05,
"loss": 27.0767,
"step": 38
},
{
"epoch": 0.025832091405861898,
"grad_norm": 17.89984130859375,
"learning_rate": 1.2408009626051137e-05,
"loss": 23.7762,
"step": 39
},
{
"epoch": 0.02649445272396092,
"grad_norm": 25.040252685546875,
"learning_rate": 1.0332332985438248e-05,
"loss": 26.2669,
"step": 40
},
{
"epoch": 0.027156814042059944,
"grad_norm": 17.308666229248047,
"learning_rate": 8.426519384872733e-06,
"loss": 22.8648,
"step": 41
},
{
"epoch": 0.027819175360158967,
"grad_norm": 17.806684494018555,
"learning_rate": 6.698729810778065e-06,
"loss": 25.445,
"step": 42
},
{
"epoch": 0.02848153667825799,
"grad_norm": 18.2435359954834,
"learning_rate": 5.156362923365588e-06,
"loss": 25.0637,
"step": 43
},
{
"epoch": 0.029143897996357013,
"grad_norm": 22.409690856933594,
"learning_rate": 3.8060233744356633e-06,
"loss": 28.0825,
"step": 44
},
{
"epoch": 0.029806259314456036,
"grad_norm": 18.422311782836914,
"learning_rate": 2.653493525244721e-06,
"loss": 26.1024,
"step": 45
},
{
"epoch": 0.03046862063255506,
"grad_norm": 21.495967864990234,
"learning_rate": 1.70370868554659e-06,
"loss": 25.523,
"step": 46
},
{
"epoch": 0.031130981950654082,
"grad_norm": 22.34445571899414,
"learning_rate": 9.607359798384785e-07,
"loss": 25.1083,
"step": 47
},
{
"epoch": 0.031793343268753105,
"grad_norm": 24.14988136291504,
"learning_rate": 4.277569313094809e-07,
"loss": 26.1674,
"step": 48
},
{
"epoch": 0.03245570458685213,
"grad_norm": 24.39285659790039,
"learning_rate": 1.0705383806982606e-07,
"loss": 27.8474,
"step": 49
},
{
"epoch": 0.03311806590495115,
"grad_norm": 28.339920043945312,
"learning_rate": 0.0,
"loss": 28.8547,
"step": 50
},
{
"epoch": 0.03311806590495115,
"eval_loss": 1.6356873512268066,
"eval_runtime": 431.7917,
"eval_samples_per_second": 23.555,
"eval_steps_per_second": 2.946,
"step": 50
}
],
"logging_steps": 1,
"max_steps": 50,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 25,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 1,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.435490789261312e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}