|
{ |
|
"best_metric": 1.6356873512268066, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-50", |
|
"epoch": 0.03311806590495115, |
|
"eval_steps": 25, |
|
"global_step": 50, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.000662361318099023, |
|
"grad_norm": 40.82220458984375, |
|
"learning_rate": 5e-05, |
|
"loss": 27.5824, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.000662361318099023, |
|
"eval_loss": 2.2696444988250732, |
|
"eval_runtime": 432.347, |
|
"eval_samples_per_second": 23.525, |
|
"eval_steps_per_second": 2.942, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.001324722636198046, |
|
"grad_norm": 36.23447799682617, |
|
"learning_rate": 0.0001, |
|
"loss": 27.0301, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.001987083954297069, |
|
"grad_norm": 41.014190673828125, |
|
"learning_rate": 9.989294616193017e-05, |
|
"loss": 29.9936, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.002649445272396092, |
|
"grad_norm": 36.32714080810547, |
|
"learning_rate": 9.957224306869053e-05, |
|
"loss": 27.1201, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.003311806590495115, |
|
"grad_norm": 36.35200500488281, |
|
"learning_rate": 9.903926402016153e-05, |
|
"loss": 30.1767, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.003974167908594138, |
|
"grad_norm": 34.66734313964844, |
|
"learning_rate": 9.829629131445342e-05, |
|
"loss": 32.16, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.004636529226693161, |
|
"grad_norm": 27.75533103942871, |
|
"learning_rate": 9.73465064747553e-05, |
|
"loss": 28.1419, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.005298890544792184, |
|
"grad_norm": 39.82036209106445, |
|
"learning_rate": 9.619397662556435e-05, |
|
"loss": 31.4935, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.005961251862891207, |
|
"grad_norm": 32.18791198730469, |
|
"learning_rate": 9.484363707663442e-05, |
|
"loss": 29.9285, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.00662361318099023, |
|
"grad_norm": 33.4499626159668, |
|
"learning_rate": 9.330127018922194e-05, |
|
"loss": 29.7212, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.007285974499089253, |
|
"grad_norm": 38.62579345703125, |
|
"learning_rate": 9.157348061512727e-05, |
|
"loss": 33.2777, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.007948335817188276, |
|
"grad_norm": 36.961639404296875, |
|
"learning_rate": 8.966766701456177e-05, |
|
"loss": 31.1595, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0086106971352873, |
|
"grad_norm": 37.11787414550781, |
|
"learning_rate": 8.759199037394887e-05, |
|
"loss": 30.9798, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.009273058453386322, |
|
"grad_norm": 35.7690315246582, |
|
"learning_rate": 8.535533905932738e-05, |
|
"loss": 26.4066, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.009935419771485345, |
|
"grad_norm": 25.68424415588379, |
|
"learning_rate": 8.296729075500344e-05, |
|
"loss": 26.8711, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.010597781089584368, |
|
"grad_norm": 20.35919189453125, |
|
"learning_rate": 8.043807145043604e-05, |
|
"loss": 24.8348, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.011260142407683391, |
|
"grad_norm": 20.65962028503418, |
|
"learning_rate": 7.777851165098012e-05, |
|
"loss": 24.6985, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.011922503725782414, |
|
"grad_norm": 20.867441177368164, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 26.0421, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.012584865043881437, |
|
"grad_norm": 23.38364028930664, |
|
"learning_rate": 7.211443451095007e-05, |
|
"loss": 25.5464, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.01324722636198046, |
|
"grad_norm": 24.096189498901367, |
|
"learning_rate": 6.91341716182545e-05, |
|
"loss": 26.3065, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.013909587680079483, |
|
"grad_norm": 25.358400344848633, |
|
"learning_rate": 6.607197326515808e-05, |
|
"loss": 27.5427, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.014571948998178506, |
|
"grad_norm": 25.04808235168457, |
|
"learning_rate": 6.294095225512603e-05, |
|
"loss": 27.4076, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.01523431031627753, |
|
"grad_norm": 29.721607208251953, |
|
"learning_rate": 5.9754516100806423e-05, |
|
"loss": 27.733, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.015896671634376552, |
|
"grad_norm": 34.963172912597656, |
|
"learning_rate": 5.6526309611002594e-05, |
|
"loss": 29.4103, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.016559032952475575, |
|
"grad_norm": 36.04214096069336, |
|
"learning_rate": 5.327015646150716e-05, |
|
"loss": 29.9496, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.016559032952475575, |
|
"eval_loss": 1.6913299560546875, |
|
"eval_runtime": 435.6417, |
|
"eval_samples_per_second": 23.347, |
|
"eval_steps_per_second": 2.92, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0172213942705746, |
|
"grad_norm": 23.54070472717285, |
|
"learning_rate": 5e-05, |
|
"loss": 22.5009, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.01788375558867362, |
|
"grad_norm": 20.027286529541016, |
|
"learning_rate": 4.6729843538492847e-05, |
|
"loss": 23.1153, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.018546116906772644, |
|
"grad_norm": 18.963481903076172, |
|
"learning_rate": 4.347369038899744e-05, |
|
"loss": 23.7903, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.019208478224871667, |
|
"grad_norm": 26.33202362060547, |
|
"learning_rate": 4.0245483899193595e-05, |
|
"loss": 27.938, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.01987083954297069, |
|
"grad_norm": 21.261442184448242, |
|
"learning_rate": 3.705904774487396e-05, |
|
"loss": 25.7281, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.020533200861069714, |
|
"grad_norm": 20.291833877563477, |
|
"learning_rate": 3.392802673484193e-05, |
|
"loss": 24.2849, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.021195562179168737, |
|
"grad_norm": 22.858013153076172, |
|
"learning_rate": 3.086582838174551e-05, |
|
"loss": 25.7152, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.02185792349726776, |
|
"grad_norm": 23.46312713623047, |
|
"learning_rate": 2.7885565489049946e-05, |
|
"loss": 26.8651, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.022520284815366783, |
|
"grad_norm": 23.533655166625977, |
|
"learning_rate": 2.500000000000001e-05, |
|
"loss": 26.9475, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.023182646133465806, |
|
"grad_norm": 23.187999725341797, |
|
"learning_rate": 2.2221488349019903e-05, |
|
"loss": 25.6272, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.02384500745156483, |
|
"grad_norm": 26.695240020751953, |
|
"learning_rate": 1.9561928549563968e-05, |
|
"loss": 28.5214, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.02450736876966385, |
|
"grad_norm": 27.083887100219727, |
|
"learning_rate": 1.703270924499656e-05, |
|
"loss": 31.241, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.025169730087762875, |
|
"grad_norm": 33.784095764160156, |
|
"learning_rate": 1.4644660940672627e-05, |
|
"loss": 27.0767, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.025832091405861898, |
|
"grad_norm": 17.89984130859375, |
|
"learning_rate": 1.2408009626051137e-05, |
|
"loss": 23.7762, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.02649445272396092, |
|
"grad_norm": 25.040252685546875, |
|
"learning_rate": 1.0332332985438248e-05, |
|
"loss": 26.2669, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.027156814042059944, |
|
"grad_norm": 17.308666229248047, |
|
"learning_rate": 8.426519384872733e-06, |
|
"loss": 22.8648, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.027819175360158967, |
|
"grad_norm": 17.806684494018555, |
|
"learning_rate": 6.698729810778065e-06, |
|
"loss": 25.445, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.02848153667825799, |
|
"grad_norm": 18.2435359954834, |
|
"learning_rate": 5.156362923365588e-06, |
|
"loss": 25.0637, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.029143897996357013, |
|
"grad_norm": 22.409690856933594, |
|
"learning_rate": 3.8060233744356633e-06, |
|
"loss": 28.0825, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.029806259314456036, |
|
"grad_norm": 18.422311782836914, |
|
"learning_rate": 2.653493525244721e-06, |
|
"loss": 26.1024, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.03046862063255506, |
|
"grad_norm": 21.495967864990234, |
|
"learning_rate": 1.70370868554659e-06, |
|
"loss": 25.523, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.031130981950654082, |
|
"grad_norm": 22.34445571899414, |
|
"learning_rate": 9.607359798384785e-07, |
|
"loss": 25.1083, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.031793343268753105, |
|
"grad_norm": 24.14988136291504, |
|
"learning_rate": 4.277569313094809e-07, |
|
"loss": 26.1674, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.03245570458685213, |
|
"grad_norm": 24.39285659790039, |
|
"learning_rate": 1.0705383806982606e-07, |
|
"loss": 27.8474, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.03311806590495115, |
|
"grad_norm": 28.339920043945312, |
|
"learning_rate": 0.0, |
|
"loss": 28.8547, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.03311806590495115, |
|
"eval_loss": 1.6356873512268066, |
|
"eval_runtime": 431.7917, |
|
"eval_samples_per_second": 23.555, |
|
"eval_steps_per_second": 2.946, |
|
"step": 50 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 50, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 1, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8.435490789261312e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|