|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"global_step": 36816, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 5.431830526887561e-06, |
|
"loss": 1.0008, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.0863661053775122e-05, |
|
"loss": 0.7988, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.6295491580662687e-05, |
|
"loss": 0.7246, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9808052151868172e-05, |
|
"loss": 0.6939, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.9204442566547958e-05, |
|
"loss": 0.6612, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.8600832981227744e-05, |
|
"loss": 0.6373, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.7997223395907527e-05, |
|
"loss": 0.6202, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.7393613810587313e-05, |
|
"loss": 0.6067, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.67900042252671e-05, |
|
"loss": 0.5993, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.618639463994688e-05, |
|
"loss": 0.5762, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.5582785054626667e-05, |
|
"loss": 0.5743, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.4979175469306453e-05, |
|
"loss": 0.5693, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.5552455186843872, |
|
"eval_runtime": 2.5925, |
|
"eval_samples_per_second": 960.462, |
|
"eval_steps_per_second": 30.087, |
|
"step": 12272 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.437556588398624e-05, |
|
"loss": 0.5115, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.3771956298666025e-05, |
|
"loss": 0.4904, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.316834671334581e-05, |
|
"loss": 0.4926, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.2564737128025594e-05, |
|
"loss": 0.4882, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.196112754270538e-05, |
|
"loss": 0.4862, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.1357517957385164e-05, |
|
"loss": 0.4829, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.0753908372064949e-05, |
|
"loss": 0.4797, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.0150298786744735e-05, |
|
"loss": 0.4749, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.546689201424519e-06, |
|
"loss": 0.4649, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.943079616104305e-06, |
|
"loss": 0.4638, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.33947003078409e-06, |
|
"loss": 0.4629, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.735860445463874e-06, |
|
"loss": 0.4666, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.4878082573413849, |
|
"eval_runtime": 2.5611, |
|
"eval_samples_per_second": 972.256, |
|
"eval_steps_per_second": 30.456, |
|
"step": 24544 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 7.1322508601436605e-06, |
|
"loss": 0.4189, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.528641274823445e-06, |
|
"loss": 0.3709, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.92503168950323e-06, |
|
"loss": 0.3749, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 5.321422104183015e-06, |
|
"loss": 0.3729, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.7178125188628e-06, |
|
"loss": 0.3657, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.114202933542585e-06, |
|
"loss": 0.3694, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.51059334822237e-06, |
|
"loss": 0.3715, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.9069837629021554e-06, |
|
"loss": 0.3666, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.30337417758194e-06, |
|
"loss": 0.3672, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.6997645922617255e-06, |
|
"loss": 0.3717, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.0961550069415102e-06, |
|
"loss": 0.3555, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.925454216212955e-07, |
|
"loss": 0.355, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.5086637735366821, |
|
"eval_runtime": 2.5595, |
|
"eval_samples_per_second": 972.842, |
|
"eval_steps_per_second": 30.475, |
|
"step": 36816 |
|
} |
|
], |
|
"max_steps": 36816, |
|
"num_train_epochs": 3, |
|
"total_flos": 7.749426953899213e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|