|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.05201222287237501, |
|
"eval_steps": 100, |
|
"global_step": 200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.000260061114361875, |
|
"eval_loss": 3.210697650909424, |
|
"eval_runtime": 46.0394, |
|
"eval_samples_per_second": 140.684, |
|
"eval_steps_per_second": 17.594, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.001300305571809375, |
|
"grad_norm": 0.1790143996477127, |
|
"learning_rate": 5e-05, |
|
"loss": 3.2141, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.00260061114361875, |
|
"grad_norm": 0.20394925773143768, |
|
"learning_rate": 0.0001, |
|
"loss": 3.1651, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0039009167154281255, |
|
"grad_norm": 0.21119458973407745, |
|
"learning_rate": 9.995944990857849e-05, |
|
"loss": 3.1673, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0052012222872375, |
|
"grad_norm": 0.2528775930404663, |
|
"learning_rate": 9.983786540671051e-05, |
|
"loss": 3.1641, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.006501527859046876, |
|
"grad_norm": 0.2846454381942749, |
|
"learning_rate": 9.96354437049027e-05, |
|
"loss": 3.0395, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.007801833430856251, |
|
"grad_norm": 0.2666814625263214, |
|
"learning_rate": 9.935251313189564e-05, |
|
"loss": 2.9841, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.009102139002665627, |
|
"grad_norm": 0.2622530460357666, |
|
"learning_rate": 9.898953260211338e-05, |
|
"loss": 2.7829, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.010402444574475, |
|
"grad_norm": 0.25730785727500916, |
|
"learning_rate": 9.85470908713026e-05, |
|
"loss": 2.911, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.011702750146284377, |
|
"grad_norm": 0.23502734303474426, |
|
"learning_rate": 9.802590558156862e-05, |
|
"loss": 2.7205, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.013003055718093752, |
|
"grad_norm": 0.22792138159275055, |
|
"learning_rate": 9.742682209735727e-05, |
|
"loss": 2.7597, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.014303361289903128, |
|
"grad_norm": 0.22460439801216125, |
|
"learning_rate": 9.675081213427076e-05, |
|
"loss": 2.624, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.015603666861712502, |
|
"grad_norm": 0.2262052595615387, |
|
"learning_rate": 9.599897218294122e-05, |
|
"loss": 2.6747, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.016903972433521878, |
|
"grad_norm": 0.21055257320404053, |
|
"learning_rate": 9.517252173051911e-05, |
|
"loss": 2.6395, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.018204278005331254, |
|
"grad_norm": 0.20293797552585602, |
|
"learning_rate": 9.42728012826605e-05, |
|
"loss": 2.5635, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01950458357714063, |
|
"grad_norm": 0.19834651052951813, |
|
"learning_rate": 9.330127018922194e-05, |
|
"loss": 2.6046, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.02080488914895, |
|
"grad_norm": 0.2112269252538681, |
|
"learning_rate": 9.225950427718975e-05, |
|
"loss": 2.5376, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.022105194720759377, |
|
"grad_norm": 0.1964295357465744, |
|
"learning_rate": 9.114919329468282e-05, |
|
"loss": 2.4841, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.023405500292568753, |
|
"grad_norm": 0.20404809713363647, |
|
"learning_rate": 8.997213817017507e-05, |
|
"loss": 2.451, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.02470580586437813, |
|
"grad_norm": 0.2313140332698822, |
|
"learning_rate": 8.873024809138272e-05, |
|
"loss": 2.487, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.026006111436187505, |
|
"grad_norm": 0.23010624945163727, |
|
"learning_rate": 8.742553740855506e-05, |
|
"loss": 2.4241, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.026006111436187505, |
|
"eval_loss": 2.4956908226013184, |
|
"eval_runtime": 45.9142, |
|
"eval_samples_per_second": 141.067, |
|
"eval_steps_per_second": 17.642, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02730641700799688, |
|
"grad_norm": 0.2048594206571579, |
|
"learning_rate": 8.606012236719073e-05, |
|
"loss": 2.473, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.028606722579806256, |
|
"grad_norm": 0.2099340707063675, |
|
"learning_rate": 8.463621767547998e-05, |
|
"loss": 2.5045, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.02990702815161563, |
|
"grad_norm": 0.18916481733322144, |
|
"learning_rate": 8.315613291203976e-05, |
|
"loss": 2.4466, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.031207333723425004, |
|
"grad_norm": 0.20611602067947388, |
|
"learning_rate": 8.162226877976887e-05, |
|
"loss": 2.4546, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03250763929523438, |
|
"grad_norm": 0.21061871945858002, |
|
"learning_rate": 8.003711321189895e-05, |
|
"loss": 2.4797, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.033807944867043756, |
|
"grad_norm": 0.20406612753868103, |
|
"learning_rate": 7.840323733655778e-05, |
|
"loss": 2.5568, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.03510825043885313, |
|
"grad_norm": 0.21400322020053864, |
|
"learning_rate": 7.672329130639005e-05, |
|
"loss": 2.4747, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.03640855601066251, |
|
"grad_norm": 0.22233155369758606, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 2.5571, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.03770886158247188, |
|
"grad_norm": 0.21888834238052368, |
|
"learning_rate": 7.323615860218843e-05, |
|
"loss": 2.4723, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.03900916715428126, |
|
"grad_norm": 0.21492037177085876, |
|
"learning_rate": 7.143462807015271e-05, |
|
"loss": 2.4496, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04030947272609063, |
|
"grad_norm": 0.229643777012825, |
|
"learning_rate": 6.959833049300377e-05, |
|
"loss": 2.4829, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.0416097782979, |
|
"grad_norm": 0.19544768333435059, |
|
"learning_rate": 6.773024435212678e-05, |
|
"loss": 2.4211, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.04291008386970938, |
|
"grad_norm": 0.2128247320652008, |
|
"learning_rate": 6.583339969007363e-05, |
|
"loss": 2.3166, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.044210389441518755, |
|
"grad_norm": 0.2158002257347107, |
|
"learning_rate": 6.391087319582264e-05, |
|
"loss": 2.4202, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.045510695013328134, |
|
"grad_norm": 0.23509222269058228, |
|
"learning_rate": 6.19657832143779e-05, |
|
"loss": 2.381, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.046811000585137506, |
|
"grad_norm": 0.21844059228897095, |
|
"learning_rate": 6.0001284688802226e-05, |
|
"loss": 2.4878, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.048111306156946886, |
|
"grad_norm": 0.22377453744411469, |
|
"learning_rate": 5.8020564042888015e-05, |
|
"loss": 2.3867, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.04941161172875626, |
|
"grad_norm": 0.22664517164230347, |
|
"learning_rate": 5.602683401276615e-05, |
|
"loss": 2.4527, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.05071191730056563, |
|
"grad_norm": 0.21447192132472992, |
|
"learning_rate": 5.402332843583631e-05, |
|
"loss": 2.3911, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.05201222287237501, |
|
"grad_norm": 0.22999480366706848, |
|
"learning_rate": 5.201329700547076e-05, |
|
"loss": 2.4612, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.05201222287237501, |
|
"eval_loss": 2.436123847961426, |
|
"eval_runtime": 45.8777, |
|
"eval_samples_per_second": 141.18, |
|
"eval_steps_per_second": 17.656, |
|
"step": 200 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 400, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.271906598649856e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|