|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.6913319238900635, |
|
"eval_steps": 500, |
|
"global_step": 200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.042283298097251586, |
|
"grad_norm": 1.2008394002914429, |
|
"learning_rate": 4.9975392245612254e-05, |
|
"loss": 4.1473, |
|
"num_input_tokens_seen": 68384, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.08456659619450317, |
|
"grad_norm": 1.1565380096435547, |
|
"learning_rate": 4.9901617425775067e-05, |
|
"loss": 4.021, |
|
"num_input_tokens_seen": 128224, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.12684989429175475, |
|
"grad_norm": 0.972896158695221, |
|
"learning_rate": 4.9778820775100185e-05, |
|
"loss": 3.8212, |
|
"num_input_tokens_seen": 197760, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.16913319238900634, |
|
"grad_norm": 1.0641047954559326, |
|
"learning_rate": 4.9607244033573156e-05, |
|
"loss": 3.7653, |
|
"num_input_tokens_seen": 262832, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.21141649048625794, |
|
"grad_norm": 1.0994701385498047, |
|
"learning_rate": 4.93872249706591e-05, |
|
"loss": 3.7434, |
|
"num_input_tokens_seen": 333472, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.2536997885835095, |
|
"grad_norm": 1.193864107131958, |
|
"learning_rate": 4.91191967203629e-05, |
|
"loss": 3.5488, |
|
"num_input_tokens_seen": 393616, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.2959830866807611, |
|
"grad_norm": 1.0215297937393188, |
|
"learning_rate": 4.8803686928552736e-05, |
|
"loss": 3.5732, |
|
"num_input_tokens_seen": 458240, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.3382663847780127, |
|
"grad_norm": 1.312198519706726, |
|
"learning_rate": 4.84413167142257e-05, |
|
"loss": 3.7108, |
|
"num_input_tokens_seen": 524576, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.38054968287526425, |
|
"grad_norm": 1.3579617738723755, |
|
"learning_rate": 4.803279944676032e-05, |
|
"loss": 3.6871, |
|
"num_input_tokens_seen": 591856, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.42283298097251587, |
|
"grad_norm": 1.4524191617965698, |
|
"learning_rate": 4.7578939341563095e-05, |
|
"loss": 3.286, |
|
"num_input_tokens_seen": 655648, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.46511627906976744, |
|
"grad_norm": 1.248831033706665, |
|
"learning_rate": 4.70806298768736e-05, |
|
"loss": 3.5377, |
|
"num_input_tokens_seen": 721280, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.507399577167019, |
|
"grad_norm": 1.2427473068237305, |
|
"learning_rate": 4.653885203484515e-05, |
|
"loss": 3.56, |
|
"num_input_tokens_seen": 784688, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.5496828752642706, |
|
"grad_norm": 1.323653221130371, |
|
"learning_rate": 4.595467237036329e-05, |
|
"loss": 3.4937, |
|
"num_input_tokens_seen": 850848, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.5919661733615222, |
|
"grad_norm": 1.3730229139328003, |
|
"learning_rate": 4.532924091140417e-05, |
|
"loss": 3.3823, |
|
"num_input_tokens_seen": 912480, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.6342494714587738, |
|
"grad_norm": 1.5328903198242188, |
|
"learning_rate": 4.466378889506607e-05, |
|
"loss": 3.3798, |
|
"num_input_tokens_seen": 978448, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.6765327695560254, |
|
"grad_norm": 1.4153543710708618, |
|
"learning_rate": 4.395962634373097e-05, |
|
"loss": 3.3044, |
|
"num_input_tokens_seen": 1041280, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.718816067653277, |
|
"grad_norm": 1.6301600933074951, |
|
"learning_rate": 4.3218139486127854e-05, |
|
"loss": 3.3661, |
|
"num_input_tokens_seen": 1102224, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.7610993657505285, |
|
"grad_norm": 1.6634522676467896, |
|
"learning_rate": 4.2440788028374624e-05, |
|
"loss": 3.3829, |
|
"num_input_tokens_seen": 1166576, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.8033826638477801, |
|
"grad_norm": 1.4539167881011963, |
|
"learning_rate": 4.1629102280370904e-05, |
|
"loss": 3.2241, |
|
"num_input_tokens_seen": 1230096, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.8456659619450317, |
|
"grad_norm": 1.453364372253418, |
|
"learning_rate": 4.0784680143198836e-05, |
|
"loss": 3.0931, |
|
"num_input_tokens_seen": 1297968, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.8879492600422833, |
|
"grad_norm": 2.0559964179992676, |
|
"learning_rate": 3.990918396346254e-05, |
|
"loss": 3.3581, |
|
"num_input_tokens_seen": 1361760, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.9302325581395349, |
|
"grad_norm": 2.0426218509674072, |
|
"learning_rate": 3.900433726075865e-05, |
|
"loss": 3.2308, |
|
"num_input_tokens_seen": 1422096, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.9725158562367865, |
|
"grad_norm": 1.7696195840835571, |
|
"learning_rate": 3.8071921334720696e-05, |
|
"loss": 3.1427, |
|
"num_input_tokens_seen": 1491120, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.014799154334038, |
|
"grad_norm": 2.3205981254577637, |
|
"learning_rate": 3.711377175831626e-05, |
|
"loss": 3.0988, |
|
"num_input_tokens_seen": 1556440, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.0570824524312896, |
|
"grad_norm": 1.7631757259368896, |
|
"learning_rate": 3.613177476430079e-05, |
|
"loss": 3.0695, |
|
"num_input_tokens_seen": 1624200, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.0993657505285412, |
|
"grad_norm": 2.32861328125, |
|
"learning_rate": 3.512786353194134e-05, |
|
"loss": 3.1109, |
|
"num_input_tokens_seen": 1686600, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.1416490486257929, |
|
"grad_norm": 2.1291067600250244, |
|
"learning_rate": 3.410401438132056e-05, |
|
"loss": 3.116, |
|
"num_input_tokens_seen": 1755144, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.1839323467230445, |
|
"grad_norm": 1.8158444166183472, |
|
"learning_rate": 3.3062242882712724e-05, |
|
"loss": 2.9528, |
|
"num_input_tokens_seen": 1819672, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.226215644820296, |
|
"grad_norm": 2.161515951156616, |
|
"learning_rate": 3.200459988869111e-05, |
|
"loss": 2.8494, |
|
"num_input_tokens_seen": 1886136, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.2684989429175475, |
|
"grad_norm": 2.321112632751465, |
|
"learning_rate": 3.093316749677788e-05, |
|
"loss": 3.1525, |
|
"num_input_tokens_seen": 1947656, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.3107822410147991, |
|
"grad_norm": 2.4329445362091064, |
|
"learning_rate": 2.985005495058446e-05, |
|
"loss": 2.8991, |
|
"num_input_tokens_seen": 2013272, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.3530655391120507, |
|
"grad_norm": 2.7929720878601074, |
|
"learning_rate": 2.875739448751176e-05, |
|
"loss": 3.026, |
|
"num_input_tokens_seen": 2079816, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.3953488372093024, |
|
"grad_norm": 2.6059908866882324, |
|
"learning_rate": 2.7657337141184138e-05, |
|
"loss": 2.8813, |
|
"num_input_tokens_seen": 2146008, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.437632135306554, |
|
"grad_norm": 2.831233501434326, |
|
"learning_rate": 2.655204850688085e-05, |
|
"loss": 2.7721, |
|
"num_input_tokens_seen": 2212840, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.4799154334038054, |
|
"grad_norm": 2.757953643798828, |
|
"learning_rate": 2.5443704478301154e-05, |
|
"loss": 2.737, |
|
"num_input_tokens_seen": 2278824, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.522198731501057, |
|
"grad_norm": 2.8655178546905518, |
|
"learning_rate": 2.433448696405563e-05, |
|
"loss": 2.6645, |
|
"num_input_tokens_seen": 2339768, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.5644820295983086, |
|
"grad_norm": 2.8346216678619385, |
|
"learning_rate": 2.3226579592316538e-05, |
|
"loss": 2.7233, |
|
"num_input_tokens_seen": 2407976, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.6067653276955602, |
|
"grad_norm": 2.432692766189575, |
|
"learning_rate": 2.2122163412082927e-05, |
|
"loss": 2.7463, |
|
"num_input_tokens_seen": 2469176, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.6490486257928119, |
|
"grad_norm": 3.0815346240997314, |
|
"learning_rate": 2.1023412599523204e-05, |
|
"loss": 2.8348, |
|
"num_input_tokens_seen": 2528152, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.6913319238900635, |
|
"grad_norm": 3.362581968307495, |
|
"learning_rate": 1.993249017784766e-05, |
|
"loss": 2.7732, |
|
"num_input_tokens_seen": 2588984, |
|
"step": 200 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 354, |
|
"num_input_tokens_seen": 2588984, |
|
"num_train_epochs": 3, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.083238062474035e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|