|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9921259842519685, |
|
"eval_steps": 16, |
|
"global_step": 63, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.015748031496062992, |
|
"grad_norm": 0.034481361508369446, |
|
"learning_rate": 4e-05, |
|
"loss": 0.1412, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.015748031496062992, |
|
"eval_loss": 0.1612786203622818, |
|
"eval_runtime": 64.5157, |
|
"eval_samples_per_second": 7.812, |
|
"eval_steps_per_second": 0.977, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.031496062992125984, |
|
"grad_norm": 0.029317770153284073, |
|
"learning_rate": 8e-05, |
|
"loss": 0.1191, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.047244094488188976, |
|
"grad_norm": 0.036621659994125366, |
|
"learning_rate": 0.00012, |
|
"loss": 0.1369, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.06299212598425197, |
|
"grad_norm": 0.04425783455371857, |
|
"learning_rate": 0.00016, |
|
"loss": 0.1321, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.07874015748031496, |
|
"grad_norm": 0.05247063934803009, |
|
"learning_rate": 0.0002, |
|
"loss": 0.1285, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.09448818897637795, |
|
"grad_norm": 0.03902214765548706, |
|
"learning_rate": 0.00019996629653035126, |
|
"loss": 0.1004, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.11023622047244094, |
|
"grad_norm": 0.03752463683485985, |
|
"learning_rate": 0.00019986520883988232, |
|
"loss": 0.0985, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.12598425196850394, |
|
"grad_norm": 0.03061060793697834, |
|
"learning_rate": 0.00019969680506871137, |
|
"loss": 0.0912, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.14173228346456693, |
|
"grad_norm": 0.034427180886268616, |
|
"learning_rate": 0.00019946119873266613, |
|
"loss": 0.0836, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.15748031496062992, |
|
"grad_norm": 0.03106631338596344, |
|
"learning_rate": 0.00019915854864676664, |
|
"loss": 0.0734, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.1732283464566929, |
|
"grad_norm": 0.02498232200741768, |
|
"learning_rate": 0.00019878905881817252, |
|
"loss": 0.0729, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.1889763779527559, |
|
"grad_norm": 0.03798564895987511, |
|
"learning_rate": 0.00019835297830866826, |
|
"loss": 0.0694, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.2047244094488189, |
|
"grad_norm": 0.046124912798404694, |
|
"learning_rate": 0.00019785060106677818, |
|
"loss": 0.0833, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.2204724409448819, |
|
"grad_norm": 0.02981509082019329, |
|
"learning_rate": 0.00019728226572962473, |
|
"loss": 0.0713, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.23622047244094488, |
|
"grad_norm": 0.02461801841855049, |
|
"learning_rate": 0.0001966483553946637, |
|
"loss": 0.0657, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.25196850393700787, |
|
"grad_norm": 0.04344266653060913, |
|
"learning_rate": 0.00019594929736144976, |
|
"loss": 0.0635, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.25196850393700787, |
|
"eval_loss": 0.06579381227493286, |
|
"eval_runtime": 64.6166, |
|
"eval_samples_per_second": 7.8, |
|
"eval_steps_per_second": 0.975, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.2677165354330709, |
|
"grad_norm": 0.0320642925798893, |
|
"learning_rate": 0.00019518556284360696, |
|
"loss": 0.0656, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.28346456692913385, |
|
"grad_norm": 0.028899891301989555, |
|
"learning_rate": 0.0001943576666511982, |
|
"loss": 0.0462, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.2992125984251969, |
|
"grad_norm": 0.02383616380393505, |
|
"learning_rate": 0.0001934661668437073, |
|
"loss": 0.0649, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.31496062992125984, |
|
"grad_norm": 0.03346535563468933, |
|
"learning_rate": 0.0001925116643538684, |
|
"loss": 0.0546, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.33070866141732286, |
|
"grad_norm": 0.020454615354537964, |
|
"learning_rate": 0.00019149480258259533, |
|
"loss": 0.0538, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.3464566929133858, |
|
"grad_norm": 0.02081696316599846, |
|
"learning_rate": 0.00019041626696528503, |
|
"loss": 0.0526, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.36220472440944884, |
|
"grad_norm": 0.028128350153565407, |
|
"learning_rate": 0.0001892767845097864, |
|
"loss": 0.0593, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.3779527559055118, |
|
"grad_norm": 0.015519126318395138, |
|
"learning_rate": 0.00018807712330634642, |
|
"loss": 0.0528, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.3937007874015748, |
|
"grad_norm": 0.03593792766332626, |
|
"learning_rate": 0.0001868180920098644, |
|
"loss": 0.0481, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.4094488188976378, |
|
"grad_norm": 0.015408644452691078, |
|
"learning_rate": 0.00018550053929480202, |
|
"loss": 0.0479, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.4251968503937008, |
|
"grad_norm": 0.021226301789283752, |
|
"learning_rate": 0.00018412535328311814, |
|
"loss": 0.054, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.4409448818897638, |
|
"grad_norm": 0.01717953570187092, |
|
"learning_rate": 0.0001826934609456129, |
|
"loss": 0.0523, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.4566929133858268, |
|
"grad_norm": 0.019626960158348083, |
|
"learning_rate": 0.00018120582747708502, |
|
"loss": 0.0512, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.47244094488188976, |
|
"grad_norm": 0.019186396151781082, |
|
"learning_rate": 0.0001796634556457236, |
|
"loss": 0.05, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.4881889763779528, |
|
"grad_norm": 0.014989328570663929, |
|
"learning_rate": 0.0001780673851171728, |
|
"loss": 0.0441, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.5039370078740157, |
|
"grad_norm": 0.012519012205302715, |
|
"learning_rate": 0.00017641869175372493, |
|
"loss": 0.0459, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.5039370078740157, |
|
"eval_loss": 0.056131936609745026, |
|
"eval_runtime": 64.4985, |
|
"eval_samples_per_second": 7.814, |
|
"eval_steps_per_second": 0.977, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.5196850393700787, |
|
"grad_norm": 0.01598576456308365, |
|
"learning_rate": 0.00017471848688911464, |
|
"loss": 0.0496, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.5354330708661418, |
|
"grad_norm": 0.017361309379339218, |
|
"learning_rate": 0.000172967916579403, |
|
"loss": 0.0534, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.5511811023622047, |
|
"grad_norm": 0.021230200305581093, |
|
"learning_rate": 0.00017116816083045602, |
|
"loss": 0.0505, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.5669291338582677, |
|
"grad_norm": 0.01624094881117344, |
|
"learning_rate": 0.0001693204328025389, |
|
"loss": 0.0568, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.5826771653543307, |
|
"grad_norm": 0.014916475862264633, |
|
"learning_rate": 0.00016742597799256182, |
|
"loss": 0.0542, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.5984251968503937, |
|
"grad_norm": 0.013211382552981377, |
|
"learning_rate": 0.00016548607339452853, |
|
"loss": 0.0507, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.6141732283464567, |
|
"grad_norm": 0.01305565144866705, |
|
"learning_rate": 0.00016350202663875386, |
|
"loss": 0.0387, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.6299212598425197, |
|
"grad_norm": 0.011459614150226116, |
|
"learning_rate": 0.0001614751751104301, |
|
"loss": 0.0433, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.6456692913385826, |
|
"grad_norm": 0.014712609350681305, |
|
"learning_rate": 0.00015940688504813662, |
|
"loss": 0.0571, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.6614173228346457, |
|
"grad_norm": 0.015662657096982002, |
|
"learning_rate": 0.00015729855062290022, |
|
"loss": 0.0504, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.6771653543307087, |
|
"grad_norm": 0.011235736310482025, |
|
"learning_rate": 0.00015515159299842707, |
|
"loss": 0.0453, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.6929133858267716, |
|
"grad_norm": 0.011984420008957386, |
|
"learning_rate": 0.00015296745937313987, |
|
"loss": 0.0402, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.7086614173228346, |
|
"grad_norm": 0.010523953475058079, |
|
"learning_rate": 0.00015074762200466556, |
|
"loss": 0.036, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.7244094488188977, |
|
"grad_norm": 0.013540665619075298, |
|
"learning_rate": 0.00014849357721743168, |
|
"loss": 0.0346, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.7401574803149606, |
|
"grad_norm": 0.012998640537261963, |
|
"learning_rate": 0.00014620684439403962, |
|
"loss": 0.0468, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.7559055118110236, |
|
"grad_norm": 0.01443515345454216, |
|
"learning_rate": 0.0001438889649510956, |
|
"loss": 0.0453, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.7559055118110236, |
|
"eval_loss": 0.05216333642601967, |
|
"eval_runtime": 64.5137, |
|
"eval_samples_per_second": 7.812, |
|
"eval_steps_per_second": 0.977, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.7716535433070866, |
|
"grad_norm": 0.01463907677680254, |
|
"learning_rate": 0.00014154150130018866, |
|
"loss": 0.0526, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.7874015748031497, |
|
"grad_norm": 0.01614455319941044, |
|
"learning_rate": 0.00013916603579471705, |
|
"loss": 0.0484, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8031496062992126, |
|
"grad_norm": 0.014042153023183346, |
|
"learning_rate": 0.000136764169663272, |
|
"loss": 0.0419, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.8188976377952756, |
|
"grad_norm": 0.015309924259781837, |
|
"learning_rate": 0.00013433752193029886, |
|
"loss": 0.0425, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.8346456692913385, |
|
"grad_norm": 0.018054217100143433, |
|
"learning_rate": 0.00013188772832476188, |
|
"loss": 0.0426, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.8503937007874016, |
|
"grad_norm": 0.012343033216893673, |
|
"learning_rate": 0.00012941644017754964, |
|
"loss": 0.0448, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.8661417322834646, |
|
"grad_norm": 0.012457596138119698, |
|
"learning_rate": 0.00012692532330836346, |
|
"loss": 0.0451, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.8818897637795275, |
|
"grad_norm": 0.013512413017451763, |
|
"learning_rate": 0.00012441605690283915, |
|
"loss": 0.0413, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.8976377952755905, |
|
"grad_norm": 0.013424846343696117, |
|
"learning_rate": 0.0001218903323806595, |
|
"loss": 0.0441, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.9133858267716536, |
|
"grad_norm": 0.014157367870211601, |
|
"learning_rate": 0.00011934985225541998, |
|
"loss": 0.0443, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.9291338582677166, |
|
"grad_norm": 0.0130110839381814, |
|
"learning_rate": 0.00011679632898701649, |
|
"loss": 0.0478, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.9448818897637795, |
|
"grad_norm": 0.012677576392889023, |
|
"learning_rate": 0.00011423148382732853, |
|
"loss": 0.0399, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.9606299212598425, |
|
"grad_norm": 0.01409006118774414, |
|
"learning_rate": 0.00011165704565997593, |
|
"loss": 0.0481, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.9763779527559056, |
|
"grad_norm": 0.013535700738430023, |
|
"learning_rate": 0.00010907474983493144, |
|
"loss": 0.0406, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.9921259842519685, |
|
"grad_norm": 0.014210895635187626, |
|
"learning_rate": 0.0001064863369987743, |
|
"loss": 0.0425, |
|
"step": 63 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 126, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 63, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.3132684787266355e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|