|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.7758046614872365, |
|
"eval_steps": 25, |
|
"global_step": 50, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03551609322974473, |
|
"grad_norm": 12.847548484802246, |
|
"learning_rate": 5e-05, |
|
"loss": 4.6931, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03551609322974473, |
|
"eval_loss": 5.996153354644775, |
|
"eval_runtime": 1.4302, |
|
"eval_samples_per_second": 34.96, |
|
"eval_steps_per_second": 9.09, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.07103218645948946, |
|
"grad_norm": 14.917906761169434, |
|
"learning_rate": 0.0001, |
|
"loss": 5.4258, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.10654827968923418, |
|
"grad_norm": 12.454079627990723, |
|
"learning_rate": 9.990365154573717e-05, |
|
"loss": 4.6077, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.14206437291897892, |
|
"grad_norm": 7.093257427215576, |
|
"learning_rate": 9.961501876182148e-05, |
|
"loss": 3.1281, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.17758046614872364, |
|
"grad_norm": 8.967534065246582, |
|
"learning_rate": 9.913533761814537e-05, |
|
"loss": 2.6663, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.21309655937846836, |
|
"grad_norm": 7.362691879272461, |
|
"learning_rate": 9.846666218300807e-05, |
|
"loss": 2.6746, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.2486126526082131, |
|
"grad_norm": 4.550023078918457, |
|
"learning_rate": 9.761185582727977e-05, |
|
"loss": 2.3725, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.28412874583795783, |
|
"grad_norm": 10.55962085723877, |
|
"learning_rate": 9.657457896300791e-05, |
|
"loss": 2.3137, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.3196448390677026, |
|
"grad_norm": 10.448192596435547, |
|
"learning_rate": 9.535927336897098e-05, |
|
"loss": 2.3299, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.3551609322974473, |
|
"grad_norm": 3.725038766860962, |
|
"learning_rate": 9.397114317029975e-05, |
|
"loss": 2.0264, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.390677025527192, |
|
"grad_norm": 6.806924343109131, |
|
"learning_rate": 9.241613255361455e-05, |
|
"loss": 2.2055, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.4261931187569367, |
|
"grad_norm": 6.795745849609375, |
|
"learning_rate": 9.070090031310558e-05, |
|
"loss": 1.8821, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.46170921198668147, |
|
"grad_norm": 3.854661226272583, |
|
"learning_rate": 8.883279133655399e-05, |
|
"loss": 1.8579, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.4972253052164262, |
|
"grad_norm": 2.821107864379883, |
|
"learning_rate": 8.681980515339464e-05, |
|
"loss": 1.969, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.532741398446171, |
|
"grad_norm": 2.6744842529296875, |
|
"learning_rate": 8.467056167950311e-05, |
|
"loss": 1.8684, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.5682574916759157, |
|
"grad_norm": 2.6326615810394287, |
|
"learning_rate": 8.239426430539243e-05, |
|
"loss": 1.9566, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.6037735849056604, |
|
"grad_norm": 2.0202972888946533, |
|
"learning_rate": 8.000066048588211e-05, |
|
"loss": 1.8665, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.6392896781354052, |
|
"grad_norm": 2.3964056968688965, |
|
"learning_rate": 7.75e-05, |
|
"loss": 1.9297, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.6748057713651499, |
|
"grad_norm": 2.168318271636963, |
|
"learning_rate": 7.490299105985507e-05, |
|
"loss": 1.7705, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.7103218645948945, |
|
"grad_norm": 1.984513521194458, |
|
"learning_rate": 7.222075445642904e-05, |
|
"loss": 1.7141, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.7458379578246392, |
|
"grad_norm": 2.0909554958343506, |
|
"learning_rate": 6.946477593864228e-05, |
|
"loss": 1.7412, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.781354051054384, |
|
"grad_norm": 1.8733512163162231, |
|
"learning_rate": 6.664685702961344e-05, |
|
"loss": 1.7218, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.8168701442841287, |
|
"grad_norm": 1.8428502082824707, |
|
"learning_rate": 6.377906449072578e-05, |
|
"loss": 1.8459, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.8523862375138734, |
|
"grad_norm": 2.086453676223755, |
|
"learning_rate": 6.087367864990233e-05, |
|
"loss": 1.8321, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.8879023307436182, |
|
"grad_norm": 2.0354926586151123, |
|
"learning_rate": 5.794314081535644e-05, |
|
"loss": 1.768, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.8879023307436182, |
|
"eval_loss": 1.8483245372772217, |
|
"eval_runtime": 1.4767, |
|
"eval_samples_per_second": 33.859, |
|
"eval_steps_per_second": 8.803, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.9234184239733629, |
|
"grad_norm": 2.222177743911743, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 1.6803, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.9589345172031076, |
|
"grad_norm": 2.106823444366455, |
|
"learning_rate": 5.205685918464356e-05, |
|
"loss": 1.7212, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.9944506104328524, |
|
"grad_norm": 2.1512210369110107, |
|
"learning_rate": 4.912632135009769e-05, |
|
"loss": 1.7802, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 1.029966703662597, |
|
"grad_norm": 4.948949337005615, |
|
"learning_rate": 4.6220935509274235e-05, |
|
"loss": 2.9527, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 1.065482796892342, |
|
"grad_norm": 1.8342108726501465, |
|
"learning_rate": 4.3353142970386564e-05, |
|
"loss": 1.637, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.1009988901220866, |
|
"grad_norm": 1.786393165588379, |
|
"learning_rate": 4.053522406135775e-05, |
|
"loss": 1.6031, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 1.1365149833518313, |
|
"grad_norm": 1.9454339742660522, |
|
"learning_rate": 3.777924554357096e-05, |
|
"loss": 1.3323, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 1.172031076581576, |
|
"grad_norm": 2.2301857471466064, |
|
"learning_rate": 3.509700894014496e-05, |
|
"loss": 1.5233, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.2075471698113207, |
|
"grad_norm": 2.044020891189575, |
|
"learning_rate": 3.250000000000001e-05, |
|
"loss": 1.5415, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.2430632630410654, |
|
"grad_norm": 2.0642783641815186, |
|
"learning_rate": 2.9999339514117912e-05, |
|
"loss": 1.4577, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 1.2785793562708103, |
|
"grad_norm": 2.05415415763855, |
|
"learning_rate": 2.760573569460757e-05, |
|
"loss": 1.434, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.314095449500555, |
|
"grad_norm": 2.0152251720428467, |
|
"learning_rate": 2.53294383204969e-05, |
|
"loss": 1.4434, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 1.3496115427302997, |
|
"grad_norm": 2.1848814487457275, |
|
"learning_rate": 2.3180194846605367e-05, |
|
"loss": 1.5678, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.3851276359600444, |
|
"grad_norm": 2.1174445152282715, |
|
"learning_rate": 2.1167208663446025e-05, |
|
"loss": 1.5425, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.420643729189789, |
|
"grad_norm": 2.277348518371582, |
|
"learning_rate": 1.9299099686894423e-05, |
|
"loss": 1.3887, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.4561598224195338, |
|
"grad_norm": 2.1913950443267822, |
|
"learning_rate": 1.758386744638546e-05, |
|
"loss": 1.4529, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.4916759156492785, |
|
"grad_norm": 2.30416202545166, |
|
"learning_rate": 1.602885682970026e-05, |
|
"loss": 1.4513, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.5271920088790232, |
|
"grad_norm": 2.0522561073303223, |
|
"learning_rate": 1.464072663102903e-05, |
|
"loss": 1.3965, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.5627081021087679, |
|
"grad_norm": 2.1286027431488037, |
|
"learning_rate": 1.3425421036992098e-05, |
|
"loss": 1.5099, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.5982241953385128, |
|
"grad_norm": 2.026170253753662, |
|
"learning_rate": 1.2388144172720251e-05, |
|
"loss": 1.4889, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.6337402885682575, |
|
"grad_norm": 2.003553628921509, |
|
"learning_rate": 1.1533337816991932e-05, |
|
"loss": 1.415, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.6692563817980022, |
|
"grad_norm": 2.1225321292877197, |
|
"learning_rate": 1.0864662381854632e-05, |
|
"loss": 1.3807, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.704772475027747, |
|
"grad_norm": 2.1674954891204834, |
|
"learning_rate": 1.0384981238178534e-05, |
|
"loss": 1.3571, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.7402885682574918, |
|
"grad_norm": 2.1650798320770264, |
|
"learning_rate": 1.0096348454262845e-05, |
|
"loss": 1.5067, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.7758046614872365, |
|
"grad_norm": 2.0062143802642822, |
|
"learning_rate": 1e-05, |
|
"loss": 1.3259, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.7758046614872365, |
|
"eval_loss": 1.8066482543945312, |
|
"eval_runtime": 1.4758, |
|
"eval_samples_per_second": 33.88, |
|
"eval_steps_per_second": 8.809, |
|
"step": 50 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 50, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.589463780163584e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|