|
{ |
|
"best_metric": 0.4896911084651947, |
|
"best_model_checkpoint": "./beans_outputs/checkpoint-1495", |
|
"epoch": 200.0, |
|
"eval_steps": 500, |
|
"global_step": 4600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.43478260869565216, |
|
"grad_norm": 3.492309093475342, |
|
"learning_rate": 1.9826086956521742e-05, |
|
"loss": 0.7305, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.8695652173913043, |
|
"grad_norm": 4.003854751586914, |
|
"learning_rate": 1.965217391304348e-05, |
|
"loss": 0.6794, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.6062992125984252, |
|
"eval_loss": 0.6559741497039795, |
|
"eval_runtime": 4.6669, |
|
"eval_samples_per_second": 54.425, |
|
"eval_steps_per_second": 0.857, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 1.3043478260869565, |
|
"grad_norm": 4.675185680389404, |
|
"learning_rate": 1.947826086956522e-05, |
|
"loss": 0.6399, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.7391304347826086, |
|
"grad_norm": 4.350035667419434, |
|
"learning_rate": 1.9304347826086957e-05, |
|
"loss": 0.6215, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7362204724409449, |
|
"eval_loss": 0.5833372473716736, |
|
"eval_runtime": 3.7107, |
|
"eval_samples_per_second": 68.451, |
|
"eval_steps_per_second": 1.078, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 2.1739130434782608, |
|
"grad_norm": 3.180147886276245, |
|
"learning_rate": 1.9130434782608697e-05, |
|
"loss": 0.5964, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 2.608695652173913, |
|
"grad_norm": 3.246190309524536, |
|
"learning_rate": 1.8956521739130434e-05, |
|
"loss": 0.5784, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7598425196850394, |
|
"eval_loss": 0.5489528179168701, |
|
"eval_runtime": 4.3517, |
|
"eval_samples_per_second": 58.367, |
|
"eval_steps_per_second": 0.919, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 3.0434782608695654, |
|
"grad_norm": 2.3400914669036865, |
|
"learning_rate": 1.8782608695652175e-05, |
|
"loss": 0.5412, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 3.4782608695652173, |
|
"grad_norm": 3.5264837741851807, |
|
"learning_rate": 1.8608695652173912e-05, |
|
"loss": 0.5659, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 3.9130434782608696, |
|
"grad_norm": 4.993140697479248, |
|
"learning_rate": 1.8434782608695653e-05, |
|
"loss": 0.5347, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.7637795275590551, |
|
"eval_loss": 0.5305963754653931, |
|
"eval_runtime": 3.9321, |
|
"eval_samples_per_second": 64.596, |
|
"eval_steps_per_second": 1.017, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 4.3478260869565215, |
|
"grad_norm": 2.20806622505188, |
|
"learning_rate": 1.8260869565217393e-05, |
|
"loss": 0.5086, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 4.782608695652174, |
|
"grad_norm": 4.256261825561523, |
|
"learning_rate": 1.808695652173913e-05, |
|
"loss": 0.5307, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.7637795275590551, |
|
"eval_loss": 0.5235078930854797, |
|
"eval_runtime": 3.7141, |
|
"eval_samples_per_second": 68.389, |
|
"eval_steps_per_second": 1.077, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 5.217391304347826, |
|
"grad_norm": 2.6543545722961426, |
|
"learning_rate": 1.791304347826087e-05, |
|
"loss": 0.5085, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 5.6521739130434785, |
|
"grad_norm": 4.274487495422363, |
|
"learning_rate": 1.773913043478261e-05, |
|
"loss": 0.5391, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.5090441703796387, |
|
"eval_runtime": 4.3438, |
|
"eval_samples_per_second": 58.475, |
|
"eval_steps_per_second": 0.921, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 6.086956521739131, |
|
"grad_norm": 3.147414445877075, |
|
"learning_rate": 1.756521739130435e-05, |
|
"loss": 0.4977, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 6.521739130434782, |
|
"grad_norm": 4.254673004150391, |
|
"learning_rate": 1.739130434782609e-05, |
|
"loss": 0.5297, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 6.956521739130435, |
|
"grad_norm": 2.083784818649292, |
|
"learning_rate": 1.721739130434783e-05, |
|
"loss": 0.48, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5108471512794495, |
|
"eval_runtime": 4.1877, |
|
"eval_samples_per_second": 60.653, |
|
"eval_steps_per_second": 0.955, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 7.391304347826087, |
|
"grad_norm": 4.193545818328857, |
|
"learning_rate": 1.7043478260869566e-05, |
|
"loss": 0.4826, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 7.826086956521739, |
|
"grad_norm": 2.05076003074646, |
|
"learning_rate": 1.6869565217391307e-05, |
|
"loss": 0.473, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.5028324127197266, |
|
"eval_runtime": 3.7021, |
|
"eval_samples_per_second": 68.61, |
|
"eval_steps_per_second": 1.08, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 8.26086956521739, |
|
"grad_norm": 3.007233142852783, |
|
"learning_rate": 1.6695652173913044e-05, |
|
"loss": 0.5255, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 8.695652173913043, |
|
"grad_norm": 2.196945905685425, |
|
"learning_rate": 1.6521739130434785e-05, |
|
"loss": 0.5014, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5054498314857483, |
|
"eval_runtime": 4.225, |
|
"eval_samples_per_second": 60.119, |
|
"eval_steps_per_second": 0.947, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 9.130434782608695, |
|
"grad_norm": 2.184353828430176, |
|
"learning_rate": 1.6347826086956525e-05, |
|
"loss": 0.5044, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 9.565217391304348, |
|
"grad_norm": 4.106619358062744, |
|
"learning_rate": 1.6173913043478262e-05, |
|
"loss": 0.4822, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 4.000082969665527, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.496, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5039955973625183, |
|
"eval_runtime": 5.3498, |
|
"eval_samples_per_second": 47.478, |
|
"eval_steps_per_second": 0.748, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 10.434782608695652, |
|
"grad_norm": 5.726933002471924, |
|
"learning_rate": 1.582608695652174e-05, |
|
"loss": 0.5101, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 10.869565217391305, |
|
"grad_norm": 4.100568771362305, |
|
"learning_rate": 1.565217391304348e-05, |
|
"loss": 0.4688, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.4972316324710846, |
|
"eval_runtime": 3.7607, |
|
"eval_samples_per_second": 67.54, |
|
"eval_steps_per_second": 1.064, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 11.304347826086957, |
|
"grad_norm": 2.6119587421417236, |
|
"learning_rate": 1.5478260869565217e-05, |
|
"loss": 0.485, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 11.73913043478261, |
|
"grad_norm": 3.003861427307129, |
|
"learning_rate": 1.5304347826086958e-05, |
|
"loss": 0.4943, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.7637795275590551, |
|
"eval_loss": 0.49771231412887573, |
|
"eval_runtime": 4.9203, |
|
"eval_samples_per_second": 51.622, |
|
"eval_steps_per_second": 0.813, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 12.173913043478262, |
|
"grad_norm": 2.9490270614624023, |
|
"learning_rate": 1.5130434782608697e-05, |
|
"loss": 0.4505, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 12.608695652173914, |
|
"grad_norm": 2.8131847381591797, |
|
"learning_rate": 1.4956521739130436e-05, |
|
"loss": 0.5012, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5057242512702942, |
|
"eval_runtime": 3.7024, |
|
"eval_samples_per_second": 68.605, |
|
"eval_steps_per_second": 1.08, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 13.043478260869565, |
|
"grad_norm": 17.65978240966797, |
|
"learning_rate": 1.4782608695652174e-05, |
|
"loss": 0.4768, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 13.478260869565217, |
|
"grad_norm": 2.085587978363037, |
|
"learning_rate": 1.4608695652173915e-05, |
|
"loss": 0.4729, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 13.91304347826087, |
|
"grad_norm": 4.59744119644165, |
|
"learning_rate": 1.4434782608695654e-05, |
|
"loss": 0.4639, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5010089874267578, |
|
"eval_runtime": 3.7018, |
|
"eval_samples_per_second": 68.616, |
|
"eval_steps_per_second": 1.081, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 14.347826086956522, |
|
"grad_norm": 2.4057395458221436, |
|
"learning_rate": 1.4260869565217392e-05, |
|
"loss": 0.4751, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 14.782608695652174, |
|
"grad_norm": 3.549567222595215, |
|
"learning_rate": 1.4086956521739133e-05, |
|
"loss": 0.4709, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.4948899447917938, |
|
"eval_runtime": 4.9714, |
|
"eval_samples_per_second": 51.092, |
|
"eval_steps_per_second": 0.805, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 15.217391304347826, |
|
"grad_norm": 6.705427646636963, |
|
"learning_rate": 1.391304347826087e-05, |
|
"loss": 0.4379, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 15.652173913043478, |
|
"grad_norm": 2.444533348083496, |
|
"learning_rate": 1.373913043478261e-05, |
|
"loss": 0.4888, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.49550917744636536, |
|
"eval_runtime": 3.6768, |
|
"eval_samples_per_second": 69.081, |
|
"eval_steps_per_second": 1.088, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 16.08695652173913, |
|
"grad_norm": 5.470461845397949, |
|
"learning_rate": 1.3565217391304348e-05, |
|
"loss": 0.4952, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 16.52173913043478, |
|
"grad_norm": 2.0678608417510986, |
|
"learning_rate": 1.3391304347826088e-05, |
|
"loss": 0.4784, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 16.956521739130434, |
|
"grad_norm": 6.63480806350708, |
|
"learning_rate": 1.3217391304347827e-05, |
|
"loss": 0.4594, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.49856194853782654, |
|
"eval_runtime": 3.7219, |
|
"eval_samples_per_second": 68.245, |
|
"eval_steps_per_second": 1.075, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 17.391304347826086, |
|
"grad_norm": 4.448991298675537, |
|
"learning_rate": 1.3043478260869566e-05, |
|
"loss": 0.4607, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 17.82608695652174, |
|
"grad_norm": 2.716780424118042, |
|
"learning_rate": 1.2869565217391305e-05, |
|
"loss": 0.4745, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.501070499420166, |
|
"eval_runtime": 4.5054, |
|
"eval_samples_per_second": 56.377, |
|
"eval_steps_per_second": 0.888, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 18.26086956521739, |
|
"grad_norm": 2.406355857849121, |
|
"learning_rate": 1.2695652173913045e-05, |
|
"loss": 0.4639, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 18.695652173913043, |
|
"grad_norm": 5.627669811248779, |
|
"learning_rate": 1.2521739130434784e-05, |
|
"loss": 0.4667, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.4928434491157532, |
|
"eval_runtime": 4.0475, |
|
"eval_samples_per_second": 62.756, |
|
"eval_steps_per_second": 0.988, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 19.130434782608695, |
|
"grad_norm": 4.074652671813965, |
|
"learning_rate": 1.2347826086956523e-05, |
|
"loss": 0.4671, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 19.565217391304348, |
|
"grad_norm": 5.88148832321167, |
|
"learning_rate": 1.2173913043478263e-05, |
|
"loss": 0.4442, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 3.00347900390625, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.4551, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5055357217788696, |
|
"eval_runtime": 3.6885, |
|
"eval_samples_per_second": 68.862, |
|
"eval_steps_per_second": 1.084, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 20.434782608695652, |
|
"grad_norm": 10.164237976074219, |
|
"learning_rate": 1.182608695652174e-05, |
|
"loss": 0.4657, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 20.869565217391305, |
|
"grad_norm": 2.1962711811065674, |
|
"learning_rate": 1.1652173913043478e-05, |
|
"loss": 0.4657, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.4928124248981476, |
|
"eval_runtime": 4.4478, |
|
"eval_samples_per_second": 57.107, |
|
"eval_steps_per_second": 0.899, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 21.304347826086957, |
|
"grad_norm": 5.0302228927612305, |
|
"learning_rate": 1.1478260869565218e-05, |
|
"loss": 0.4564, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 21.73913043478261, |
|
"grad_norm": 3.5275819301605225, |
|
"learning_rate": 1.1304347826086957e-05, |
|
"loss": 0.4818, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.5001721978187561, |
|
"eval_runtime": 4.0355, |
|
"eval_samples_per_second": 62.942, |
|
"eval_steps_per_second": 0.991, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 22.17391304347826, |
|
"grad_norm": 6.920666694641113, |
|
"learning_rate": 1.1130434782608696e-05, |
|
"loss": 0.4608, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 22.608695652173914, |
|
"grad_norm": 2.2840707302093506, |
|
"learning_rate": 1.0956521739130435e-05, |
|
"loss": 0.4633, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.49459317326545715, |
|
"eval_runtime": 3.7179, |
|
"eval_samples_per_second": 68.319, |
|
"eval_steps_per_second": 1.076, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 23.043478260869566, |
|
"grad_norm": 6.509201526641846, |
|
"learning_rate": 1.0782608695652175e-05, |
|
"loss": 0.4694, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 23.47826086956522, |
|
"grad_norm": 2.403275489807129, |
|
"learning_rate": 1.0608695652173914e-05, |
|
"loss": 0.4874, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 23.91304347826087, |
|
"grad_norm": 2.1320598125457764, |
|
"learning_rate": 1.0434782608695653e-05, |
|
"loss": 0.4779, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.49417200684547424, |
|
"eval_runtime": 4.3215, |
|
"eval_samples_per_second": 58.776, |
|
"eval_steps_per_second": 0.926, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 24.347826086956523, |
|
"grad_norm": 3.7421488761901855, |
|
"learning_rate": 1.0260869565217393e-05, |
|
"loss": 0.4579, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 24.782608695652176, |
|
"grad_norm": 3.07060170173645, |
|
"learning_rate": 1.008695652173913e-05, |
|
"loss": 0.4718, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.49625155329704285, |
|
"eval_runtime": 5.5612, |
|
"eval_samples_per_second": 45.674, |
|
"eval_steps_per_second": 0.719, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 25.217391304347824, |
|
"grad_norm": 4.446998596191406, |
|
"learning_rate": 9.913043478260871e-06, |
|
"loss": 0.443, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 25.652173913043477, |
|
"grad_norm": 2.4786624908447266, |
|
"learning_rate": 9.73913043478261e-06, |
|
"loss": 0.4511, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5011107325553894, |
|
"eval_runtime": 3.7637, |
|
"eval_samples_per_second": 67.487, |
|
"eval_steps_per_second": 1.063, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 26.08695652173913, |
|
"grad_norm": 5.552999019622803, |
|
"learning_rate": 9.565217391304349e-06, |
|
"loss": 0.4631, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 26.52173913043478, |
|
"grad_norm": 5.050811290740967, |
|
"learning_rate": 9.391304347826087e-06, |
|
"loss": 0.4564, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 26.956521739130434, |
|
"grad_norm": 2.0711512565612793, |
|
"learning_rate": 9.217391304347826e-06, |
|
"loss": 0.4798, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.4903908967971802, |
|
"eval_runtime": 4.9056, |
|
"eval_samples_per_second": 51.777, |
|
"eval_steps_per_second": 0.815, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 27.391304347826086, |
|
"grad_norm": 4.117509365081787, |
|
"learning_rate": 9.043478260869565e-06, |
|
"loss": 0.4411, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 27.82608695652174, |
|
"grad_norm": 4.448685646057129, |
|
"learning_rate": 8.869565217391306e-06, |
|
"loss": 0.4868, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.4982087016105652, |
|
"eval_runtime": 3.7322, |
|
"eval_samples_per_second": 68.057, |
|
"eval_steps_per_second": 1.072, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 28.26086956521739, |
|
"grad_norm": 3.0993807315826416, |
|
"learning_rate": 8.695652173913044e-06, |
|
"loss": 0.4414, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 28.695652173913043, |
|
"grad_norm": 4.982347011566162, |
|
"learning_rate": 8.521739130434783e-06, |
|
"loss": 0.4653, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.498798668384552, |
|
"eval_runtime": 3.7347, |
|
"eval_samples_per_second": 68.012, |
|
"eval_steps_per_second": 1.071, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 29.130434782608695, |
|
"grad_norm": 3.081833600997925, |
|
"learning_rate": 8.347826086956522e-06, |
|
"loss": 0.4503, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 29.565217391304348, |
|
"grad_norm": 4.352429389953613, |
|
"learning_rate": 8.173913043478263e-06, |
|
"loss": 0.4674, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"grad_norm": 5.281393051147461, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.4613, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.49851593375205994, |
|
"eval_runtime": 4.8766, |
|
"eval_samples_per_second": 52.085, |
|
"eval_steps_per_second": 0.82, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 30.434782608695652, |
|
"grad_norm": 2.2079997062683105, |
|
"learning_rate": 7.82608695652174e-06, |
|
"loss": 0.4574, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 30.869565217391305, |
|
"grad_norm": 4.6935858726501465, |
|
"learning_rate": 7.652173913043479e-06, |
|
"loss": 0.4675, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5060083270072937, |
|
"eval_runtime": 3.7305, |
|
"eval_samples_per_second": 68.087, |
|
"eval_steps_per_second": 1.072, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 31.304347826086957, |
|
"grad_norm": 4.8790602684021, |
|
"learning_rate": 7.478260869565218e-06, |
|
"loss": 0.4802, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 31.73913043478261, |
|
"grad_norm": 5.6365485191345215, |
|
"learning_rate": 7.304347826086957e-06, |
|
"loss": 0.4587, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5059147477149963, |
|
"eval_runtime": 3.7699, |
|
"eval_samples_per_second": 67.376, |
|
"eval_steps_per_second": 1.061, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 32.17391304347826, |
|
"grad_norm": 5.480165004730225, |
|
"learning_rate": 7.130434782608696e-06, |
|
"loss": 0.4541, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 32.608695652173914, |
|
"grad_norm": 2.053098440170288, |
|
"learning_rate": 6.956521739130435e-06, |
|
"loss": 0.464, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5041583180427551, |
|
"eval_runtime": 4.9981, |
|
"eval_samples_per_second": 50.82, |
|
"eval_steps_per_second": 0.8, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 33.04347826086956, |
|
"grad_norm": 3.6429481506347656, |
|
"learning_rate": 6.782608695652174e-06, |
|
"loss": 0.454, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 33.47826086956522, |
|
"grad_norm": 2.436143636703491, |
|
"learning_rate": 6.6086956521739135e-06, |
|
"loss": 0.4612, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 33.91304347826087, |
|
"grad_norm": 2.5793776512145996, |
|
"learning_rate": 6.434782608695652e-06, |
|
"loss": 0.4374, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.5063456296920776, |
|
"eval_runtime": 3.7117, |
|
"eval_samples_per_second": 68.432, |
|
"eval_steps_per_second": 1.078, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 34.34782608695652, |
|
"grad_norm": 3.71374773979187, |
|
"learning_rate": 6.260869565217392e-06, |
|
"loss": 0.4667, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 34.78260869565217, |
|
"grad_norm": 4.282368183135986, |
|
"learning_rate": 6.086956521739132e-06, |
|
"loss": 0.4864, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.5039507150650024, |
|
"eval_runtime": 3.6837, |
|
"eval_samples_per_second": 68.952, |
|
"eval_steps_per_second": 1.086, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 35.21739130434783, |
|
"grad_norm": 2.896638870239258, |
|
"learning_rate": 5.91304347826087e-06, |
|
"loss": 0.4922, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 35.65217391304348, |
|
"grad_norm": 2.2342097759246826, |
|
"learning_rate": 5.739130434782609e-06, |
|
"loss": 0.4354, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5108994841575623, |
|
"eval_runtime": 4.9899, |
|
"eval_samples_per_second": 50.902, |
|
"eval_steps_per_second": 0.802, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 36.08695652173913, |
|
"grad_norm": 8.385408401489258, |
|
"learning_rate": 5.565217391304348e-06, |
|
"loss": 0.4585, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 36.52173913043478, |
|
"grad_norm": 2.839411497116089, |
|
"learning_rate": 5.391304347826088e-06, |
|
"loss": 0.4497, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 36.95652173913044, |
|
"grad_norm": 2.479076623916626, |
|
"learning_rate": 5.2173913043478265e-06, |
|
"loss": 0.4655, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.510716438293457, |
|
"eval_runtime": 3.6997, |
|
"eval_samples_per_second": 68.653, |
|
"eval_steps_per_second": 1.081, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 37.391304347826086, |
|
"grad_norm": 2.271686553955078, |
|
"learning_rate": 5.043478260869565e-06, |
|
"loss": 0.4462, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 37.82608695652174, |
|
"grad_norm": 3.4210402965545654, |
|
"learning_rate": 4.869565217391305e-06, |
|
"loss": 0.4691, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.5093376636505127, |
|
"eval_runtime": 3.7287, |
|
"eval_samples_per_second": 68.119, |
|
"eval_steps_per_second": 1.073, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 38.26086956521739, |
|
"grad_norm": 5.694761276245117, |
|
"learning_rate": 4.695652173913044e-06, |
|
"loss": 0.4592, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 38.69565217391305, |
|
"grad_norm": 2.2949883937835693, |
|
"learning_rate": 4.5217391304347826e-06, |
|
"loss": 0.4826, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5044277906417847, |
|
"eval_runtime": 4.9781, |
|
"eval_samples_per_second": 51.024, |
|
"eval_steps_per_second": 0.804, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 39.130434782608695, |
|
"grad_norm": 3.4144210815429688, |
|
"learning_rate": 4.347826086956522e-06, |
|
"loss": 0.4407, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 39.56521739130435, |
|
"grad_norm": 2.22868013381958, |
|
"learning_rate": 4.173913043478261e-06, |
|
"loss": 0.4482, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 3.2193689346313477, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.4577, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.4999626874923706, |
|
"eval_runtime": 3.6952, |
|
"eval_samples_per_second": 68.738, |
|
"eval_steps_per_second": 1.082, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 40.43478260869565, |
|
"grad_norm": 4.500718593597412, |
|
"learning_rate": 3.8260869565217395e-06, |
|
"loss": 0.4585, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 40.869565217391305, |
|
"grad_norm": 1.9281222820281982, |
|
"learning_rate": 3.6521739130434787e-06, |
|
"loss": 0.4636, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.4962589144706726, |
|
"eval_runtime": 3.6977, |
|
"eval_samples_per_second": 68.69, |
|
"eval_steps_per_second": 1.082, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 41.30434782608695, |
|
"grad_norm": 2.193452835083008, |
|
"learning_rate": 3.4782608695652175e-06, |
|
"loss": 0.4306, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 41.73913043478261, |
|
"grad_norm": 2.2370336055755615, |
|
"learning_rate": 3.3043478260869567e-06, |
|
"loss": 0.4361, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.4958040118217468, |
|
"eval_runtime": 4.9548, |
|
"eval_samples_per_second": 51.264, |
|
"eval_steps_per_second": 0.807, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 42.17391304347826, |
|
"grad_norm": 3.6354355812072754, |
|
"learning_rate": 3.130434782608696e-06, |
|
"loss": 0.4514, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 42.608695652173914, |
|
"grad_norm": 1.8955118656158447, |
|
"learning_rate": 2.956521739130435e-06, |
|
"loss": 0.4534, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5007808208465576, |
|
"eval_runtime": 3.7121, |
|
"eval_samples_per_second": 68.424, |
|
"eval_steps_per_second": 1.078, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 43.04347826086956, |
|
"grad_norm": 2.2034902572631836, |
|
"learning_rate": 2.782608695652174e-06, |
|
"loss": 0.4176, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 43.47826086956522, |
|
"grad_norm": 4.387076377868652, |
|
"learning_rate": 2.6086956521739132e-06, |
|
"loss": 0.4748, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 43.91304347826087, |
|
"grad_norm": 5.444644927978516, |
|
"learning_rate": 2.4347826086956525e-06, |
|
"loss": 0.4559, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5025174021720886, |
|
"eval_runtime": 3.7093, |
|
"eval_samples_per_second": 68.476, |
|
"eval_steps_per_second": 1.078, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 44.34782608695652, |
|
"grad_norm": 2.2067017555236816, |
|
"learning_rate": 2.2608695652173913e-06, |
|
"loss": 0.4882, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 44.78260869565217, |
|
"grad_norm": 3.562736988067627, |
|
"learning_rate": 2.0869565217391305e-06, |
|
"loss": 0.4189, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.5014046430587769, |
|
"eval_runtime": 4.9992, |
|
"eval_samples_per_second": 50.808, |
|
"eval_steps_per_second": 0.8, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 45.21739130434783, |
|
"grad_norm": 10.402663230895996, |
|
"learning_rate": 1.9130434782608697e-06, |
|
"loss": 0.4432, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 45.65217391304348, |
|
"grad_norm": 4.949878215789795, |
|
"learning_rate": 1.7391304347826088e-06, |
|
"loss": 0.4861, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_accuracy": 0.7677165354330708, |
|
"eval_loss": 0.5003762245178223, |
|
"eval_runtime": 3.7019, |
|
"eval_samples_per_second": 68.614, |
|
"eval_steps_per_second": 1.081, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 46.08695652173913, |
|
"grad_norm": 1.938593864440918, |
|
"learning_rate": 1.565217391304348e-06, |
|
"loss": 0.4326, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 46.52173913043478, |
|
"grad_norm": 3.236699342727661, |
|
"learning_rate": 1.391304347826087e-06, |
|
"loss": 0.4726, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 46.95652173913044, |
|
"grad_norm": 3.047184944152832, |
|
"learning_rate": 1.2173913043478262e-06, |
|
"loss": 0.4709, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5004997849464417, |
|
"eval_runtime": 3.7143, |
|
"eval_samples_per_second": 68.384, |
|
"eval_steps_per_second": 1.077, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 47.391304347826086, |
|
"grad_norm": 2.8639461994171143, |
|
"learning_rate": 1.0434782608695653e-06, |
|
"loss": 0.4649, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 47.82608695652174, |
|
"grad_norm": 3.7704715728759766, |
|
"learning_rate": 8.695652173913044e-07, |
|
"loss": 0.4726, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.5007592439651489, |
|
"eval_runtime": 4.8498, |
|
"eval_samples_per_second": 52.373, |
|
"eval_steps_per_second": 0.825, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 48.26086956521739, |
|
"grad_norm": 4.941337585449219, |
|
"learning_rate": 6.956521739130435e-07, |
|
"loss": 0.4314, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 48.69565217391305, |
|
"grad_norm": 3.2265655994415283, |
|
"learning_rate": 5.217391304347826e-07, |
|
"loss": 0.4441, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.4987953305244446, |
|
"eval_runtime": 3.6681, |
|
"eval_samples_per_second": 69.246, |
|
"eval_steps_per_second": 1.09, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 49.130434782608695, |
|
"grad_norm": 3.7678611278533936, |
|
"learning_rate": 3.4782608695652175e-07, |
|
"loss": 0.4571, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 49.56521739130435, |
|
"grad_norm": 3.657460927963257, |
|
"learning_rate": 1.7391304347826088e-07, |
|
"loss": 0.4558, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"grad_norm": 3.096832513809204, |
|
"learning_rate": 0.0, |
|
"loss": 0.4579, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.499985009431839, |
|
"eval_runtime": 3.8189, |
|
"eval_samples_per_second": 66.512, |
|
"eval_steps_per_second": 1.047, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 50.43478260869565, |
|
"grad_norm": 2.7469470500946045, |
|
"learning_rate": 1.4956521739130436e-05, |
|
"loss": 0.4877, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 50.869565217391305, |
|
"grad_norm": 2.5254504680633545, |
|
"learning_rate": 1.491304347826087e-05, |
|
"loss": 0.4366, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.497986376285553, |
|
"eval_runtime": 3.9676, |
|
"eval_samples_per_second": 64.019, |
|
"eval_steps_per_second": 0.504, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 51.30434782608695, |
|
"grad_norm": 2.032457113265991, |
|
"learning_rate": 1.4869565217391306e-05, |
|
"loss": 0.4663, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 51.73913043478261, |
|
"grad_norm": 2.802882432937622, |
|
"learning_rate": 1.4826086956521741e-05, |
|
"loss": 0.4467, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.49471431970596313, |
|
"eval_runtime": 4.8494, |
|
"eval_samples_per_second": 52.378, |
|
"eval_steps_per_second": 0.412, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 52.17391304347826, |
|
"grad_norm": 2.2359466552734375, |
|
"learning_rate": 1.4782608695652174e-05, |
|
"loss": 0.4424, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 52.608695652173914, |
|
"grad_norm": 3.219308376312256, |
|
"learning_rate": 1.473913043478261e-05, |
|
"loss": 0.4797, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 53.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.4950390160083771, |
|
"eval_runtime": 3.8997, |
|
"eval_samples_per_second": 65.133, |
|
"eval_steps_per_second": 0.513, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 53.04347826086956, |
|
"grad_norm": 2.6939969062805176, |
|
"learning_rate": 1.4695652173913045e-05, |
|
"loss": 0.4256, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 53.47826086956522, |
|
"grad_norm": 2.6343085765838623, |
|
"learning_rate": 1.465217391304348e-05, |
|
"loss": 0.4192, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 53.91304347826087, |
|
"grad_norm": 10.655885696411133, |
|
"learning_rate": 1.4608695652173915e-05, |
|
"loss": 0.4544, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 54.0, |
|
"eval_accuracy": 0.7716535433070866, |
|
"eval_loss": 0.49978330731391907, |
|
"eval_runtime": 3.7833, |
|
"eval_samples_per_second": 67.137, |
|
"eval_steps_per_second": 0.529, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 54.34782608695652, |
|
"grad_norm": 4.838284969329834, |
|
"learning_rate": 1.456521739130435e-05, |
|
"loss": 0.4361, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 54.78260869565217, |
|
"grad_norm": 3.1171820163726807, |
|
"learning_rate": 1.4521739130434785e-05, |
|
"loss": 0.4466, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.49803978204727173, |
|
"eval_runtime": 3.7735, |
|
"eval_samples_per_second": 67.311, |
|
"eval_steps_per_second": 0.53, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 55.21739130434783, |
|
"grad_norm": 2.5744450092315674, |
|
"learning_rate": 1.447826086956522e-05, |
|
"loss": 0.4511, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 55.65217391304348, |
|
"grad_norm": 7.211576461791992, |
|
"learning_rate": 1.4434782608695654e-05, |
|
"loss": 0.4599, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.4962967336177826, |
|
"eval_runtime": 4.8613, |
|
"eval_samples_per_second": 52.249, |
|
"eval_steps_per_second": 0.411, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 56.08695652173913, |
|
"grad_norm": 1.932460069656372, |
|
"learning_rate": 1.4391304347826087e-05, |
|
"loss": 0.4168, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 56.52173913043478, |
|
"grad_norm": 5.841196537017822, |
|
"learning_rate": 1.4347826086956522e-05, |
|
"loss": 0.4622, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 56.95652173913044, |
|
"grad_norm": 1.948188066482544, |
|
"learning_rate": 1.4304347826086957e-05, |
|
"loss": 0.4458, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 57.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.49557480216026306, |
|
"eval_runtime": 3.6949, |
|
"eval_samples_per_second": 68.744, |
|
"eval_steps_per_second": 0.541, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 57.391304347826086, |
|
"grad_norm": 4.304020881652832, |
|
"learning_rate": 1.4260869565217392e-05, |
|
"loss": 0.4378, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 57.82608695652174, |
|
"grad_norm": 2.710130453109741, |
|
"learning_rate": 1.4217391304347828e-05, |
|
"loss": 0.4296, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 58.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.49939388036727905, |
|
"eval_runtime": 3.7356, |
|
"eval_samples_per_second": 67.995, |
|
"eval_steps_per_second": 0.535, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 58.26086956521739, |
|
"grad_norm": 3.730140209197998, |
|
"learning_rate": 1.4173913043478263e-05, |
|
"loss": 0.4664, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 58.69565217391305, |
|
"grad_norm": 9.71405029296875, |
|
"learning_rate": 1.4130434782608698e-05, |
|
"loss": 0.4415, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 59.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.4997561573982239, |
|
"eval_runtime": 4.9859, |
|
"eval_samples_per_second": 50.943, |
|
"eval_steps_per_second": 0.401, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 59.130434782608695, |
|
"grad_norm": 2.7752935886383057, |
|
"learning_rate": 1.4086956521739133e-05, |
|
"loss": 0.4436, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 59.56521739130435, |
|
"grad_norm": 4.0491251945495605, |
|
"learning_rate": 1.4043478260869568e-05, |
|
"loss": 0.4442, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"grad_norm": 3.6015145778656006, |
|
"learning_rate": 1.4e-05, |
|
"loss": 0.4036, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.4996122717857361, |
|
"eval_runtime": 3.8039, |
|
"eval_samples_per_second": 66.774, |
|
"eval_steps_per_second": 0.526, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 60.43478260869565, |
|
"grad_norm": 2.5297908782958984, |
|
"learning_rate": 1.3956521739130435e-05, |
|
"loss": 0.4364, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 60.869565217391305, |
|
"grad_norm": 2.8682429790496826, |
|
"learning_rate": 1.391304347826087e-05, |
|
"loss": 0.4406, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 61.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5022182464599609, |
|
"eval_runtime": 3.7333, |
|
"eval_samples_per_second": 68.036, |
|
"eval_steps_per_second": 0.536, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 61.30434782608695, |
|
"grad_norm": 3.3014872074127197, |
|
"learning_rate": 1.3869565217391305e-05, |
|
"loss": 0.4346, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 61.73913043478261, |
|
"grad_norm": 3.4654860496520996, |
|
"learning_rate": 1.382608695652174e-05, |
|
"loss": 0.4235, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 62.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5018435120582581, |
|
"eval_runtime": 4.9797, |
|
"eval_samples_per_second": 51.007, |
|
"eval_steps_per_second": 0.402, |
|
"step": 1426 |
|
}, |
|
{ |
|
"epoch": 62.17391304347826, |
|
"grad_norm": 6.609365940093994, |
|
"learning_rate": 1.3782608695652175e-05, |
|
"loss": 0.4884, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 62.608695652173914, |
|
"grad_norm": 3.425076484680176, |
|
"learning_rate": 1.373913043478261e-05, |
|
"loss": 0.4492, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 63.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.4963783323764801, |
|
"eval_runtime": 3.7677, |
|
"eval_samples_per_second": 67.416, |
|
"eval_steps_per_second": 0.531, |
|
"step": 1449 |
|
}, |
|
{ |
|
"epoch": 63.04347826086956, |
|
"grad_norm": 4.069096088409424, |
|
"learning_rate": 1.3695652173913046e-05, |
|
"loss": 0.4119, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 63.47826086956522, |
|
"grad_norm": 2.3584377765655518, |
|
"learning_rate": 1.3652173913043479e-05, |
|
"loss": 0.4356, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 63.91304347826087, |
|
"grad_norm": 12.776151657104492, |
|
"learning_rate": 1.3608695652173913e-05, |
|
"loss": 0.4065, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.4952879250049591, |
|
"eval_runtime": 3.7751, |
|
"eval_samples_per_second": 67.284, |
|
"eval_steps_per_second": 0.53, |
|
"step": 1472 |
|
}, |
|
{ |
|
"epoch": 64.34782608695652, |
|
"grad_norm": 2.501909017562866, |
|
"learning_rate": 1.3565217391304348e-05, |
|
"loss": 0.4356, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 64.78260869565217, |
|
"grad_norm": 2.650075674057007, |
|
"learning_rate": 1.3521739130434783e-05, |
|
"loss": 0.4474, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.4896911084651947, |
|
"eval_runtime": 5.1494, |
|
"eval_samples_per_second": 49.326, |
|
"eval_steps_per_second": 0.388, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 65.21739130434783, |
|
"grad_norm": 4.465973854064941, |
|
"learning_rate": 1.3478260869565218e-05, |
|
"loss": 0.4246, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 65.65217391304348, |
|
"grad_norm": 2.057035207748413, |
|
"learning_rate": 1.3434782608695653e-05, |
|
"loss": 0.4605, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 66.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5039426684379578, |
|
"eval_runtime": 3.7424, |
|
"eval_samples_per_second": 67.871, |
|
"eval_steps_per_second": 0.534, |
|
"step": 1518 |
|
}, |
|
{ |
|
"epoch": 66.08695652173913, |
|
"grad_norm": 3.064012050628662, |
|
"learning_rate": 1.3391304347826088e-05, |
|
"loss": 0.4157, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 66.52173913043478, |
|
"grad_norm": 3.7584011554718018, |
|
"learning_rate": 1.3347826086956523e-05, |
|
"loss": 0.4553, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 66.95652173913044, |
|
"grad_norm": 2.945054054260254, |
|
"learning_rate": 1.3304347826086958e-05, |
|
"loss": 0.436, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 67.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.5024412274360657, |
|
"eval_runtime": 3.7621, |
|
"eval_samples_per_second": 67.516, |
|
"eval_steps_per_second": 0.532, |
|
"step": 1541 |
|
}, |
|
{ |
|
"epoch": 67.3913043478261, |
|
"grad_norm": 3.1257166862487793, |
|
"learning_rate": 1.3260869565217392e-05, |
|
"loss": 0.4173, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 67.82608695652173, |
|
"grad_norm": 5.225259304046631, |
|
"learning_rate": 1.3217391304347827e-05, |
|
"loss": 0.4746, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5006521344184875, |
|
"eval_runtime": 5.0155, |
|
"eval_samples_per_second": 50.643, |
|
"eval_steps_per_second": 0.399, |
|
"step": 1564 |
|
}, |
|
{ |
|
"epoch": 68.26086956521739, |
|
"grad_norm": 3.3438003063201904, |
|
"learning_rate": 1.3173913043478262e-05, |
|
"loss": 0.3957, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 68.69565217391305, |
|
"grad_norm": 2.6640641689300537, |
|
"learning_rate": 1.3130434782608697e-05, |
|
"loss": 0.4555, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 69.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5053796172142029, |
|
"eval_runtime": 3.7207, |
|
"eval_samples_per_second": 68.266, |
|
"eval_steps_per_second": 0.538, |
|
"step": 1587 |
|
}, |
|
{ |
|
"epoch": 69.1304347826087, |
|
"grad_norm": 6.726771831512451, |
|
"learning_rate": 1.308695652173913e-05, |
|
"loss": 0.4322, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 69.56521739130434, |
|
"grad_norm": 3.231029748916626, |
|
"learning_rate": 1.3043478260869566e-05, |
|
"loss": 0.3949, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"grad_norm": 6.560612201690674, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 0.433, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.49738696217536926, |
|
"eval_runtime": 3.716, |
|
"eval_samples_per_second": 68.353, |
|
"eval_steps_per_second": 0.538, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 70.43478260869566, |
|
"grad_norm": 2.6361474990844727, |
|
"learning_rate": 1.2956521739130436e-05, |
|
"loss": 0.4158, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 70.8695652173913, |
|
"grad_norm": 2.7182960510253906, |
|
"learning_rate": 1.2913043478260871e-05, |
|
"loss": 0.4503, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 71.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.509588897228241, |
|
"eval_runtime": 5.0488, |
|
"eval_samples_per_second": 50.309, |
|
"eval_steps_per_second": 0.396, |
|
"step": 1633 |
|
}, |
|
{ |
|
"epoch": 71.30434782608695, |
|
"grad_norm": 1.938330888748169, |
|
"learning_rate": 1.2869565217391305e-05, |
|
"loss": 0.4285, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 71.73913043478261, |
|
"grad_norm": 3.1797378063201904, |
|
"learning_rate": 1.282608695652174e-05, |
|
"loss": 0.4424, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.5040333867073059, |
|
"eval_runtime": 3.7598, |
|
"eval_samples_per_second": 67.557, |
|
"eval_steps_per_second": 0.532, |
|
"step": 1656 |
|
}, |
|
{ |
|
"epoch": 72.17391304347827, |
|
"grad_norm": 3.028841257095337, |
|
"learning_rate": 1.2782608695652175e-05, |
|
"loss": 0.4278, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 72.6086956521739, |
|
"grad_norm": 3.0137178897857666, |
|
"learning_rate": 1.273913043478261e-05, |
|
"loss": 0.4331, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 73.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.505591869354248, |
|
"eval_runtime": 3.7202, |
|
"eval_samples_per_second": 68.277, |
|
"eval_steps_per_second": 0.538, |
|
"step": 1679 |
|
}, |
|
{ |
|
"epoch": 73.04347826086956, |
|
"grad_norm": 3.519934892654419, |
|
"learning_rate": 1.2695652173913045e-05, |
|
"loss": 0.4523, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 73.47826086956522, |
|
"grad_norm": 2.6839394569396973, |
|
"learning_rate": 1.265217391304348e-05, |
|
"loss": 0.4143, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 73.91304347826087, |
|
"grad_norm": 4.223355770111084, |
|
"learning_rate": 1.2608695652173915e-05, |
|
"loss": 0.4263, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 74.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5025500059127808, |
|
"eval_runtime": 5.0354, |
|
"eval_samples_per_second": 50.442, |
|
"eval_steps_per_second": 0.397, |
|
"step": 1702 |
|
}, |
|
{ |
|
"epoch": 74.34782608695652, |
|
"grad_norm": 2.633610248565674, |
|
"learning_rate": 1.2565217391304349e-05, |
|
"loss": 0.4451, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 74.78260869565217, |
|
"grad_norm": 4.227041721343994, |
|
"learning_rate": 1.2521739130434784e-05, |
|
"loss": 0.4305, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5032832026481628, |
|
"eval_runtime": 3.7074, |
|
"eval_samples_per_second": 68.512, |
|
"eval_steps_per_second": 0.539, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 75.21739130434783, |
|
"grad_norm": 3.885732412338257, |
|
"learning_rate": 1.2478260869565217e-05, |
|
"loss": 0.4177, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 75.65217391304348, |
|
"grad_norm": 6.669870853424072, |
|
"learning_rate": 1.2434782608695652e-05, |
|
"loss": 0.4271, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5014809966087341, |
|
"eval_runtime": 3.6911, |
|
"eval_samples_per_second": 68.814, |
|
"eval_steps_per_second": 0.542, |
|
"step": 1748 |
|
}, |
|
{ |
|
"epoch": 76.08695652173913, |
|
"grad_norm": 1.820388913154602, |
|
"learning_rate": 1.2391304347826088e-05, |
|
"loss": 0.4457, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 76.52173913043478, |
|
"grad_norm": 2.142805337905884, |
|
"learning_rate": 1.2347826086956523e-05, |
|
"loss": 0.3962, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 76.95652173913044, |
|
"grad_norm": 3.5151073932647705, |
|
"learning_rate": 1.2304347826086958e-05, |
|
"loss": 0.4635, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 77.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.49884113669395447, |
|
"eval_runtime": 4.3354, |
|
"eval_samples_per_second": 58.587, |
|
"eval_steps_per_second": 0.461, |
|
"step": 1771 |
|
}, |
|
{ |
|
"epoch": 77.3913043478261, |
|
"grad_norm": 3.867955207824707, |
|
"learning_rate": 1.2260869565217393e-05, |
|
"loss": 0.4616, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 77.82608695652173, |
|
"grad_norm": 2.6050870418548584, |
|
"learning_rate": 1.2217391304347828e-05, |
|
"loss": 0.4212, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 78.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.4993511438369751, |
|
"eval_runtime": 4.267, |
|
"eval_samples_per_second": 59.527, |
|
"eval_steps_per_second": 0.469, |
|
"step": 1794 |
|
}, |
|
{ |
|
"epoch": 78.26086956521739, |
|
"grad_norm": 2.1961538791656494, |
|
"learning_rate": 1.2173913043478263e-05, |
|
"loss": 0.4191, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 78.69565217391305, |
|
"grad_norm": 6.02454948425293, |
|
"learning_rate": 1.2130434782608698e-05, |
|
"loss": 0.4154, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 79.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5044043660163879, |
|
"eval_runtime": 3.7036, |
|
"eval_samples_per_second": 68.581, |
|
"eval_steps_per_second": 0.54, |
|
"step": 1817 |
|
}, |
|
{ |
|
"epoch": 79.1304347826087, |
|
"grad_norm": 2.1048858165740967, |
|
"learning_rate": 1.208695652173913e-05, |
|
"loss": 0.4196, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 79.56521739130434, |
|
"grad_norm": 2.8622193336486816, |
|
"learning_rate": 1.2043478260869565e-05, |
|
"loss": 0.4314, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"grad_norm": 6.1558427810668945, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.4288, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5033003687858582, |
|
"eval_runtime": 3.7575, |
|
"eval_samples_per_second": 67.598, |
|
"eval_steps_per_second": 0.532, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 80.43478260869566, |
|
"grad_norm": 3.3254945278167725, |
|
"learning_rate": 1.1956521739130435e-05, |
|
"loss": 0.4297, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 80.8695652173913, |
|
"grad_norm": 2.2818620204925537, |
|
"learning_rate": 1.191304347826087e-05, |
|
"loss": 0.4211, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 81.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5050157904624939, |
|
"eval_runtime": 5.0113, |
|
"eval_samples_per_second": 50.685, |
|
"eval_steps_per_second": 0.399, |
|
"step": 1863 |
|
}, |
|
{ |
|
"epoch": 81.30434782608695, |
|
"grad_norm": 4.174459934234619, |
|
"learning_rate": 1.1869565217391306e-05, |
|
"loss": 0.4229, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 81.73913043478261, |
|
"grad_norm": 2.87514066696167, |
|
"learning_rate": 1.182608695652174e-05, |
|
"loss": 0.4022, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 82.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5021248459815979, |
|
"eval_runtime": 3.7629, |
|
"eval_samples_per_second": 67.5, |
|
"eval_steps_per_second": 0.531, |
|
"step": 1886 |
|
}, |
|
{ |
|
"epoch": 82.17391304347827, |
|
"grad_norm": 5.307149410247803, |
|
"learning_rate": 1.1782608695652176e-05, |
|
"loss": 0.4564, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 82.6086956521739, |
|
"grad_norm": 4.411511421203613, |
|
"learning_rate": 1.1739130434782611e-05, |
|
"loss": 0.4477, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 83.0, |
|
"eval_accuracy": 0.7755905511811023, |
|
"eval_loss": 0.509568452835083, |
|
"eval_runtime": 5.3605, |
|
"eval_samples_per_second": 47.384, |
|
"eval_steps_per_second": 0.373, |
|
"step": 1909 |
|
}, |
|
{ |
|
"epoch": 83.04347826086956, |
|
"grad_norm": 2.478482246398926, |
|
"learning_rate": 1.1695652173913043e-05, |
|
"loss": 0.4118, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 83.47826086956522, |
|
"grad_norm": 2.000185012817383, |
|
"learning_rate": 1.1652173913043478e-05, |
|
"loss": 0.4486, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 83.91304347826087, |
|
"grad_norm": 4.231175422668457, |
|
"learning_rate": 1.1608695652173913e-05, |
|
"loss": 0.4091, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.501672625541687, |
|
"eval_runtime": 4.6714, |
|
"eval_samples_per_second": 54.374, |
|
"eval_steps_per_second": 0.428, |
|
"step": 1932 |
|
}, |
|
{ |
|
"epoch": 84.34782608695652, |
|
"grad_norm": 8.062799453735352, |
|
"learning_rate": 1.1565217391304348e-05, |
|
"loss": 0.4108, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 84.78260869565217, |
|
"grad_norm": 3.525912046432495, |
|
"learning_rate": 1.1521739130434783e-05, |
|
"loss": 0.4284, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5094006061553955, |
|
"eval_runtime": 3.7057, |
|
"eval_samples_per_second": 68.544, |
|
"eval_steps_per_second": 0.54, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 85.21739130434783, |
|
"grad_norm": 2.8294172286987305, |
|
"learning_rate": 1.1478260869565218e-05, |
|
"loss": 0.4341, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 85.65217391304348, |
|
"grad_norm": 2.6164603233337402, |
|
"learning_rate": 1.1434782608695654e-05, |
|
"loss": 0.4317, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 86.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5055702328681946, |
|
"eval_runtime": 3.78, |
|
"eval_samples_per_second": 67.195, |
|
"eval_steps_per_second": 0.529, |
|
"step": 1978 |
|
}, |
|
{ |
|
"epoch": 86.08695652173913, |
|
"grad_norm": 5.29531717300415, |
|
"learning_rate": 1.1391304347826089e-05, |
|
"loss": 0.3842, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 86.52173913043478, |
|
"grad_norm": 3.8016159534454346, |
|
"learning_rate": 1.1347826086956524e-05, |
|
"loss": 0.4294, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 86.95652173913044, |
|
"grad_norm": 2.229055643081665, |
|
"learning_rate": 1.1304347826086957e-05, |
|
"loss": 0.4011, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 87.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.4991566836833954, |
|
"eval_runtime": 4.9616, |
|
"eval_samples_per_second": 51.194, |
|
"eval_steps_per_second": 0.403, |
|
"step": 2001 |
|
}, |
|
{ |
|
"epoch": 87.3913043478261, |
|
"grad_norm": 4.449975490570068, |
|
"learning_rate": 1.1260869565217392e-05, |
|
"loss": 0.4413, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 87.82608695652173, |
|
"grad_norm": 3.4843342304229736, |
|
"learning_rate": 1.1217391304347827e-05, |
|
"loss": 0.4043, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5106358528137207, |
|
"eval_runtime": 3.759, |
|
"eval_samples_per_second": 67.571, |
|
"eval_steps_per_second": 0.532, |
|
"step": 2024 |
|
}, |
|
{ |
|
"epoch": 88.26086956521739, |
|
"grad_norm": 3.2311477661132812, |
|
"learning_rate": 1.1173913043478261e-05, |
|
"loss": 0.4127, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 88.69565217391305, |
|
"grad_norm": 3.511033058166504, |
|
"learning_rate": 1.1130434782608696e-05, |
|
"loss": 0.4233, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 89.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5082967877388, |
|
"eval_runtime": 3.7279, |
|
"eval_samples_per_second": 68.135, |
|
"eval_steps_per_second": 0.536, |
|
"step": 2047 |
|
}, |
|
{ |
|
"epoch": 89.1304347826087, |
|
"grad_norm": 3.1737847328186035, |
|
"learning_rate": 1.1086956521739131e-05, |
|
"loss": 0.4449, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 89.56521739130434, |
|
"grad_norm": 3.3332552909851074, |
|
"learning_rate": 1.1043478260869566e-05, |
|
"loss": 0.4148, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"grad_norm": 5.011209487915039, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"loss": 0.4383, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5015798211097717, |
|
"eval_runtime": 5.0126, |
|
"eval_samples_per_second": 50.672, |
|
"eval_steps_per_second": 0.399, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 90.43478260869566, |
|
"grad_norm": 2.4368808269500732, |
|
"learning_rate": 1.0956521739130435e-05, |
|
"loss": 0.4133, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 90.8695652173913, |
|
"grad_norm": 5.885110378265381, |
|
"learning_rate": 1.091304347826087e-05, |
|
"loss": 0.4328, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 91.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5062097311019897, |
|
"eval_runtime": 3.7299, |
|
"eval_samples_per_second": 68.098, |
|
"eval_steps_per_second": 0.536, |
|
"step": 2093 |
|
}, |
|
{ |
|
"epoch": 91.30434782608695, |
|
"grad_norm": 2.2072901725769043, |
|
"learning_rate": 1.0869565217391305e-05, |
|
"loss": 0.4137, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 91.73913043478261, |
|
"grad_norm": 1.9848076105117798, |
|
"learning_rate": 1.082608695652174e-05, |
|
"loss": 0.3978, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5026075839996338, |
|
"eval_runtime": 3.7759, |
|
"eval_samples_per_second": 67.268, |
|
"eval_steps_per_second": 0.53, |
|
"step": 2116 |
|
}, |
|
{ |
|
"epoch": 92.17391304347827, |
|
"grad_norm": 3.738398313522339, |
|
"learning_rate": 1.0782608695652175e-05, |
|
"loss": 0.4459, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 92.6086956521739, |
|
"grad_norm": 3.0096168518066406, |
|
"learning_rate": 1.073913043478261e-05, |
|
"loss": 0.4052, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 93.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.49642127752304077, |
|
"eval_runtime": 5.0236, |
|
"eval_samples_per_second": 50.562, |
|
"eval_steps_per_second": 0.398, |
|
"step": 2139 |
|
}, |
|
{ |
|
"epoch": 93.04347826086956, |
|
"grad_norm": 3.7452170848846436, |
|
"learning_rate": 1.0695652173913046e-05, |
|
"loss": 0.4205, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 93.47826086956522, |
|
"grad_norm": 3.8985049724578857, |
|
"learning_rate": 1.0652173913043479e-05, |
|
"loss": 0.4171, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 93.91304347826087, |
|
"grad_norm": 2.283020496368408, |
|
"learning_rate": 1.0608695652173914e-05, |
|
"loss": 0.3938, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 94.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5036487579345703, |
|
"eval_runtime": 3.7298, |
|
"eval_samples_per_second": 68.101, |
|
"eval_steps_per_second": 0.536, |
|
"step": 2162 |
|
}, |
|
{ |
|
"epoch": 94.34782608695652, |
|
"grad_norm": 7.054046630859375, |
|
"learning_rate": 1.0565217391304348e-05, |
|
"loss": 0.4336, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 94.78260869565217, |
|
"grad_norm": 3.131002902984619, |
|
"learning_rate": 1.0521739130434783e-05, |
|
"loss": 0.393, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5102458596229553, |
|
"eval_runtime": 3.6839, |
|
"eval_samples_per_second": 68.949, |
|
"eval_steps_per_second": 0.543, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 95.21739130434783, |
|
"grad_norm": 2.4622268676757812, |
|
"learning_rate": 1.0478260869565218e-05, |
|
"loss": 0.3997, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 95.65217391304348, |
|
"grad_norm": 3.815375566482544, |
|
"learning_rate": 1.0434782608695653e-05, |
|
"loss": 0.4294, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5002910494804382, |
|
"eval_runtime": 4.8997, |
|
"eval_samples_per_second": 51.84, |
|
"eval_steps_per_second": 0.408, |
|
"step": 2208 |
|
}, |
|
{ |
|
"epoch": 96.08695652173913, |
|
"grad_norm": 8.787290573120117, |
|
"learning_rate": 1.0391304347826088e-05, |
|
"loss": 0.4155, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 96.52173913043478, |
|
"grad_norm": 2.8499906063079834, |
|
"learning_rate": 1.0347826086956523e-05, |
|
"loss": 0.4095, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 96.95652173913044, |
|
"grad_norm": 6.26355504989624, |
|
"learning_rate": 1.0304347826086958e-05, |
|
"loss": 0.4122, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 97.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5013226270675659, |
|
"eval_runtime": 3.7744, |
|
"eval_samples_per_second": 67.296, |
|
"eval_steps_per_second": 0.53, |
|
"step": 2231 |
|
}, |
|
{ |
|
"epoch": 97.3913043478261, |
|
"grad_norm": 3.257772445678711, |
|
"learning_rate": 1.0260869565217393e-05, |
|
"loss": 0.3522, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 97.82608695652173, |
|
"grad_norm": 2.788611888885498, |
|
"learning_rate": 1.0217391304347829e-05, |
|
"loss": 0.4207, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 98.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.507587730884552, |
|
"eval_runtime": 3.8534, |
|
"eval_samples_per_second": 65.915, |
|
"eval_steps_per_second": 0.519, |
|
"step": 2254 |
|
}, |
|
{ |
|
"epoch": 98.26086956521739, |
|
"grad_norm": 2.974043846130371, |
|
"learning_rate": 1.017391304347826e-05, |
|
"loss": 0.4352, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 98.69565217391305, |
|
"grad_norm": 3.231869697570801, |
|
"learning_rate": 1.0130434782608695e-05, |
|
"loss": 0.4127, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 99.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.503979504108429, |
|
"eval_runtime": 5.0998, |
|
"eval_samples_per_second": 49.806, |
|
"eval_steps_per_second": 0.392, |
|
"step": 2277 |
|
}, |
|
{ |
|
"epoch": 99.1304347826087, |
|
"grad_norm": 2.597999095916748, |
|
"learning_rate": 1.008695652173913e-05, |
|
"loss": 0.3888, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 99.56521739130434, |
|
"grad_norm": 4.4219889640808105, |
|
"learning_rate": 1.0043478260869566e-05, |
|
"loss": 0.3921, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"grad_norm": 4.641758441925049, |
|
"learning_rate": 1e-05, |
|
"loss": 0.441, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.502194881439209, |
|
"eval_runtime": 3.7077, |
|
"eval_samples_per_second": 68.506, |
|
"eval_steps_per_second": 0.539, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 100.43478260869566, |
|
"grad_norm": 2.948529005050659, |
|
"learning_rate": 9.956521739130436e-06, |
|
"loss": 0.4324, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 100.8695652173913, |
|
"grad_norm": 2.4855594635009766, |
|
"learning_rate": 9.913043478260871e-06, |
|
"loss": 0.3938, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 101.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.4974897503852844, |
|
"eval_runtime": 3.7364, |
|
"eval_samples_per_second": 67.98, |
|
"eval_steps_per_second": 0.535, |
|
"step": 2323 |
|
}, |
|
{ |
|
"epoch": 101.30434782608695, |
|
"grad_norm": 4.753269195556641, |
|
"learning_rate": 9.869565217391304e-06, |
|
"loss": 0.3918, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 101.73913043478261, |
|
"grad_norm": 5.000470161437988, |
|
"learning_rate": 9.82608695652174e-06, |
|
"loss": 0.4109, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 102.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5018798112869263, |
|
"eval_runtime": 4.8425, |
|
"eval_samples_per_second": 52.452, |
|
"eval_steps_per_second": 0.413, |
|
"step": 2346 |
|
}, |
|
{ |
|
"epoch": 102.17391304347827, |
|
"grad_norm": 2.8584697246551514, |
|
"learning_rate": 9.782608695652175e-06, |
|
"loss": 0.4199, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 102.6086956521739, |
|
"grad_norm": 2.773083448410034, |
|
"learning_rate": 9.73913043478261e-06, |
|
"loss": 0.4299, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 103.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5060404539108276, |
|
"eval_runtime": 3.7179, |
|
"eval_samples_per_second": 68.318, |
|
"eval_steps_per_second": 0.538, |
|
"step": 2369 |
|
}, |
|
{ |
|
"epoch": 103.04347826086956, |
|
"grad_norm": 1.847158670425415, |
|
"learning_rate": 9.695652173913043e-06, |
|
"loss": 0.3834, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 103.47826086956522, |
|
"grad_norm": 4.114128112792969, |
|
"learning_rate": 9.652173913043478e-06, |
|
"loss": 0.4061, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 103.91304347826087, |
|
"grad_norm": 5.080406665802002, |
|
"learning_rate": 9.608695652173914e-06, |
|
"loss": 0.4148, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 104.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5038026571273804, |
|
"eval_runtime": 3.7535, |
|
"eval_samples_per_second": 67.671, |
|
"eval_steps_per_second": 0.533, |
|
"step": 2392 |
|
}, |
|
{ |
|
"epoch": 104.34782608695652, |
|
"grad_norm": 3.291896104812622, |
|
"learning_rate": 9.565217391304349e-06, |
|
"loss": 0.4272, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 104.78260869565217, |
|
"grad_norm": 2.7959041595458984, |
|
"learning_rate": 9.521739130434784e-06, |
|
"loss": 0.4179, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 105.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5064316391944885, |
|
"eval_runtime": 4.8627, |
|
"eval_samples_per_second": 52.235, |
|
"eval_steps_per_second": 0.411, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 105.21739130434783, |
|
"grad_norm": 5.880518913269043, |
|
"learning_rate": 9.478260869565217e-06, |
|
"loss": 0.4155, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 105.65217391304348, |
|
"grad_norm": 2.2435200214385986, |
|
"learning_rate": 9.434782608695652e-06, |
|
"loss": 0.4352, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 106.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5059410929679871, |
|
"eval_runtime": 3.7149, |
|
"eval_samples_per_second": 68.373, |
|
"eval_steps_per_second": 0.538, |
|
"step": 2438 |
|
}, |
|
{ |
|
"epoch": 106.08695652173913, |
|
"grad_norm": 3.1865811347961426, |
|
"learning_rate": 9.391304347826087e-06, |
|
"loss": 0.3997, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 106.52173913043478, |
|
"grad_norm": 4.0479936599731445, |
|
"learning_rate": 9.347826086956523e-06, |
|
"loss": 0.401, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 106.95652173913044, |
|
"grad_norm": 2.87663197517395, |
|
"learning_rate": 9.304347826086956e-06, |
|
"loss": 0.4027, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 107.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5025486350059509, |
|
"eval_runtime": 3.7614, |
|
"eval_samples_per_second": 67.528, |
|
"eval_steps_per_second": 0.532, |
|
"step": 2461 |
|
}, |
|
{ |
|
"epoch": 107.3913043478261, |
|
"grad_norm": 2.630986452102661, |
|
"learning_rate": 9.260869565217391e-06, |
|
"loss": 0.3828, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 107.82608695652173, |
|
"grad_norm": 2.9700822830200195, |
|
"learning_rate": 9.217391304347826e-06, |
|
"loss": 0.4002, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 108.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5020495653152466, |
|
"eval_runtime": 4.6331, |
|
"eval_samples_per_second": 54.823, |
|
"eval_steps_per_second": 0.432, |
|
"step": 2484 |
|
}, |
|
{ |
|
"epoch": 108.26086956521739, |
|
"grad_norm": 4.361221790313721, |
|
"learning_rate": 9.173913043478261e-06, |
|
"loss": 0.405, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 108.69565217391305, |
|
"grad_norm": 2.9328296184539795, |
|
"learning_rate": 9.130434782608697e-06, |
|
"loss": 0.3988, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 109.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5063354969024658, |
|
"eval_runtime": 3.8012, |
|
"eval_samples_per_second": 66.821, |
|
"eval_steps_per_second": 0.526, |
|
"step": 2507 |
|
}, |
|
{ |
|
"epoch": 109.1304347826087, |
|
"grad_norm": 2.3236513137817383, |
|
"learning_rate": 9.086956521739132e-06, |
|
"loss": 0.3894, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 109.56521739130434, |
|
"grad_norm": 3.4379804134368896, |
|
"learning_rate": 9.043478260869565e-06, |
|
"loss": 0.4023, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"grad_norm": 4.300137042999268, |
|
"learning_rate": 9e-06, |
|
"loss": 0.4095, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5034452676773071, |
|
"eval_runtime": 3.7021, |
|
"eval_samples_per_second": 68.61, |
|
"eval_steps_per_second": 0.54, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 110.43478260869566, |
|
"grad_norm": 2.190524101257324, |
|
"learning_rate": 8.956521739130435e-06, |
|
"loss": 0.4072, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 110.8695652173913, |
|
"grad_norm": 2.2291879653930664, |
|
"learning_rate": 8.91304347826087e-06, |
|
"loss": 0.4001, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 111.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.505436360836029, |
|
"eval_runtime": 4.2919, |
|
"eval_samples_per_second": 59.182, |
|
"eval_steps_per_second": 0.466, |
|
"step": 2553 |
|
}, |
|
{ |
|
"epoch": 111.30434782608695, |
|
"grad_norm": 3.1182541847229004, |
|
"learning_rate": 8.869565217391306e-06, |
|
"loss": 0.3904, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 111.73913043478261, |
|
"grad_norm": 3.8375625610351562, |
|
"learning_rate": 8.82608695652174e-06, |
|
"loss": 0.4201, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 112.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5076125860214233, |
|
"eval_runtime": 4.2691, |
|
"eval_samples_per_second": 59.497, |
|
"eval_steps_per_second": 0.468, |
|
"step": 2576 |
|
}, |
|
{ |
|
"epoch": 112.17391304347827, |
|
"grad_norm": 2.4231808185577393, |
|
"learning_rate": 8.782608695652174e-06, |
|
"loss": 0.3925, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 112.6086956521739, |
|
"grad_norm": 4.854309558868408, |
|
"learning_rate": 8.73913043478261e-06, |
|
"loss": 0.4134, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 113.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5069688558578491, |
|
"eval_runtime": 3.7367, |
|
"eval_samples_per_second": 67.974, |
|
"eval_steps_per_second": 0.535, |
|
"step": 2599 |
|
}, |
|
{ |
|
"epoch": 113.04347826086956, |
|
"grad_norm": 4.327704429626465, |
|
"learning_rate": 8.695652173913044e-06, |
|
"loss": 0.3959, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 113.47826086956522, |
|
"grad_norm": 2.8718910217285156, |
|
"learning_rate": 8.65217391304348e-06, |
|
"loss": 0.3806, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 113.91304347826087, |
|
"grad_norm": 5.400497913360596, |
|
"learning_rate": 8.608695652173915e-06, |
|
"loss": 0.3614, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 114.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5032684206962585, |
|
"eval_runtime": 4.4717, |
|
"eval_samples_per_second": 56.802, |
|
"eval_steps_per_second": 0.447, |
|
"step": 2622 |
|
}, |
|
{ |
|
"epoch": 114.34782608695652, |
|
"grad_norm": 2.7276597023010254, |
|
"learning_rate": 8.56521739130435e-06, |
|
"loss": 0.3956, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 114.78260869565217, |
|
"grad_norm": 3.339860200881958, |
|
"learning_rate": 8.521739130434783e-06, |
|
"loss": 0.3928, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 115.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5042973160743713, |
|
"eval_runtime": 4.1216, |
|
"eval_samples_per_second": 61.627, |
|
"eval_steps_per_second": 0.485, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 115.21739130434783, |
|
"grad_norm": 2.435579538345337, |
|
"learning_rate": 8.478260869565218e-06, |
|
"loss": 0.4149, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 115.65217391304348, |
|
"grad_norm": 3.9001612663269043, |
|
"learning_rate": 8.434782608695653e-06, |
|
"loss": 0.435, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 116.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.4998602271080017, |
|
"eval_runtime": 3.7703, |
|
"eval_samples_per_second": 67.369, |
|
"eval_steps_per_second": 0.53, |
|
"step": 2668 |
|
}, |
|
{ |
|
"epoch": 116.08695652173913, |
|
"grad_norm": 4.031954288482666, |
|
"learning_rate": 8.391304347826089e-06, |
|
"loss": 0.3575, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 116.52173913043478, |
|
"grad_norm": 3.1172120571136475, |
|
"learning_rate": 8.347826086956522e-06, |
|
"loss": 0.4062, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 116.95652173913044, |
|
"grad_norm": 2.6061761379241943, |
|
"learning_rate": 8.304347826086957e-06, |
|
"loss": 0.4162, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 117.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5132189393043518, |
|
"eval_runtime": 4.4279, |
|
"eval_samples_per_second": 57.363, |
|
"eval_steps_per_second": 0.452, |
|
"step": 2691 |
|
}, |
|
{ |
|
"epoch": 117.3913043478261, |
|
"grad_norm": 1.7457960844039917, |
|
"learning_rate": 8.260869565217392e-06, |
|
"loss": 0.3887, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 117.82608695652173, |
|
"grad_norm": 5.013397216796875, |
|
"learning_rate": 8.217391304347827e-06, |
|
"loss": 0.4078, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 118.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5088200569152832, |
|
"eval_runtime": 4.1897, |
|
"eval_samples_per_second": 60.625, |
|
"eval_steps_per_second": 0.477, |
|
"step": 2714 |
|
}, |
|
{ |
|
"epoch": 118.26086956521739, |
|
"grad_norm": 3.4758872985839844, |
|
"learning_rate": 8.173913043478263e-06, |
|
"loss": 0.4251, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 118.69565217391305, |
|
"grad_norm": 1.8225319385528564, |
|
"learning_rate": 8.130434782608696e-06, |
|
"loss": 0.4025, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 119.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.507527768611908, |
|
"eval_runtime": 3.721, |
|
"eval_samples_per_second": 68.261, |
|
"eval_steps_per_second": 0.537, |
|
"step": 2737 |
|
}, |
|
{ |
|
"epoch": 119.1304347826087, |
|
"grad_norm": 4.636626720428467, |
|
"learning_rate": 8.086956521739131e-06, |
|
"loss": 0.4024, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 119.56521739130434, |
|
"grad_norm": 2.249758720397949, |
|
"learning_rate": 8.043478260869566e-06, |
|
"loss": 0.3917, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"grad_norm": 6.408204555511475, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.4096, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.502310574054718, |
|
"eval_runtime": 4.469, |
|
"eval_samples_per_second": 56.836, |
|
"eval_steps_per_second": 0.448, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 120.43478260869566, |
|
"grad_norm": 2.495302200317383, |
|
"learning_rate": 7.956521739130435e-06, |
|
"loss": 0.3791, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 120.8695652173913, |
|
"grad_norm": 2.840449571609497, |
|
"learning_rate": 7.91304347826087e-06, |
|
"loss": 0.3879, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 121.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5062641501426697, |
|
"eval_runtime": 3.9315, |
|
"eval_samples_per_second": 64.606, |
|
"eval_steps_per_second": 0.509, |
|
"step": 2783 |
|
}, |
|
{ |
|
"epoch": 121.30434782608695, |
|
"grad_norm": 4.82555627822876, |
|
"learning_rate": 7.869565217391305e-06, |
|
"loss": 0.4232, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 121.73913043478261, |
|
"grad_norm": 3.220736503601074, |
|
"learning_rate": 7.82608695652174e-06, |
|
"loss": 0.4033, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 122.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5001329183578491, |
|
"eval_runtime": 3.7903, |
|
"eval_samples_per_second": 67.012, |
|
"eval_steps_per_second": 0.528, |
|
"step": 2806 |
|
}, |
|
{ |
|
"epoch": 122.17391304347827, |
|
"grad_norm": 4.516547203063965, |
|
"learning_rate": 7.782608695652174e-06, |
|
"loss": 0.4144, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 122.6086956521739, |
|
"grad_norm": 2.559272289276123, |
|
"learning_rate": 7.739130434782609e-06, |
|
"loss": 0.3927, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 123.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5087068676948547, |
|
"eval_runtime": 4.404, |
|
"eval_samples_per_second": 57.675, |
|
"eval_steps_per_second": 0.454, |
|
"step": 2829 |
|
}, |
|
{ |
|
"epoch": 123.04347826086956, |
|
"grad_norm": 3.344332695007324, |
|
"learning_rate": 7.695652173913044e-06, |
|
"loss": 0.4016, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 123.47826086956522, |
|
"grad_norm": 2.610856533050537, |
|
"learning_rate": 7.652173913043479e-06, |
|
"loss": 0.3925, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 123.91304347826087, |
|
"grad_norm": 3.501596689224243, |
|
"learning_rate": 7.608695652173914e-06, |
|
"loss": 0.3803, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 124.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5149940848350525, |
|
"eval_runtime": 4.2503, |
|
"eval_samples_per_second": 59.76, |
|
"eval_steps_per_second": 0.471, |
|
"step": 2852 |
|
}, |
|
{ |
|
"epoch": 124.34782608695652, |
|
"grad_norm": 4.040353298187256, |
|
"learning_rate": 7.565217391304348e-06, |
|
"loss": 0.4101, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 124.78260869565217, |
|
"grad_norm": 3.1806752681732178, |
|
"learning_rate": 7.5217391304347835e-06, |
|
"loss": 0.4248, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.515027642250061, |
|
"eval_runtime": 3.7006, |
|
"eval_samples_per_second": 68.638, |
|
"eval_steps_per_second": 0.54, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 125.21739130434783, |
|
"grad_norm": 2.976123332977295, |
|
"learning_rate": 7.478260869565218e-06, |
|
"loss": 0.3806, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 125.65217391304348, |
|
"grad_norm": 4.0399250984191895, |
|
"learning_rate": 7.434782608695653e-06, |
|
"loss": 0.3874, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 126.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5157892107963562, |
|
"eval_runtime": 3.8292, |
|
"eval_samples_per_second": 66.332, |
|
"eval_steps_per_second": 0.522, |
|
"step": 2898 |
|
}, |
|
{ |
|
"epoch": 126.08695652173913, |
|
"grad_norm": 2.8186984062194824, |
|
"learning_rate": 7.391304347826087e-06, |
|
"loss": 0.4068, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 126.52173913043478, |
|
"grad_norm": 1.7811031341552734, |
|
"learning_rate": 7.347826086956522e-06, |
|
"loss": 0.4188, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 126.95652173913044, |
|
"grad_norm": 2.591479539871216, |
|
"learning_rate": 7.304347826086957e-06, |
|
"loss": 0.3646, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 127.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.4979710578918457, |
|
"eval_runtime": 4.9476, |
|
"eval_samples_per_second": 51.338, |
|
"eval_steps_per_second": 0.404, |
|
"step": 2921 |
|
}, |
|
{ |
|
"epoch": 127.3913043478261, |
|
"grad_norm": 3.097064733505249, |
|
"learning_rate": 7.2608695652173925e-06, |
|
"loss": 0.3809, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 127.82608695652173, |
|
"grad_norm": 4.5358805656433105, |
|
"learning_rate": 7.217391304347827e-06, |
|
"loss": 0.4115, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 128.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.507692813873291, |
|
"eval_runtime": 3.7086, |
|
"eval_samples_per_second": 68.489, |
|
"eval_steps_per_second": 0.539, |
|
"step": 2944 |
|
}, |
|
{ |
|
"epoch": 128.2608695652174, |
|
"grad_norm": 4.192093372344971, |
|
"learning_rate": 7.173913043478261e-06, |
|
"loss": 0.3931, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 128.69565217391303, |
|
"grad_norm": 2.4763779640197754, |
|
"learning_rate": 7.130434782608696e-06, |
|
"loss": 0.385, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 129.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5153175592422485, |
|
"eval_runtime": 3.726, |
|
"eval_samples_per_second": 68.17, |
|
"eval_steps_per_second": 0.537, |
|
"step": 2967 |
|
}, |
|
{ |
|
"epoch": 129.1304347826087, |
|
"grad_norm": 2.906510353088379, |
|
"learning_rate": 7.086956521739131e-06, |
|
"loss": 0.4009, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 129.56521739130434, |
|
"grad_norm": 5.497567653656006, |
|
"learning_rate": 7.0434782608695665e-06, |
|
"loss": 0.4091, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 130.0, |
|
"grad_norm": 4.277368068695068, |
|
"learning_rate": 7e-06, |
|
"loss": 0.4064, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 130.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.511443018913269, |
|
"eval_runtime": 5.0723, |
|
"eval_samples_per_second": 50.076, |
|
"eval_steps_per_second": 0.394, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 130.43478260869566, |
|
"grad_norm": 2.3368613719940186, |
|
"learning_rate": 6.956521739130435e-06, |
|
"loss": 0.3762, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 130.8695652173913, |
|
"grad_norm": 2.983280897140503, |
|
"learning_rate": 6.91304347826087e-06, |
|
"loss": 0.4168, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 131.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5056832432746887, |
|
"eval_runtime": 3.7472, |
|
"eval_samples_per_second": 67.785, |
|
"eval_steps_per_second": 0.534, |
|
"step": 3013 |
|
}, |
|
{ |
|
"epoch": 131.30434782608697, |
|
"grad_norm": 5.6472978591918945, |
|
"learning_rate": 6.869565217391305e-06, |
|
"loss": 0.3454, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 131.7391304347826, |
|
"grad_norm": 2.710934638977051, |
|
"learning_rate": 6.8260869565217395e-06, |
|
"loss": 0.4319, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 132.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5041180849075317, |
|
"eval_runtime": 3.7165, |
|
"eval_samples_per_second": 68.344, |
|
"eval_steps_per_second": 0.538, |
|
"step": 3036 |
|
}, |
|
{ |
|
"epoch": 132.17391304347825, |
|
"grad_norm": 2.8998305797576904, |
|
"learning_rate": 6.782608695652174e-06, |
|
"loss": 0.3769, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 132.6086956521739, |
|
"grad_norm": 3.503068208694458, |
|
"learning_rate": 6.739130434782609e-06, |
|
"loss": 0.4234, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 133.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5119389891624451, |
|
"eval_runtime": 4.8197, |
|
"eval_samples_per_second": 52.701, |
|
"eval_steps_per_second": 0.415, |
|
"step": 3059 |
|
}, |
|
{ |
|
"epoch": 133.04347826086956, |
|
"grad_norm": 2.628817319869995, |
|
"learning_rate": 6.695652173913044e-06, |
|
"loss": 0.3984, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 133.47826086956522, |
|
"grad_norm": 3.1060750484466553, |
|
"learning_rate": 6.652173913043479e-06, |
|
"loss": 0.4147, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 133.91304347826087, |
|
"grad_norm": 6.7668328285217285, |
|
"learning_rate": 6.6086956521739135e-06, |
|
"loss": 0.3721, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 134.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.51175457239151, |
|
"eval_runtime": 3.7909, |
|
"eval_samples_per_second": 67.003, |
|
"eval_steps_per_second": 0.528, |
|
"step": 3082 |
|
}, |
|
{ |
|
"epoch": 134.34782608695653, |
|
"grad_norm": 6.763729572296143, |
|
"learning_rate": 6.565217391304349e-06, |
|
"loss": 0.386, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 134.7826086956522, |
|
"grad_norm": 4.876804828643799, |
|
"learning_rate": 6.521739130434783e-06, |
|
"loss": 0.3709, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 135.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5078221559524536, |
|
"eval_runtime": 3.7684, |
|
"eval_samples_per_second": 67.402, |
|
"eval_steps_per_second": 0.531, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 135.2173913043478, |
|
"grad_norm": 3.7445313930511475, |
|
"learning_rate": 6.478260869565218e-06, |
|
"loss": 0.3592, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 135.65217391304347, |
|
"grad_norm": 5.715231418609619, |
|
"learning_rate": 6.434782608695652e-06, |
|
"loss": 0.4149, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 136.0, |
|
"eval_accuracy": 0.7795275590551181, |
|
"eval_loss": 0.5163589715957642, |
|
"eval_runtime": 4.6746, |
|
"eval_samples_per_second": 54.336, |
|
"eval_steps_per_second": 0.428, |
|
"step": 3128 |
|
}, |
|
{ |
|
"epoch": 136.08695652173913, |
|
"grad_norm": 3.3850629329681396, |
|
"learning_rate": 6.391304347826087e-06, |
|
"loss": 0.3681, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 136.52173913043478, |
|
"grad_norm": 5.502380847930908, |
|
"learning_rate": 6.3478260869565225e-06, |
|
"loss": 0.3629, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 136.95652173913044, |
|
"grad_norm": 4.158088684082031, |
|
"learning_rate": 6.304347826086958e-06, |
|
"loss": 0.416, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 137.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5123007297515869, |
|
"eval_runtime": 3.789, |
|
"eval_samples_per_second": 67.036, |
|
"eval_steps_per_second": 0.528, |
|
"step": 3151 |
|
}, |
|
{ |
|
"epoch": 137.3913043478261, |
|
"grad_norm": 2.241478681564331, |
|
"learning_rate": 6.260869565217392e-06, |
|
"loss": 0.4089, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 137.82608695652175, |
|
"grad_norm": 4.336514472961426, |
|
"learning_rate": 6.217391304347826e-06, |
|
"loss": 0.406, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 138.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5115824937820435, |
|
"eval_runtime": 3.7195, |
|
"eval_samples_per_second": 68.288, |
|
"eval_steps_per_second": 0.538, |
|
"step": 3174 |
|
}, |
|
{ |
|
"epoch": 138.2608695652174, |
|
"grad_norm": 2.154179334640503, |
|
"learning_rate": 6.173913043478261e-06, |
|
"loss": 0.4018, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 138.69565217391303, |
|
"grad_norm": 3.2215845584869385, |
|
"learning_rate": 6.1304347826086965e-06, |
|
"loss": 0.3613, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 139.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5169662237167358, |
|
"eval_runtime": 4.4593, |
|
"eval_samples_per_second": 56.959, |
|
"eval_steps_per_second": 0.448, |
|
"step": 3197 |
|
}, |
|
{ |
|
"epoch": 139.1304347826087, |
|
"grad_norm": 2.800915241241455, |
|
"learning_rate": 6.086956521739132e-06, |
|
"loss": 0.3863, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 139.56521739130434, |
|
"grad_norm": 7.433578014373779, |
|
"learning_rate": 6.043478260869565e-06, |
|
"loss": 0.4278, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"grad_norm": 3.887300968170166, |
|
"learning_rate": 6e-06, |
|
"loss": 0.3786, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5098868608474731, |
|
"eval_runtime": 4.1343, |
|
"eval_samples_per_second": 61.437, |
|
"eval_steps_per_second": 0.484, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 140.43478260869566, |
|
"grad_norm": 3.3379013538360596, |
|
"learning_rate": 5.956521739130435e-06, |
|
"loss": 0.405, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 140.8695652173913, |
|
"grad_norm": 3.2763419151306152, |
|
"learning_rate": 5.91304347826087e-06, |
|
"loss": 0.3976, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 141.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5111474394798279, |
|
"eval_runtime": 3.8029, |
|
"eval_samples_per_second": 66.792, |
|
"eval_steps_per_second": 0.526, |
|
"step": 3243 |
|
}, |
|
{ |
|
"epoch": 141.30434782608697, |
|
"grad_norm": 3.1908023357391357, |
|
"learning_rate": 5.8695652173913055e-06, |
|
"loss": 0.3856, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 141.7391304347826, |
|
"grad_norm": 3.875778913497925, |
|
"learning_rate": 5.826086956521739e-06, |
|
"loss": 0.371, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 142.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5081124901771545, |
|
"eval_runtime": 4.5605, |
|
"eval_samples_per_second": 55.696, |
|
"eval_steps_per_second": 0.439, |
|
"step": 3266 |
|
}, |
|
{ |
|
"epoch": 142.17391304347825, |
|
"grad_norm": 2.925506353378296, |
|
"learning_rate": 5.782608695652174e-06, |
|
"loss": 0.4169, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 142.6086956521739, |
|
"grad_norm": 9.266388893127441, |
|
"learning_rate": 5.739130434782609e-06, |
|
"loss": 0.4056, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 143.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5098369717597961, |
|
"eval_runtime": 3.9928, |
|
"eval_samples_per_second": 63.615, |
|
"eval_steps_per_second": 0.501, |
|
"step": 3289 |
|
}, |
|
{ |
|
"epoch": 143.04347826086956, |
|
"grad_norm": 25.856365203857422, |
|
"learning_rate": 5.695652173913044e-06, |
|
"loss": 0.3757, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 143.47826086956522, |
|
"grad_norm": 2.937258720397949, |
|
"learning_rate": 5.652173913043479e-06, |
|
"loss": 0.3745, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 143.91304347826087, |
|
"grad_norm": 3.236806631088257, |
|
"learning_rate": 5.608695652173914e-06, |
|
"loss": 0.4214, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 144.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5085259675979614, |
|
"eval_runtime": 3.7621, |
|
"eval_samples_per_second": 67.516, |
|
"eval_steps_per_second": 0.532, |
|
"step": 3312 |
|
}, |
|
{ |
|
"epoch": 144.34782608695653, |
|
"grad_norm": 3.6454241275787354, |
|
"learning_rate": 5.565217391304348e-06, |
|
"loss": 0.3659, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 144.7826086956522, |
|
"grad_norm": 3.4510464668273926, |
|
"learning_rate": 5.521739130434783e-06, |
|
"loss": 0.3832, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 145.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5084368586540222, |
|
"eval_runtime": 4.2307, |
|
"eval_samples_per_second": 60.037, |
|
"eval_steps_per_second": 0.473, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 145.2173913043478, |
|
"grad_norm": 2.4478542804718018, |
|
"learning_rate": 5.478260869565217e-06, |
|
"loss": 0.3522, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 145.65217391304347, |
|
"grad_norm": 4.097745895385742, |
|
"learning_rate": 5.4347826086956525e-06, |
|
"loss": 0.3762, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 146.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5060733556747437, |
|
"eval_runtime": 4.203, |
|
"eval_samples_per_second": 60.433, |
|
"eval_steps_per_second": 0.476, |
|
"step": 3358 |
|
}, |
|
{ |
|
"epoch": 146.08695652173913, |
|
"grad_norm": 3.384960651397705, |
|
"learning_rate": 5.391304347826088e-06, |
|
"loss": 0.3761, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 146.52173913043478, |
|
"grad_norm": 2.909395217895508, |
|
"learning_rate": 5.347826086956523e-06, |
|
"loss": 0.3902, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 146.95652173913044, |
|
"grad_norm": 2.538163900375366, |
|
"learning_rate": 5.304347826086957e-06, |
|
"loss": 0.4118, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 147.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5111083984375, |
|
"eval_runtime": 3.8334, |
|
"eval_samples_per_second": 66.26, |
|
"eval_steps_per_second": 0.522, |
|
"step": 3381 |
|
}, |
|
{ |
|
"epoch": 147.3913043478261, |
|
"grad_norm": 2.9644970893859863, |
|
"learning_rate": 5.260869565217391e-06, |
|
"loss": 0.3802, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 147.82608695652175, |
|
"grad_norm": 3.0972464084625244, |
|
"learning_rate": 5.2173913043478265e-06, |
|
"loss": 0.3866, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 148.0, |
|
"eval_accuracy": 0.8070866141732284, |
|
"eval_loss": 0.5092455148696899, |
|
"eval_runtime": 5.6174, |
|
"eval_samples_per_second": 45.216, |
|
"eval_steps_per_second": 0.356, |
|
"step": 3404 |
|
}, |
|
{ |
|
"epoch": 148.2608695652174, |
|
"grad_norm": 1.625214695930481, |
|
"learning_rate": 5.173913043478262e-06, |
|
"loss": 0.3584, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 148.69565217391303, |
|
"grad_norm": 15.01403522491455, |
|
"learning_rate": 5.130434782608697e-06, |
|
"loss": 0.3869, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 149.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.512187659740448, |
|
"eval_runtime": 3.8515, |
|
"eval_samples_per_second": 65.948, |
|
"eval_steps_per_second": 0.519, |
|
"step": 3427 |
|
}, |
|
{ |
|
"epoch": 149.1304347826087, |
|
"grad_norm": 1.9776344299316406, |
|
"learning_rate": 5.08695652173913e-06, |
|
"loss": 0.3921, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 149.56521739130434, |
|
"grad_norm": 2.336129665374756, |
|
"learning_rate": 5.043478260869565e-06, |
|
"loss": 0.4048, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"grad_norm": 3.6398816108703613, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3734, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5116916298866272, |
|
"eval_runtime": 3.7475, |
|
"eval_samples_per_second": 67.779, |
|
"eval_steps_per_second": 0.534, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 150.43478260869566, |
|
"grad_norm": 2.299021005630493, |
|
"learning_rate": 4.9565217391304355e-06, |
|
"loss": 0.3734, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 150.8695652173913, |
|
"grad_norm": 3.107494831085205, |
|
"learning_rate": 4.91304347826087e-06, |
|
"loss": 0.4061, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 151.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5094764232635498, |
|
"eval_runtime": 4.4075, |
|
"eval_samples_per_second": 57.629, |
|
"eval_steps_per_second": 0.454, |
|
"step": 3473 |
|
}, |
|
{ |
|
"epoch": 151.30434782608697, |
|
"grad_norm": 2.319066286087036, |
|
"learning_rate": 4.869565217391305e-06, |
|
"loss": 0.3681, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 151.7391304347826, |
|
"grad_norm": 2.7603538036346436, |
|
"learning_rate": 4.826086956521739e-06, |
|
"loss": 0.3705, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 152.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5171404480934143, |
|
"eval_runtime": 4.362, |
|
"eval_samples_per_second": 58.23, |
|
"eval_steps_per_second": 0.459, |
|
"step": 3496 |
|
}, |
|
{ |
|
"epoch": 152.17391304347825, |
|
"grad_norm": 2.0375826358795166, |
|
"learning_rate": 4.782608695652174e-06, |
|
"loss": 0.3882, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 152.6086956521739, |
|
"grad_norm": 2.8498833179473877, |
|
"learning_rate": 4.739130434782609e-06, |
|
"loss": 0.3873, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 153.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5179200768470764, |
|
"eval_runtime": 3.7588, |
|
"eval_samples_per_second": 67.575, |
|
"eval_steps_per_second": 0.532, |
|
"step": 3519 |
|
}, |
|
{ |
|
"epoch": 153.04347826086956, |
|
"grad_norm": 2.707977533340454, |
|
"learning_rate": 4.695652173913044e-06, |
|
"loss": 0.3979, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 153.47826086956522, |
|
"grad_norm": 3.5183486938476562, |
|
"learning_rate": 4.652173913043478e-06, |
|
"loss": 0.4025, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 153.91304347826087, |
|
"grad_norm": 2.90291166305542, |
|
"learning_rate": 4.608695652173913e-06, |
|
"loss": 0.3927, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 154.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5117496252059937, |
|
"eval_runtime": 3.7541, |
|
"eval_samples_per_second": 67.659, |
|
"eval_steps_per_second": 0.533, |
|
"step": 3542 |
|
}, |
|
{ |
|
"epoch": 154.34782608695653, |
|
"grad_norm": 4.005958080291748, |
|
"learning_rate": 4.565217391304348e-06, |
|
"loss": 0.4011, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 154.7826086956522, |
|
"grad_norm": 2.469202995300293, |
|
"learning_rate": 4.5217391304347826e-06, |
|
"loss": 0.3807, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 155.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5133464932441711, |
|
"eval_runtime": 5.8154, |
|
"eval_samples_per_second": 43.677, |
|
"eval_steps_per_second": 0.344, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 155.2173913043478, |
|
"grad_norm": 3.2248237133026123, |
|
"learning_rate": 4.478260869565218e-06, |
|
"loss": 0.4498, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 155.65217391304347, |
|
"grad_norm": 3.463270425796509, |
|
"learning_rate": 4.434782608695653e-06, |
|
"loss": 0.3761, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 156.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5140319466590881, |
|
"eval_runtime": 3.7668, |
|
"eval_samples_per_second": 67.432, |
|
"eval_steps_per_second": 0.531, |
|
"step": 3588 |
|
}, |
|
{ |
|
"epoch": 156.08695652173913, |
|
"grad_norm": 3.640611171722412, |
|
"learning_rate": 4.391304347826087e-06, |
|
"loss": 0.3609, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 156.52173913043478, |
|
"grad_norm": 4.198793888092041, |
|
"learning_rate": 4.347826086956522e-06, |
|
"loss": 0.3984, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 156.95652173913044, |
|
"grad_norm": 2.9035775661468506, |
|
"learning_rate": 4.304347826086957e-06, |
|
"loss": 0.3964, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 157.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5117691159248352, |
|
"eval_runtime": 3.7832, |
|
"eval_samples_per_second": 67.138, |
|
"eval_steps_per_second": 0.529, |
|
"step": 3611 |
|
}, |
|
{ |
|
"epoch": 157.3913043478261, |
|
"grad_norm": 5.13762092590332, |
|
"learning_rate": 4.260869565217392e-06, |
|
"loss": 0.3818, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 157.82608695652175, |
|
"grad_norm": 8.948963165283203, |
|
"learning_rate": 4.217391304347827e-06, |
|
"loss": 0.39, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 158.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5122236609458923, |
|
"eval_runtime": 4.6309, |
|
"eval_samples_per_second": 54.849, |
|
"eval_steps_per_second": 0.432, |
|
"step": 3634 |
|
}, |
|
{ |
|
"epoch": 158.2608695652174, |
|
"grad_norm": 2.4759654998779297, |
|
"learning_rate": 4.173913043478261e-06, |
|
"loss": 0.3784, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 158.69565217391303, |
|
"grad_norm": 2.407663106918335, |
|
"learning_rate": 4.130434782608696e-06, |
|
"loss": 0.3943, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 159.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5125917196273804, |
|
"eval_runtime": 3.7278, |
|
"eval_samples_per_second": 68.138, |
|
"eval_steps_per_second": 0.537, |
|
"step": 3657 |
|
}, |
|
{ |
|
"epoch": 159.1304347826087, |
|
"grad_norm": 2.2464840412139893, |
|
"learning_rate": 4.086956521739131e-06, |
|
"loss": 0.3675, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 159.56521739130434, |
|
"grad_norm": 3.0186944007873535, |
|
"learning_rate": 4.0434782608695655e-06, |
|
"loss": 0.355, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"grad_norm": 9.606362342834473, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.3417, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5096677541732788, |
|
"eval_runtime": 3.7505, |
|
"eval_samples_per_second": 67.724, |
|
"eval_steps_per_second": 0.533, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 160.43478260869566, |
|
"grad_norm": 3.155024766921997, |
|
"learning_rate": 3.956521739130435e-06, |
|
"loss": 0.3951, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 160.8695652173913, |
|
"grad_norm": 2.3195645809173584, |
|
"learning_rate": 3.91304347826087e-06, |
|
"loss": 0.3996, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 161.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5048008561134338, |
|
"eval_runtime": 4.9463, |
|
"eval_samples_per_second": 51.351, |
|
"eval_steps_per_second": 0.404, |
|
"step": 3703 |
|
}, |
|
{ |
|
"epoch": 161.30434782608697, |
|
"grad_norm": 16.818618774414062, |
|
"learning_rate": 3.869565217391304e-06, |
|
"loss": 0.3613, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 161.7391304347826, |
|
"grad_norm": 5.290389060974121, |
|
"learning_rate": 3.8260869565217395e-06, |
|
"loss": 0.4, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 162.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5148473978042603, |
|
"eval_runtime": 3.7348, |
|
"eval_samples_per_second": 68.008, |
|
"eval_steps_per_second": 0.535, |
|
"step": 3726 |
|
}, |
|
{ |
|
"epoch": 162.17391304347825, |
|
"grad_norm": 4.7519330978393555, |
|
"learning_rate": 3.782608695652174e-06, |
|
"loss": 0.3983, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 162.6086956521739, |
|
"grad_norm": 2.433164358139038, |
|
"learning_rate": 3.739130434782609e-06, |
|
"loss": 0.4051, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 163.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5150399804115295, |
|
"eval_runtime": 3.7013, |
|
"eval_samples_per_second": 68.625, |
|
"eval_steps_per_second": 0.54, |
|
"step": 3749 |
|
}, |
|
{ |
|
"epoch": 163.04347826086956, |
|
"grad_norm": 2.870962381362915, |
|
"learning_rate": 3.6956521739130436e-06, |
|
"loss": 0.3903, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 163.47826086956522, |
|
"grad_norm": 3.3795669078826904, |
|
"learning_rate": 3.6521739130434787e-06, |
|
"loss": 0.3981, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 163.91304347826087, |
|
"grad_norm": 4.447073936462402, |
|
"learning_rate": 3.6086956521739134e-06, |
|
"loss": 0.3973, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 164.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5036624073982239, |
|
"eval_runtime": 4.8343, |
|
"eval_samples_per_second": 52.541, |
|
"eval_steps_per_second": 0.414, |
|
"step": 3772 |
|
}, |
|
{ |
|
"epoch": 164.34782608695653, |
|
"grad_norm": 2.5403716564178467, |
|
"learning_rate": 3.565217391304348e-06, |
|
"loss": 0.3586, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 164.7826086956522, |
|
"grad_norm": 2.5216853618621826, |
|
"learning_rate": 3.5217391304347832e-06, |
|
"loss": 0.3963, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 165.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5048288702964783, |
|
"eval_runtime": 3.7404, |
|
"eval_samples_per_second": 67.907, |
|
"eval_steps_per_second": 0.535, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 165.2173913043478, |
|
"grad_norm": 3.382376194000244, |
|
"learning_rate": 3.4782608695652175e-06, |
|
"loss": 0.4012, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 165.65217391304347, |
|
"grad_norm": 3.0021872520446777, |
|
"learning_rate": 3.4347826086956526e-06, |
|
"loss": 0.3568, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 166.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5167564749717712, |
|
"eval_runtime": 3.6895, |
|
"eval_samples_per_second": 68.845, |
|
"eval_steps_per_second": 0.542, |
|
"step": 3818 |
|
}, |
|
{ |
|
"epoch": 166.08695652173913, |
|
"grad_norm": 4.209798812866211, |
|
"learning_rate": 3.391304347826087e-06, |
|
"loss": 0.4217, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 166.52173913043478, |
|
"grad_norm": 2.3605332374572754, |
|
"learning_rate": 3.347826086956522e-06, |
|
"loss": 0.3897, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 166.95652173913044, |
|
"grad_norm": 7.9494733810424805, |
|
"learning_rate": 3.3043478260869567e-06, |
|
"loss": 0.3995, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 167.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5096150636672974, |
|
"eval_runtime": 4.9956, |
|
"eval_samples_per_second": 50.845, |
|
"eval_steps_per_second": 0.4, |
|
"step": 3841 |
|
}, |
|
{ |
|
"epoch": 167.3913043478261, |
|
"grad_norm": 3.431043863296509, |
|
"learning_rate": 3.2608695652173914e-06, |
|
"loss": 0.3765, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 167.82608695652175, |
|
"grad_norm": 3.4384922981262207, |
|
"learning_rate": 3.217391304347826e-06, |
|
"loss": 0.3628, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 168.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5101594924926758, |
|
"eval_runtime": 3.7705, |
|
"eval_samples_per_second": 67.365, |
|
"eval_steps_per_second": 0.53, |
|
"step": 3864 |
|
}, |
|
{ |
|
"epoch": 168.2608695652174, |
|
"grad_norm": 8.502880096435547, |
|
"learning_rate": 3.1739130434782613e-06, |
|
"loss": 0.3857, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 168.69565217391303, |
|
"grad_norm": 2.5634241104125977, |
|
"learning_rate": 3.130434782608696e-06, |
|
"loss": 0.3836, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 169.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5133307576179504, |
|
"eval_runtime": 3.7532, |
|
"eval_samples_per_second": 67.676, |
|
"eval_steps_per_second": 0.533, |
|
"step": 3887 |
|
}, |
|
{ |
|
"epoch": 169.1304347826087, |
|
"grad_norm": 3.617677927017212, |
|
"learning_rate": 3.0869565217391307e-06, |
|
"loss": 0.4251, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 169.56521739130434, |
|
"grad_norm": 3.9091439247131348, |
|
"learning_rate": 3.043478260869566e-06, |
|
"loss": 0.3747, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 170.0, |
|
"grad_norm": 12.626005172729492, |
|
"learning_rate": 3e-06, |
|
"loss": 0.3646, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 170.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5099019408226013, |
|
"eval_runtime": 4.85, |
|
"eval_samples_per_second": 52.372, |
|
"eval_steps_per_second": 0.412, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 170.43478260869566, |
|
"grad_norm": 5.3712263107299805, |
|
"learning_rate": 2.956521739130435e-06, |
|
"loss": 0.3732, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 170.8695652173913, |
|
"grad_norm": 2.489645481109619, |
|
"learning_rate": 2.9130434782608695e-06, |
|
"loss": 0.3789, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 171.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5151440501213074, |
|
"eval_runtime": 3.7265, |
|
"eval_samples_per_second": 68.161, |
|
"eval_steps_per_second": 0.537, |
|
"step": 3933 |
|
}, |
|
{ |
|
"epoch": 171.30434782608697, |
|
"grad_norm": 60.279747009277344, |
|
"learning_rate": 2.8695652173913046e-06, |
|
"loss": 0.3372, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 171.7391304347826, |
|
"grad_norm": 5.177385330200195, |
|
"learning_rate": 2.8260869565217393e-06, |
|
"loss": 0.3832, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 172.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5148643255233765, |
|
"eval_runtime": 3.7835, |
|
"eval_samples_per_second": 67.134, |
|
"eval_steps_per_second": 0.529, |
|
"step": 3956 |
|
}, |
|
{ |
|
"epoch": 172.17391304347825, |
|
"grad_norm": 2.5841851234436035, |
|
"learning_rate": 2.782608695652174e-06, |
|
"loss": 0.405, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 172.6086956521739, |
|
"grad_norm": 2.6472222805023193, |
|
"learning_rate": 2.7391304347826087e-06, |
|
"loss": 0.3476, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 173.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5178123116493225, |
|
"eval_runtime": 5.0055, |
|
"eval_samples_per_second": 50.745, |
|
"eval_steps_per_second": 0.4, |
|
"step": 3979 |
|
}, |
|
{ |
|
"epoch": 173.04347826086956, |
|
"grad_norm": 2.3995625972747803, |
|
"learning_rate": 2.695652173913044e-06, |
|
"loss": 0.4347, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 173.47826086956522, |
|
"grad_norm": 4.958439826965332, |
|
"learning_rate": 2.6521739130434785e-06, |
|
"loss": 0.3886, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 173.91304347826087, |
|
"grad_norm": 4.661713600158691, |
|
"learning_rate": 2.6086956521739132e-06, |
|
"loss": 0.3806, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 174.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5080812573432922, |
|
"eval_runtime": 3.6838, |
|
"eval_samples_per_second": 68.951, |
|
"eval_steps_per_second": 0.543, |
|
"step": 4002 |
|
}, |
|
{ |
|
"epoch": 174.34782608695653, |
|
"grad_norm": 2.979862928390503, |
|
"learning_rate": 2.5652173913043484e-06, |
|
"loss": 0.3429, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 174.7826086956522, |
|
"grad_norm": 1.8571139574050903, |
|
"learning_rate": 2.5217391304347826e-06, |
|
"loss": 0.4053, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 175.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5099707245826721, |
|
"eval_runtime": 3.7194, |
|
"eval_samples_per_second": 68.291, |
|
"eval_steps_per_second": 0.538, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 175.2173913043478, |
|
"grad_norm": 2.364047050476074, |
|
"learning_rate": 2.4782608695652178e-06, |
|
"loss": 0.3774, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 175.65217391304347, |
|
"grad_norm": 4.220658779144287, |
|
"learning_rate": 2.4347826086956525e-06, |
|
"loss": 0.3986, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 176.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5189133286476135, |
|
"eval_runtime": 5.0478, |
|
"eval_samples_per_second": 50.319, |
|
"eval_steps_per_second": 0.396, |
|
"step": 4048 |
|
}, |
|
{ |
|
"epoch": 176.08695652173913, |
|
"grad_norm": 2.9689295291900635, |
|
"learning_rate": 2.391304347826087e-06, |
|
"loss": 0.4225, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 176.52173913043478, |
|
"grad_norm": 3.78476881980896, |
|
"learning_rate": 2.347826086956522e-06, |
|
"loss": 0.3798, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 176.95652173913044, |
|
"grad_norm": 2.3258774280548096, |
|
"learning_rate": 2.3043478260869566e-06, |
|
"loss": 0.3827, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 177.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5128843784332275, |
|
"eval_runtime": 3.7539, |
|
"eval_samples_per_second": 67.663, |
|
"eval_steps_per_second": 0.533, |
|
"step": 4071 |
|
}, |
|
{ |
|
"epoch": 177.3913043478261, |
|
"grad_norm": 2.329585313796997, |
|
"learning_rate": 2.2608695652173913e-06, |
|
"loss": 0.329, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 177.82608695652175, |
|
"grad_norm": 3.0889029502868652, |
|
"learning_rate": 2.2173913043478264e-06, |
|
"loss": 0.3892, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 178.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5099364519119263, |
|
"eval_runtime": 3.764, |
|
"eval_samples_per_second": 67.482, |
|
"eval_steps_per_second": 0.531, |
|
"step": 4094 |
|
}, |
|
{ |
|
"epoch": 178.2608695652174, |
|
"grad_norm": 3.167226791381836, |
|
"learning_rate": 2.173913043478261e-06, |
|
"loss": 0.3801, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 178.69565217391303, |
|
"grad_norm": 2.857957601547241, |
|
"learning_rate": 2.130434782608696e-06, |
|
"loss": 0.3955, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 179.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5212357640266418, |
|
"eval_runtime": 4.8308, |
|
"eval_samples_per_second": 52.579, |
|
"eval_steps_per_second": 0.414, |
|
"step": 4117 |
|
}, |
|
{ |
|
"epoch": 179.1304347826087, |
|
"grad_norm": 8.153979301452637, |
|
"learning_rate": 2.0869565217391305e-06, |
|
"loss": 0.4062, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 179.56521739130434, |
|
"grad_norm": 3.2647910118103027, |
|
"learning_rate": 2.0434782608695656e-06, |
|
"loss": 0.3603, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"grad_norm": 4.87031364440918, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.4077, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5101702213287354, |
|
"eval_runtime": 3.8052, |
|
"eval_samples_per_second": 66.75, |
|
"eval_steps_per_second": 0.526, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 180.43478260869566, |
|
"grad_norm": 3.3625569343566895, |
|
"learning_rate": 1.956521739130435e-06, |
|
"loss": 0.3881, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 180.8695652173913, |
|
"grad_norm": 3.717646360397339, |
|
"learning_rate": 1.9130434782608697e-06, |
|
"loss": 0.3579, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 181.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5099858641624451, |
|
"eval_runtime": 3.707, |
|
"eval_samples_per_second": 68.52, |
|
"eval_steps_per_second": 0.54, |
|
"step": 4163 |
|
}, |
|
{ |
|
"epoch": 181.30434782608697, |
|
"grad_norm": 2.5178964138031006, |
|
"learning_rate": 1.8695652173913044e-06, |
|
"loss": 0.3828, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 181.7391304347826, |
|
"grad_norm": 3.244948625564575, |
|
"learning_rate": 1.8260869565217394e-06, |
|
"loss": 0.3666, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 182.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5247715711593628, |
|
"eval_runtime": 4.2228, |
|
"eval_samples_per_second": 60.149, |
|
"eval_steps_per_second": 0.474, |
|
"step": 4186 |
|
}, |
|
{ |
|
"epoch": 182.17391304347825, |
|
"grad_norm": 3.418851613998413, |
|
"learning_rate": 1.782608695652174e-06, |
|
"loss": 0.4, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 182.6086956521739, |
|
"grad_norm": 2.247349262237549, |
|
"learning_rate": 1.7391304347826088e-06, |
|
"loss": 0.3746, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 183.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5220462083816528, |
|
"eval_runtime": 4.3605, |
|
"eval_samples_per_second": 58.25, |
|
"eval_steps_per_second": 0.459, |
|
"step": 4209 |
|
}, |
|
{ |
|
"epoch": 183.04347826086956, |
|
"grad_norm": 5.591789245605469, |
|
"learning_rate": 1.6956521739130435e-06, |
|
"loss": 0.3971, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 183.47826086956522, |
|
"grad_norm": 2.8663575649261475, |
|
"learning_rate": 1.6521739130434784e-06, |
|
"loss": 0.3516, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 183.91304347826087, |
|
"grad_norm": 5.791408061981201, |
|
"learning_rate": 1.608695652173913e-06, |
|
"loss": 0.3867, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 184.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5172824859619141, |
|
"eval_runtime": 3.8331, |
|
"eval_samples_per_second": 66.265, |
|
"eval_steps_per_second": 0.522, |
|
"step": 4232 |
|
}, |
|
{ |
|
"epoch": 184.34782608695653, |
|
"grad_norm": 3.3605191707611084, |
|
"learning_rate": 1.565217391304348e-06, |
|
"loss": 0.3911, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 184.7826086956522, |
|
"grad_norm": 3.4683103561401367, |
|
"learning_rate": 1.521739130434783e-06, |
|
"loss": 0.4024, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 185.0, |
|
"eval_accuracy": 0.7874015748031497, |
|
"eval_loss": 0.5248106122016907, |
|
"eval_runtime": 4.5705, |
|
"eval_samples_per_second": 55.574, |
|
"eval_steps_per_second": 0.438, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 185.2173913043478, |
|
"grad_norm": 4.495180130004883, |
|
"learning_rate": 1.4782608695652176e-06, |
|
"loss": 0.3931, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 185.65217391304347, |
|
"grad_norm": 4.51051139831543, |
|
"learning_rate": 1.4347826086956523e-06, |
|
"loss": 0.4014, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 186.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5084752440452576, |
|
"eval_runtime": 4.1594, |
|
"eval_samples_per_second": 61.066, |
|
"eval_steps_per_second": 0.481, |
|
"step": 4278 |
|
}, |
|
{ |
|
"epoch": 186.08695652173913, |
|
"grad_norm": 6.847979545593262, |
|
"learning_rate": 1.391304347826087e-06, |
|
"loss": 0.3887, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 186.52173913043478, |
|
"grad_norm": 8.414494514465332, |
|
"learning_rate": 1.347826086956522e-06, |
|
"loss": 0.3876, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 186.95652173913044, |
|
"grad_norm": 2.0459609031677246, |
|
"learning_rate": 1.3043478260869566e-06, |
|
"loss": 0.3445, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 187.0, |
|
"eval_accuracy": 0.8031496062992126, |
|
"eval_loss": 0.5136986970901489, |
|
"eval_runtime": 3.7104, |
|
"eval_samples_per_second": 68.456, |
|
"eval_steps_per_second": 0.539, |
|
"step": 4301 |
|
}, |
|
{ |
|
"epoch": 187.3913043478261, |
|
"grad_norm": 2.7707877159118652, |
|
"learning_rate": 1.2608695652173913e-06, |
|
"loss": 0.4067, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 187.82608695652175, |
|
"grad_norm": 2.2277884483337402, |
|
"learning_rate": 1.2173913043478262e-06, |
|
"loss": 0.382, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 188.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.521314799785614, |
|
"eval_runtime": 4.1528, |
|
"eval_samples_per_second": 61.164, |
|
"eval_steps_per_second": 0.482, |
|
"step": 4324 |
|
}, |
|
{ |
|
"epoch": 188.2608695652174, |
|
"grad_norm": 4.299314498901367, |
|
"learning_rate": 1.173913043478261e-06, |
|
"loss": 0.3717, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 188.69565217391303, |
|
"grad_norm": 2.479510545730591, |
|
"learning_rate": 1.1304347826086956e-06, |
|
"loss": 0.3673, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 189.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5241702795028687, |
|
"eval_runtime": 4.1853, |
|
"eval_samples_per_second": 60.689, |
|
"eval_steps_per_second": 0.478, |
|
"step": 4347 |
|
}, |
|
{ |
|
"epoch": 189.1304347826087, |
|
"grad_norm": 3.9942944049835205, |
|
"learning_rate": 1.0869565217391306e-06, |
|
"loss": 0.4158, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 189.56521739130434, |
|
"grad_norm": 2.8651175498962402, |
|
"learning_rate": 1.0434782608695653e-06, |
|
"loss": 0.3919, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 190.0, |
|
"grad_norm": 3.1065168380737305, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.3631, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 190.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5146118402481079, |
|
"eval_runtime": 3.7356, |
|
"eval_samples_per_second": 67.995, |
|
"eval_steps_per_second": 0.535, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 190.43478260869566, |
|
"grad_norm": 3.8796093463897705, |
|
"learning_rate": 9.565217391304349e-07, |
|
"loss": 0.3893, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 190.8695652173913, |
|
"grad_norm": 3.2894842624664307, |
|
"learning_rate": 9.130434782608697e-07, |
|
"loss": 0.393, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 191.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5097819566726685, |
|
"eval_runtime": 3.7404, |
|
"eval_samples_per_second": 67.908, |
|
"eval_steps_per_second": 0.535, |
|
"step": 4393 |
|
}, |
|
{ |
|
"epoch": 191.30434782608697, |
|
"grad_norm": 2.4112348556518555, |
|
"learning_rate": 8.695652173913044e-07, |
|
"loss": 0.4037, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 191.7391304347826, |
|
"grad_norm": 2.4510791301727295, |
|
"learning_rate": 8.260869565217392e-07, |
|
"loss": 0.3806, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 192.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5133717656135559, |
|
"eval_runtime": 5.016, |
|
"eval_samples_per_second": 50.638, |
|
"eval_steps_per_second": 0.399, |
|
"step": 4416 |
|
}, |
|
{ |
|
"epoch": 192.17391304347825, |
|
"grad_norm": 3.1017332077026367, |
|
"learning_rate": 7.82608695652174e-07, |
|
"loss": 0.3598, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 192.6086956521739, |
|
"grad_norm": 3.5164568424224854, |
|
"learning_rate": 7.391304347826088e-07, |
|
"loss": 0.3789, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 193.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5127285718917847, |
|
"eval_runtime": 3.736, |
|
"eval_samples_per_second": 67.988, |
|
"eval_steps_per_second": 0.535, |
|
"step": 4439 |
|
}, |
|
{ |
|
"epoch": 193.04347826086956, |
|
"grad_norm": 86.44344329833984, |
|
"learning_rate": 6.956521739130435e-07, |
|
"loss": 0.3858, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 193.47826086956522, |
|
"grad_norm": 2.892185688018799, |
|
"learning_rate": 6.521739130434783e-07, |
|
"loss": 0.3894, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 193.91304347826087, |
|
"grad_norm": 2.0254733562469482, |
|
"learning_rate": 6.086956521739131e-07, |
|
"loss": 0.3717, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 194.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5183544158935547, |
|
"eval_runtime": 3.7197, |
|
"eval_samples_per_second": 68.285, |
|
"eval_steps_per_second": 0.538, |
|
"step": 4462 |
|
}, |
|
{ |
|
"epoch": 194.34782608695653, |
|
"grad_norm": 4.124297618865967, |
|
"learning_rate": 5.652173913043478e-07, |
|
"loss": 0.4098, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 194.7826086956522, |
|
"grad_norm": 4.1497955322265625, |
|
"learning_rate": 5.217391304347826e-07, |
|
"loss": 0.361, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 195.0, |
|
"eval_accuracy": 0.7834645669291339, |
|
"eval_loss": 0.5185708999633789, |
|
"eval_runtime": 4.9741, |
|
"eval_samples_per_second": 51.064, |
|
"eval_steps_per_second": 0.402, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 195.2173913043478, |
|
"grad_norm": 11.268845558166504, |
|
"learning_rate": 4.782608695652174e-07, |
|
"loss": 0.3786, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 195.65217391304347, |
|
"grad_norm": 3.9937920570373535, |
|
"learning_rate": 4.347826086956522e-07, |
|
"loss": 0.3722, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 196.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5107359886169434, |
|
"eval_runtime": 3.7506, |
|
"eval_samples_per_second": 67.723, |
|
"eval_steps_per_second": 0.533, |
|
"step": 4508 |
|
}, |
|
{ |
|
"epoch": 196.08695652173913, |
|
"grad_norm": 2.869596004486084, |
|
"learning_rate": 3.91304347826087e-07, |
|
"loss": 0.3985, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 196.52173913043478, |
|
"grad_norm": 6.21280574798584, |
|
"learning_rate": 3.4782608695652175e-07, |
|
"loss": 0.4019, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 196.95652173913044, |
|
"grad_norm": 2.2324206829071045, |
|
"learning_rate": 3.0434782608695656e-07, |
|
"loss": 0.3551, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 197.0, |
|
"eval_accuracy": 0.7952755905511811, |
|
"eval_loss": 0.5174936056137085, |
|
"eval_runtime": 3.6975, |
|
"eval_samples_per_second": 68.695, |
|
"eval_steps_per_second": 0.541, |
|
"step": 4531 |
|
}, |
|
{ |
|
"epoch": 197.3913043478261, |
|
"grad_norm": 2.6415905952453613, |
|
"learning_rate": 2.608695652173913e-07, |
|
"loss": 0.3919, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 197.82608695652175, |
|
"grad_norm": 5.146513938903809, |
|
"learning_rate": 2.173913043478261e-07, |
|
"loss": 0.3649, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 198.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5135703682899475, |
|
"eval_runtime": 4.9875, |
|
"eval_samples_per_second": 50.928, |
|
"eval_steps_per_second": 0.401, |
|
"step": 4554 |
|
}, |
|
{ |
|
"epoch": 198.2608695652174, |
|
"grad_norm": 3.1943461894989014, |
|
"learning_rate": 1.7391304347826088e-07, |
|
"loss": 0.3763, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 198.69565217391303, |
|
"grad_norm": 2.8955743312835693, |
|
"learning_rate": 1.3043478260869566e-07, |
|
"loss": 0.3749, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 199.0, |
|
"eval_accuracy": 0.7913385826771654, |
|
"eval_loss": 0.5192672610282898, |
|
"eval_runtime": 3.6944, |
|
"eval_samples_per_second": 68.753, |
|
"eval_steps_per_second": 0.541, |
|
"step": 4577 |
|
}, |
|
{ |
|
"epoch": 199.1304347826087, |
|
"grad_norm": 12.166488647460938, |
|
"learning_rate": 8.695652173913044e-08, |
|
"loss": 0.3869, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 199.56521739130434, |
|
"grad_norm": 2.9687561988830566, |
|
"learning_rate": 4.347826086956522e-08, |
|
"loss": 0.3926, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"grad_norm": 4.834624290466309, |
|
"learning_rate": 0.0, |
|
"loss": 0.3782, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"eval_accuracy": 0.7992125984251969, |
|
"eval_loss": 0.5181651711463928, |
|
"eval_runtime": 3.7789, |
|
"eval_samples_per_second": 67.216, |
|
"eval_steps_per_second": 0.529, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"step": 4600, |
|
"total_flos": 1.089869514338304e+18, |
|
"train_loss": 0.30267460563908455, |
|
"train_runtime": 4787.5341, |
|
"train_samples_per_second": 59.947, |
|
"train_steps_per_second": 0.961 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 4600, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 200, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.089869514338304e+18, |
|
"train_batch_size": 64, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|