|
{ |
|
"best_metric": 0.36585365853658536, |
|
"best_model_checkpoint": "google/vivit-b-16x2-kinetics400-CAER-SAMPLE\\checkpoint-588", |
|
"epoch": 10.066666666666666, |
|
"eval_steps": 500, |
|
"global_step": 2100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 21.040367126464844, |
|
"learning_rate": 2.3809523809523808e-06, |
|
"loss": 2.0819, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 20.962032318115234, |
|
"learning_rate": 4.7619047619047615e-06, |
|
"loss": 2.1082, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 19.943674087524414, |
|
"learning_rate": 7.142857142857143e-06, |
|
"loss": 2.148, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 20.1109619140625, |
|
"learning_rate": 9.523809523809523e-06, |
|
"loss": 1.9797, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 20.655811309814453, |
|
"learning_rate": 1.1904761904761905e-05, |
|
"loss": 2.1647, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 19.211088180541992, |
|
"learning_rate": 1.4285714285714285e-05, |
|
"loss": 2.0541, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 19.060794830322266, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.9982, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 18.562660217285156, |
|
"learning_rate": 1.9047619047619046e-05, |
|
"loss": 1.9553, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 21.629850387573242, |
|
"learning_rate": 2.1428571428571428e-05, |
|
"loss": 2.0218, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 22.143321990966797, |
|
"learning_rate": 2.380952380952381e-05, |
|
"loss": 2.0226, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 21.463300704956055, |
|
"learning_rate": 2.6190476190476192e-05, |
|
"loss": 2.1175, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 20.069828033447266, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 1.9767, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 21.68781280517578, |
|
"learning_rate": 3.095238095238095e-05, |
|
"loss": 1.944, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 22.342853546142578, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 1.8881, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 22.58480453491211, |
|
"learning_rate": 3.571428571428572e-05, |
|
"loss": 1.957, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 21.043750762939453, |
|
"learning_rate": 3.809523809523809e-05, |
|
"loss": 1.9067, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 18.581403732299805, |
|
"learning_rate": 4.047619047619048e-05, |
|
"loss": 1.8434, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 19.873624801635742, |
|
"learning_rate": 4.2857142857142856e-05, |
|
"loss": 1.7584, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 27.060535430908203, |
|
"learning_rate": 4.523809523809524e-05, |
|
"loss": 2.4781, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_accuracy": 0.24390243902439024, |
|
"eval_loss": 1.8166238069534302, |
|
"eval_runtime": 97.0837, |
|
"eval_samples_per_second": 0.422, |
|
"eval_steps_per_second": 0.216, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 22.22335433959961, |
|
"learning_rate": 4.761904761904762e-05, |
|
"loss": 1.8185, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 18.064006805419922, |
|
"learning_rate": 5e-05, |
|
"loss": 1.513, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 21.092018127441406, |
|
"learning_rate": 4.973544973544973e-05, |
|
"loss": 1.6385, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 24.412302017211914, |
|
"learning_rate": 4.9470899470899475e-05, |
|
"loss": 1.6291, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 23.470930099487305, |
|
"learning_rate": 4.9206349206349204e-05, |
|
"loss": 1.4416, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 24.228782653808594, |
|
"learning_rate": 4.894179894179895e-05, |
|
"loss": 2.2057, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 17.363630294799805, |
|
"learning_rate": 4.8677248677248676e-05, |
|
"loss": 1.8102, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 25.69906234741211, |
|
"learning_rate": 4.841269841269841e-05, |
|
"loss": 1.7703, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 24.245840072631836, |
|
"learning_rate": 4.814814814814815e-05, |
|
"loss": 1.8854, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 18.425830841064453, |
|
"learning_rate": 4.7883597883597884e-05, |
|
"loss": 1.4344, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 25.066003799438477, |
|
"learning_rate": 4.761904761904762e-05, |
|
"loss": 1.8341, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 18.590608596801758, |
|
"learning_rate": 4.7354497354497356e-05, |
|
"loss": 1.5598, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 18.962934494018555, |
|
"learning_rate": 4.708994708994709e-05, |
|
"loss": 1.6827, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 15.114640235900879, |
|
"learning_rate": 4.682539682539683e-05, |
|
"loss": 1.5371, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 10.92546558380127, |
|
"learning_rate": 4.656084656084656e-05, |
|
"loss": 1.6233, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 24.024572372436523, |
|
"learning_rate": 4.62962962962963e-05, |
|
"loss": 1.7803, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 22.18260383605957, |
|
"learning_rate": 4.603174603174603e-05, |
|
"loss": 2.0079, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 23.039701461791992, |
|
"learning_rate": 4.576719576719577e-05, |
|
"loss": 1.582, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 20.918445587158203, |
|
"learning_rate": 4.55026455026455e-05, |
|
"loss": 1.9383, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 20.540185928344727, |
|
"learning_rate": 4.523809523809524e-05, |
|
"loss": 2.0142, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"eval_accuracy": 0.1951219512195122, |
|
"eval_loss": 2.294609546661377, |
|
"eval_runtime": 97.734, |
|
"eval_samples_per_second": 0.42, |
|
"eval_steps_per_second": 0.215, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 20.112930297851562, |
|
"learning_rate": 4.4973544973544974e-05, |
|
"loss": 1.281, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 16.186555862426758, |
|
"learning_rate": 4.470899470899471e-05, |
|
"loss": 1.1811, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 17.308746337890625, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 1.6046, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 15.936869621276855, |
|
"learning_rate": 4.417989417989418e-05, |
|
"loss": 1.1209, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 18.442317962646484, |
|
"learning_rate": 4.391534391534391e-05, |
|
"loss": 1.613, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 25.79071044921875, |
|
"learning_rate": 4.3650793650793655e-05, |
|
"loss": 1.1109, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 17.710105895996094, |
|
"learning_rate": 4.3386243386243384e-05, |
|
"loss": 1.0664, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 14.870391845703125, |
|
"learning_rate": 4.312169312169313e-05, |
|
"loss": 1.2283, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 10.361430168151855, |
|
"learning_rate": 4.2857142857142856e-05, |
|
"loss": 0.9415, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 22.23798179626465, |
|
"learning_rate": 4.259259259259259e-05, |
|
"loss": 1.3095, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 15.113299369812012, |
|
"learning_rate": 4.232804232804233e-05, |
|
"loss": 1.2173, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 22.383806228637695, |
|
"learning_rate": 4.2063492063492065e-05, |
|
"loss": 1.6995, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 18.106990814208984, |
|
"learning_rate": 4.17989417989418e-05, |
|
"loss": 1.5246, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 17.58225440979004, |
|
"learning_rate": 4.153439153439154e-05, |
|
"loss": 1.6346, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 20.450359344482422, |
|
"learning_rate": 4.126984126984127e-05, |
|
"loss": 1.402, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 19.060672760009766, |
|
"learning_rate": 4.100529100529101e-05, |
|
"loss": 1.3466, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 19.76554298400879, |
|
"learning_rate": 4.074074074074074e-05, |
|
"loss": 1.3471, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 19.519485473632812, |
|
"learning_rate": 4.047619047619048e-05, |
|
"loss": 1.3535, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 27.53349494934082, |
|
"learning_rate": 4.021164021164021e-05, |
|
"loss": 1.2947, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"eval_accuracy": 0.36585365853658536, |
|
"eval_loss": 1.6997814178466797, |
|
"eval_runtime": 97.5161, |
|
"eval_samples_per_second": 0.42, |
|
"eval_steps_per_second": 0.215, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 12.76953411102295, |
|
"learning_rate": 3.9947089947089946e-05, |
|
"loss": 1.0455, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"grad_norm": 23.524166107177734, |
|
"learning_rate": 3.968253968253968e-05, |
|
"loss": 0.9096, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"grad_norm": 16.023509979248047, |
|
"learning_rate": 3.941798941798942e-05, |
|
"loss": 0.7188, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 6.053380012512207, |
|
"learning_rate": 3.9153439153439155e-05, |
|
"loss": 1.0134, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 21.445356369018555, |
|
"learning_rate": 3.888888888888889e-05, |
|
"loss": 1.1377, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 2.62589168548584, |
|
"learning_rate": 3.862433862433863e-05, |
|
"loss": 0.7862, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"grad_norm": 17.238386154174805, |
|
"learning_rate": 3.835978835978836e-05, |
|
"loss": 0.8987, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"grad_norm": 21.664026260375977, |
|
"learning_rate": 3.809523809523809e-05, |
|
"loss": 0.8078, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"grad_norm": 17.513721466064453, |
|
"learning_rate": 3.7830687830687835e-05, |
|
"loss": 1.0894, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"grad_norm": 28.274599075317383, |
|
"learning_rate": 3.7566137566137564e-05, |
|
"loss": 0.8099, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 12.51369571685791, |
|
"learning_rate": 3.730158730158731e-05, |
|
"loss": 0.7437, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 18.50059700012207, |
|
"learning_rate": 3.7037037037037037e-05, |
|
"loss": 0.966, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 16.658267974853516, |
|
"learning_rate": 3.677248677248677e-05, |
|
"loss": 0.9597, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 12.815518379211426, |
|
"learning_rate": 3.650793650793651e-05, |
|
"loss": 1.0438, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"grad_norm": 10.553025245666504, |
|
"learning_rate": 3.6243386243386245e-05, |
|
"loss": 1.0394, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"grad_norm": 19.39080238342285, |
|
"learning_rate": 3.597883597883598e-05, |
|
"loss": 0.7301, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 20.088712692260742, |
|
"learning_rate": 3.571428571428572e-05, |
|
"loss": 1.0226, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 21.335155487060547, |
|
"learning_rate": 3.5449735449735446e-05, |
|
"loss": 1.0559, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 13.45548152923584, |
|
"learning_rate": 3.518518518518519e-05, |
|
"loss": 0.7884, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 3.4905335903167725, |
|
"learning_rate": 3.492063492063492e-05, |
|
"loss": 0.8486, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"eval_accuracy": 0.21951219512195122, |
|
"eval_loss": 2.036907196044922, |
|
"eval_runtime": 97.631, |
|
"eval_samples_per_second": 0.42, |
|
"eval_steps_per_second": 0.215, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 14.28947639465332, |
|
"learning_rate": 3.465608465608466e-05, |
|
"loss": 0.4492, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"grad_norm": 3.365736722946167, |
|
"learning_rate": 3.439153439153439e-05, |
|
"loss": 0.3086, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"grad_norm": 0.7734526991844177, |
|
"learning_rate": 3.412698412698413e-05, |
|
"loss": 0.6843, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"grad_norm": 15.79121208190918, |
|
"learning_rate": 3.386243386243386e-05, |
|
"loss": 0.3582, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"grad_norm": 8.443323135375977, |
|
"learning_rate": 3.35978835978836e-05, |
|
"loss": 0.6707, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"grad_norm": 11.17960262298584, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 0.5133, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"grad_norm": 15.559344291687012, |
|
"learning_rate": 3.306878306878307e-05, |
|
"loss": 0.4993, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"grad_norm": 21.119354248046875, |
|
"learning_rate": 3.280423280423281e-05, |
|
"loss": 0.4755, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"grad_norm": 19.267467498779297, |
|
"learning_rate": 3.253968253968254e-05, |
|
"loss": 0.4796, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"grad_norm": 18.882789611816406, |
|
"learning_rate": 3.227513227513227e-05, |
|
"loss": 0.5674, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"grad_norm": 21.429229736328125, |
|
"learning_rate": 3.2010582010582015e-05, |
|
"loss": 0.575, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"grad_norm": 1.5254511833190918, |
|
"learning_rate": 3.1746031746031745e-05, |
|
"loss": 0.7652, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"grad_norm": 27.608917236328125, |
|
"learning_rate": 3.148148148148148e-05, |
|
"loss": 0.4358, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"grad_norm": 4.12730598449707, |
|
"learning_rate": 3.121693121693122e-05, |
|
"loss": 0.7179, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"grad_norm": 0.7054173946380615, |
|
"learning_rate": 3.095238095238095e-05, |
|
"loss": 0.4201, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"grad_norm": 6.384374618530273, |
|
"learning_rate": 3.068783068783069e-05, |
|
"loss": 0.523, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"grad_norm": 13.877901077270508, |
|
"learning_rate": 3.0423280423280425e-05, |
|
"loss": 1.1374, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"grad_norm": 11.547405242919922, |
|
"learning_rate": 3.0158730158730158e-05, |
|
"loss": 0.4007, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"grad_norm": 17.14991569519043, |
|
"learning_rate": 2.9894179894179897e-05, |
|
"loss": 0.6621, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"grad_norm": 4.131192684173584, |
|
"learning_rate": 2.962962962962963e-05, |
|
"loss": 0.2636, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"eval_accuracy": 0.3170731707317073, |
|
"eval_loss": 1.9747871160507202, |
|
"eval_runtime": 96.8808, |
|
"eval_samples_per_second": 0.423, |
|
"eval_steps_per_second": 0.217, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 31.432811737060547, |
|
"learning_rate": 2.9365079365079366e-05, |
|
"loss": 0.4902, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"grad_norm": 3.502074956893921, |
|
"learning_rate": 2.91005291005291e-05, |
|
"loss": 0.1312, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"grad_norm": 11.63001823425293, |
|
"learning_rate": 2.8835978835978838e-05, |
|
"loss": 0.2772, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"grad_norm": 18.965436935424805, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 0.6744, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"grad_norm": 0.373548686504364, |
|
"learning_rate": 2.830687830687831e-05, |
|
"loss": 0.1485, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"grad_norm": 1.2408517599105835, |
|
"learning_rate": 2.8042328042328043e-05, |
|
"loss": 0.5443, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"grad_norm": 7.223196506500244, |
|
"learning_rate": 2.777777777777778e-05, |
|
"loss": 0.212, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"grad_norm": 0.37929368019104004, |
|
"learning_rate": 2.7513227513227512e-05, |
|
"loss": 0.6033, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"grad_norm": 1.4948322772979736, |
|
"learning_rate": 2.724867724867725e-05, |
|
"loss": 0.3575, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"grad_norm": 23.476411819458008, |
|
"learning_rate": 2.6984126984126984e-05, |
|
"loss": 0.4821, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"grad_norm": 21.29241180419922, |
|
"learning_rate": 2.6719576719576723e-05, |
|
"loss": 0.2678, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"grad_norm": 1.714184045791626, |
|
"learning_rate": 2.6455026455026456e-05, |
|
"loss": 0.3042, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"grad_norm": 17.73127555847168, |
|
"learning_rate": 2.6190476190476192e-05, |
|
"loss": 0.4311, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"grad_norm": 12.709320068359375, |
|
"learning_rate": 2.5925925925925925e-05, |
|
"loss": 0.4813, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"grad_norm": 4.548611640930176, |
|
"learning_rate": 2.5661375661375664e-05, |
|
"loss": 0.1838, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"grad_norm": 27.118480682373047, |
|
"learning_rate": 2.5396825396825397e-05, |
|
"loss": 0.3369, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"grad_norm": 20.047382354736328, |
|
"learning_rate": 2.5132275132275137e-05, |
|
"loss": 0.1552, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"grad_norm": 2.572249412536621, |
|
"learning_rate": 2.4867724867724866e-05, |
|
"loss": 0.2454, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"grad_norm": 16.45228385925293, |
|
"learning_rate": 2.4603174603174602e-05, |
|
"loss": 0.2805, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"eval_accuracy": 0.36585365853658536, |
|
"eval_loss": 2.356266975402832, |
|
"eval_runtime": 97.8432, |
|
"eval_samples_per_second": 0.419, |
|
"eval_steps_per_second": 0.215, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 1.568260908126831, |
|
"learning_rate": 2.4338624338624338e-05, |
|
"loss": 0.3522, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"grad_norm": 0.9600222706794739, |
|
"learning_rate": 2.4074074074074074e-05, |
|
"loss": 0.199, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"grad_norm": 15.724946975708008, |
|
"learning_rate": 2.380952380952381e-05, |
|
"loss": 0.1719, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"grad_norm": 1.4226208925247192, |
|
"learning_rate": 2.3544973544973546e-05, |
|
"loss": 0.1337, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"grad_norm": 0.574043333530426, |
|
"learning_rate": 2.328042328042328e-05, |
|
"loss": 0.0584, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"grad_norm": 1.265620231628418, |
|
"learning_rate": 2.3015873015873015e-05, |
|
"loss": 0.1366, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"grad_norm": 0.19037747383117676, |
|
"learning_rate": 2.275132275132275e-05, |
|
"loss": 0.052, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"grad_norm": 2.8592171669006348, |
|
"learning_rate": 2.2486772486772487e-05, |
|
"loss": 0.0912, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"grad_norm": 32.83649444580078, |
|
"learning_rate": 2.2222222222222223e-05, |
|
"loss": 0.6816, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"grad_norm": 26.58733558654785, |
|
"learning_rate": 2.1957671957671956e-05, |
|
"loss": 0.1562, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"grad_norm": 0.8090476989746094, |
|
"learning_rate": 2.1693121693121692e-05, |
|
"loss": 0.0661, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"grad_norm": 12.49517822265625, |
|
"learning_rate": 2.1428571428571428e-05, |
|
"loss": 0.097, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"grad_norm": 1.2869700193405151, |
|
"learning_rate": 2.1164021164021164e-05, |
|
"loss": 0.2118, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"grad_norm": 1.337388038635254, |
|
"learning_rate": 2.08994708994709e-05, |
|
"loss": 0.1251, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"grad_norm": 15.6812162399292, |
|
"learning_rate": 2.0634920634920636e-05, |
|
"loss": 0.2407, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"grad_norm": 4.392534255981445, |
|
"learning_rate": 2.037037037037037e-05, |
|
"loss": 0.1948, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"grad_norm": 3.4775161743164062, |
|
"learning_rate": 2.0105820105820105e-05, |
|
"loss": 0.056, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"grad_norm": 0.3342803716659546, |
|
"learning_rate": 1.984126984126984e-05, |
|
"loss": 0.2194, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"grad_norm": 2.376802444458008, |
|
"learning_rate": 1.9576719576719577e-05, |
|
"loss": 0.1855, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"grad_norm": 0.15384900569915771, |
|
"learning_rate": 1.9312169312169313e-05, |
|
"loss": 0.0923, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"eval_accuracy": 0.36585365853658536, |
|
"eval_loss": 2.3754231929779053, |
|
"eval_runtime": 98.5579, |
|
"eval_samples_per_second": 0.416, |
|
"eval_steps_per_second": 0.213, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 0.41682931780815125, |
|
"learning_rate": 1.9047619047619046e-05, |
|
"loss": 0.097, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"grad_norm": 0.6600458025932312, |
|
"learning_rate": 1.8783068783068782e-05, |
|
"loss": 0.1678, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"grad_norm": 6.393650531768799, |
|
"learning_rate": 1.8518518518518518e-05, |
|
"loss": 0.0715, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"grad_norm": 0.031914032995700836, |
|
"learning_rate": 1.8253968253968254e-05, |
|
"loss": 0.1708, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"grad_norm": 0.04353192821145058, |
|
"learning_rate": 1.798941798941799e-05, |
|
"loss": 0.1562, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"grad_norm": 0.5575181841850281, |
|
"learning_rate": 1.7724867724867723e-05, |
|
"loss": 0.0242, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"grad_norm": 0.2302282154560089, |
|
"learning_rate": 1.746031746031746e-05, |
|
"loss": 0.039, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"grad_norm": 0.12757529318332672, |
|
"learning_rate": 1.7195767195767195e-05, |
|
"loss": 0.1855, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"grad_norm": 27.270219802856445, |
|
"learning_rate": 1.693121693121693e-05, |
|
"loss": 0.204, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"grad_norm": 0.17601485550403595, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 0.1326, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"grad_norm": 7.770025253295898, |
|
"learning_rate": 1.6402116402116404e-05, |
|
"loss": 0.0323, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"grad_norm": 0.36580923199653625, |
|
"learning_rate": 1.6137566137566136e-05, |
|
"loss": 0.1058, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"grad_norm": 4.40885591506958, |
|
"learning_rate": 1.5873015873015872e-05, |
|
"loss": 0.0768, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"grad_norm": 0.18270359933376312, |
|
"learning_rate": 1.560846560846561e-05, |
|
"loss": 0.2691, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"grad_norm": 2.5134904384613037, |
|
"learning_rate": 1.5343915343915344e-05, |
|
"loss": 0.1627, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"grad_norm": 0.8382402658462524, |
|
"learning_rate": 1.5079365079365079e-05, |
|
"loss": 0.0194, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"grad_norm": 22.632375717163086, |
|
"learning_rate": 1.4814814814814815e-05, |
|
"loss": 0.2512, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"grad_norm": 0.267007052898407, |
|
"learning_rate": 1.455026455026455e-05, |
|
"loss": 0.0539, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"grad_norm": 0.03810029849410057, |
|
"learning_rate": 1.4285714285714285e-05, |
|
"loss": 0.1543, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"eval_accuracy": 0.3170731707317073, |
|
"eval_loss": 2.7736659049987793, |
|
"eval_runtime": 98.6426, |
|
"eval_samples_per_second": 0.416, |
|
"eval_steps_per_second": 0.213, |
|
"step": 1568 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 1.9330545663833618, |
|
"learning_rate": 1.4021164021164022e-05, |
|
"loss": 0.3434, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"grad_norm": 0.2689897119998932, |
|
"learning_rate": 1.3756613756613756e-05, |
|
"loss": 0.0472, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"grad_norm": 0.09701512008905411, |
|
"learning_rate": 1.3492063492063492e-05, |
|
"loss": 0.0198, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"grad_norm": 0.07738398015499115, |
|
"learning_rate": 1.3227513227513228e-05, |
|
"loss": 0.01, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"grad_norm": 0.07364221662282944, |
|
"learning_rate": 1.2962962962962962e-05, |
|
"loss": 0.0218, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"grad_norm": 0.06300120055675507, |
|
"learning_rate": 1.2698412698412699e-05, |
|
"loss": 0.0517, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"grad_norm": 0.2190474420785904, |
|
"learning_rate": 1.2433862433862433e-05, |
|
"loss": 0.0538, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"grad_norm": 13.809772491455078, |
|
"learning_rate": 1.2169312169312169e-05, |
|
"loss": 0.046, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"grad_norm": 5.0279645919799805, |
|
"learning_rate": 1.1904761904761905e-05, |
|
"loss": 0.1286, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"grad_norm": 0.25265267491340637, |
|
"learning_rate": 1.164021164021164e-05, |
|
"loss": 0.0477, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"grad_norm": 12.444536209106445, |
|
"learning_rate": 1.1375661375661376e-05, |
|
"loss": 0.0547, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"grad_norm": 0.8364056348800659, |
|
"learning_rate": 1.1111111111111112e-05, |
|
"loss": 0.0147, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"grad_norm": 0.10071144253015518, |
|
"learning_rate": 1.0846560846560846e-05, |
|
"loss": 0.0119, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"grad_norm": 0.4113464653491974, |
|
"learning_rate": 1.0582010582010582e-05, |
|
"loss": 0.0098, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"grad_norm": 0.29613247513771057, |
|
"learning_rate": 1.0317460317460318e-05, |
|
"loss": 0.0096, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"grad_norm": 0.28390398621559143, |
|
"learning_rate": 1.0052910052910053e-05, |
|
"loss": 0.0841, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"grad_norm": 2.4915988445281982, |
|
"learning_rate": 9.788359788359789e-06, |
|
"loss": 0.0831, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"grad_norm": 0.16037584841251373, |
|
"learning_rate": 9.523809523809523e-06, |
|
"loss": 0.0245, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"grad_norm": 0.21573053300380707, |
|
"learning_rate": 9.259259259259259e-06, |
|
"loss": 0.206, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"grad_norm": 0.03257734328508377, |
|
"learning_rate": 8.994708994708995e-06, |
|
"loss": 0.0387, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"eval_accuracy": 0.36585365853658536, |
|
"eval_loss": 2.6675825119018555, |
|
"eval_runtime": 98.7265, |
|
"eval_samples_per_second": 0.415, |
|
"eval_steps_per_second": 0.213, |
|
"step": 1764 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 0.25477078557014465, |
|
"learning_rate": 8.73015873015873e-06, |
|
"loss": 0.0048, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"grad_norm": 0.08968080580234528, |
|
"learning_rate": 8.465608465608466e-06, |
|
"loss": 0.0213, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"grad_norm": 0.12380950897932053, |
|
"learning_rate": 8.201058201058202e-06, |
|
"loss": 0.0471, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"grad_norm": 0.1933353841304779, |
|
"learning_rate": 7.936507936507936e-06, |
|
"loss": 0.3223, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"grad_norm": 0.09426797181367874, |
|
"learning_rate": 7.671957671957672e-06, |
|
"loss": 0.0566, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"grad_norm": 0.1309937983751297, |
|
"learning_rate": 7.4074074074074075e-06, |
|
"loss": 0.0268, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"grad_norm": 0.7364773750305176, |
|
"learning_rate": 7.142857142857143e-06, |
|
"loss": 0.0418, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"grad_norm": 0.38073354959487915, |
|
"learning_rate": 6.878306878306878e-06, |
|
"loss": 0.0053, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"grad_norm": 0.09456183016300201, |
|
"learning_rate": 6.613756613756614e-06, |
|
"loss": 0.07, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"grad_norm": 0.9321669936180115, |
|
"learning_rate": 6.349206349206349e-06, |
|
"loss": 0.0108, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"grad_norm": 0.7253817915916443, |
|
"learning_rate": 6.0846560846560845e-06, |
|
"loss": 0.0041, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"grad_norm": 3.511643886566162, |
|
"learning_rate": 5.82010582010582e-06, |
|
"loss": 0.057, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"grad_norm": 0.09484612196683884, |
|
"learning_rate": 5.555555555555556e-06, |
|
"loss": 0.05, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"grad_norm": 0.05964544415473938, |
|
"learning_rate": 5.291005291005291e-06, |
|
"loss": 0.0089, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"grad_norm": 0.03321171924471855, |
|
"learning_rate": 5.026455026455026e-06, |
|
"loss": 0.0041, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"grad_norm": 0.12247146666049957, |
|
"learning_rate": 4.7619047619047615e-06, |
|
"loss": 0.0272, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"grad_norm": 6.011055946350098, |
|
"learning_rate": 4.497354497354498e-06, |
|
"loss": 0.0125, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"grad_norm": 0.07328727841377258, |
|
"learning_rate": 4.232804232804233e-06, |
|
"loss": 0.0493, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"grad_norm": 0.03230283781886101, |
|
"learning_rate": 3.968253968253968e-06, |
|
"loss": 0.0104, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"grad_norm": 0.10232985019683838, |
|
"learning_rate": 3.7037037037037037e-06, |
|
"loss": 0.0101, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"eval_accuracy": 0.34146341463414637, |
|
"eval_loss": 2.7895021438598633, |
|
"eval_runtime": 98.3982, |
|
"eval_samples_per_second": 0.417, |
|
"eval_steps_per_second": 0.213, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 5.483851909637451, |
|
"learning_rate": 3.439153439153439e-06, |
|
"loss": 0.2193, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"grad_norm": 2.402463912963867, |
|
"learning_rate": 3.1746031746031746e-06, |
|
"loss": 0.005, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"grad_norm": 0.08023607730865479, |
|
"learning_rate": 2.91005291005291e-06, |
|
"loss": 0.0125, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"grad_norm": 0.023654110729694366, |
|
"learning_rate": 2.6455026455026455e-06, |
|
"loss": 0.0094, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"grad_norm": 0.052570246160030365, |
|
"learning_rate": 2.3809523809523808e-06, |
|
"loss": 0.11, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"grad_norm": 0.08791643381118774, |
|
"learning_rate": 2.1164021164021164e-06, |
|
"loss": 0.0189, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"grad_norm": 0.21485354006290436, |
|
"learning_rate": 1.8518518518518519e-06, |
|
"loss": 0.0285, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"grad_norm": 0.09601633995771408, |
|
"learning_rate": 1.5873015873015873e-06, |
|
"loss": 0.0326, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"grad_norm": 0.05465522035956383, |
|
"learning_rate": 1.3227513227513228e-06, |
|
"loss": 0.0235, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 10.05, |
|
"grad_norm": 0.12677225470542908, |
|
"learning_rate": 1.0582010582010582e-06, |
|
"loss": 0.0122, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 10.05, |
|
"grad_norm": 0.05187565088272095, |
|
"learning_rate": 7.936507936507937e-07, |
|
"loss": 0.0277, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 10.06, |
|
"grad_norm": 0.02051164023578167, |
|
"learning_rate": 5.291005291005291e-07, |
|
"loss": 0.1015, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 10.06, |
|
"grad_norm": 0.02192351035773754, |
|
"learning_rate": 2.6455026455026455e-07, |
|
"loss": 0.1089, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"grad_norm": 0.19547943770885468, |
|
"learning_rate": 0.0, |
|
"loss": 0.0662, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"eval_accuracy": 0.34146341463414637, |
|
"eval_loss": 2.772752523422241, |
|
"eval_runtime": 100.1099, |
|
"eval_samples_per_second": 0.41, |
|
"eval_steps_per_second": 0.21, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"step": 2100, |
|
"total_flos": 1.073544417178878e+19, |
|
"train_loss": 0.6815972964820408, |
|
"train_runtime": 32185.1971, |
|
"train_samples_per_second": 0.13, |
|
"train_steps_per_second": 0.065 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"eval_accuracy": 0.24271844660194175, |
|
"eval_loss": 1.9485329389572144, |
|
"eval_runtime": 275.3036, |
|
"eval_samples_per_second": 0.374, |
|
"eval_steps_per_second": 0.189, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"eval_accuracy": 0.24271844660194175, |
|
"eval_loss": 1.9485328197479248, |
|
"eval_runtime": 251.8324, |
|
"eval_samples_per_second": 0.409, |
|
"eval_steps_per_second": 0.206, |
|
"step": 2100 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 9223372036854775807, |
|
"save_steps": 500, |
|
"total_flos": 1.073544417178878e+19, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|