|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.6455026455026456, |
|
"eval_steps": 500, |
|
"global_step": 3500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.7509137392044067, |
|
"learning_rate": 9.974804736709499e-06, |
|
"loss": 1.9586, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.5435916185379028, |
|
"learning_rate": 9.949609473418998e-06, |
|
"loss": 1.764, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.209004521369934, |
|
"learning_rate": 9.924414210128497e-06, |
|
"loss": 1.5738, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.036686897277832, |
|
"learning_rate": 9.899218946837995e-06, |
|
"loss": 1.456, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.8039250373840332, |
|
"learning_rate": 9.874023683547493e-06, |
|
"loss": 1.3517, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.668268620967865, |
|
"learning_rate": 9.848828420256992e-06, |
|
"loss": 1.2773, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.6465680003166199, |
|
"learning_rate": 9.823633156966492e-06, |
|
"loss": 1.2494, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.669222891330719, |
|
"learning_rate": 9.79843789367599e-06, |
|
"loss": 1.2465, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.7393785715103149, |
|
"learning_rate": 9.773242630385489e-06, |
|
"loss": 1.2073, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.6821632385253906, |
|
"learning_rate": 9.748047367094986e-06, |
|
"loss": 1.1725, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.698268711566925, |
|
"learning_rate": 9.722852103804486e-06, |
|
"loss": 1.1724, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.6640620231628418, |
|
"learning_rate": 9.697656840513983e-06, |
|
"loss": 1.1656, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.8017461895942688, |
|
"learning_rate": 9.672461577223483e-06, |
|
"loss": 1.1764, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.6652551293373108, |
|
"learning_rate": 9.64726631393298e-06, |
|
"loss": 1.1706, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8237988352775574, |
|
"learning_rate": 9.62207105064248e-06, |
|
"loss": 1.1231, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.8982694149017334, |
|
"learning_rate": 9.59687578735198e-06, |
|
"loss": 1.1417, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.7859475016593933, |
|
"learning_rate": 9.571680524061477e-06, |
|
"loss": 1.1382, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9249159693717957, |
|
"learning_rate": 9.546485260770976e-06, |
|
"loss": 1.1363, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.221252202987671, |
|
"learning_rate": 9.521289997480474e-06, |
|
"loss": 1.1382, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 4.636524200439453, |
|
"learning_rate": 9.498614260519024e-06, |
|
"loss": 1.0979, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.7521413564682007, |
|
"learning_rate": 9.473418997228522e-06, |
|
"loss": 1.0645, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.8328731060028076, |
|
"learning_rate": 9.44822373393802e-06, |
|
"loss": 1.0585, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.8342303037643433, |
|
"learning_rate": 9.423028470647519e-06, |
|
"loss": 1.0655, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8196636438369751, |
|
"learning_rate": 9.397833207357018e-06, |
|
"loss": 1.0221, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.7511851787567139, |
|
"learning_rate": 9.372637944066516e-06, |
|
"loss": 1.046, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8341330885887146, |
|
"learning_rate": 9.347442680776014e-06, |
|
"loss": 1.0331, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9003998637199402, |
|
"learning_rate": 9.322247417485513e-06, |
|
"loss": 1.0386, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.79107266664505, |
|
"learning_rate": 9.297052154195013e-06, |
|
"loss": 1.0152, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.7756459712982178, |
|
"learning_rate": 9.27185689090451e-06, |
|
"loss": 1.0224, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.762424647808075, |
|
"learning_rate": 9.24666162761401e-06, |
|
"loss": 1.0517, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.7143459916114807, |
|
"learning_rate": 9.221466364323507e-06, |
|
"loss": 1.0106, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.7883962988853455, |
|
"learning_rate": 9.196271101033007e-06, |
|
"loss": 1.0004, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.7985430359840393, |
|
"learning_rate": 9.171075837742504e-06, |
|
"loss": 1.0171, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8012735247612, |
|
"learning_rate": 9.145880574452004e-06, |
|
"loss": 1.0106, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.7331275939941406, |
|
"learning_rate": 9.120685311161502e-06, |
|
"loss": 1.0044, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8667876124382019, |
|
"learning_rate": 9.095490047871001e-06, |
|
"loss": 1.0195, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.8839796781539917, |
|
"learning_rate": 9.0702947845805e-06, |
|
"loss": 1.0145, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8900341391563416, |
|
"learning_rate": 9.045099521289998e-06, |
|
"loss": 1.0171, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8628876805305481, |
|
"learning_rate": 9.019904257999496e-06, |
|
"loss": 1.0039, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8550145626068115, |
|
"learning_rate": 8.994708994708995e-06, |
|
"loss": 1.015, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8471419215202332, |
|
"learning_rate": 8.969513731418495e-06, |
|
"loss": 1.0114, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8051427006721497, |
|
"learning_rate": 8.944318468127992e-06, |
|
"loss": 0.9998, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8776746392250061, |
|
"learning_rate": 8.919123204837492e-06, |
|
"loss": 1.0109, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9933453798294067, |
|
"learning_rate": 8.893927941546991e-06, |
|
"loss": 1.0099, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8829633593559265, |
|
"learning_rate": 8.868732678256489e-06, |
|
"loss": 1.0201, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8752363324165344, |
|
"learning_rate": 8.843537414965987e-06, |
|
"loss": 1.0022, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9361273050308228, |
|
"learning_rate": 8.818342151675486e-06, |
|
"loss": 1.0075, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8231716156005859, |
|
"learning_rate": 8.793146888384985e-06, |
|
"loss": 0.984, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.8183861374855042, |
|
"learning_rate": 8.767951625094483e-06, |
|
"loss": 0.9871, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9514871835708618, |
|
"learning_rate": 8.74275636180398e-06, |
|
"loss": 1.024, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9753857254981995, |
|
"learning_rate": 8.71756109851348e-06, |
|
"loss": 0.9926, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9237353205680847, |
|
"learning_rate": 8.69236583522298e-06, |
|
"loss": 0.9709, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9065663814544678, |
|
"learning_rate": 8.667170571932477e-06, |
|
"loss": 0.9848, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8715184926986694, |
|
"learning_rate": 8.641975308641975e-06, |
|
"loss": 1.0091, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8322390913963318, |
|
"learning_rate": 8.616780045351474e-06, |
|
"loss": 0.9876, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9560284614562988, |
|
"learning_rate": 8.591584782060974e-06, |
|
"loss": 1.0234, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8381783366203308, |
|
"learning_rate": 8.566389518770471e-06, |
|
"loss": 0.9788, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8607673645019531, |
|
"learning_rate": 8.541194255479971e-06, |
|
"loss": 0.9816, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8169811367988586, |
|
"learning_rate": 8.515998992189469e-06, |
|
"loss": 1.0092, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.9223540425300598, |
|
"learning_rate": 8.490803728898968e-06, |
|
"loss": 0.9885, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.7612660527229309, |
|
"learning_rate": 8.465608465608466e-06, |
|
"loss": 0.97, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.860916256904602, |
|
"learning_rate": 8.440413202317965e-06, |
|
"loss": 0.9921, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.807750403881073, |
|
"learning_rate": 8.415217939027463e-06, |
|
"loss": 0.987, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.7962011694908142, |
|
"learning_rate": 8.390022675736962e-06, |
|
"loss": 1.0134, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8323624730110168, |
|
"learning_rate": 8.364827412446462e-06, |
|
"loss": 0.9769, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.9545472860336304, |
|
"learning_rate": 8.33963214915596e-06, |
|
"loss": 0.9862, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.0507547855377197, |
|
"learning_rate": 8.314436885865457e-06, |
|
"loss": 0.9977, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.9005000591278076, |
|
"learning_rate": 8.289241622574956e-06, |
|
"loss": 0.9897, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8091966509819031, |
|
"learning_rate": 8.264046359284456e-06, |
|
"loss": 1.0188, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8602038025856018, |
|
"learning_rate": 8.238851095993954e-06, |
|
"loss": 0.9997, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.9347642660140991, |
|
"learning_rate": 8.213655832703453e-06, |
|
"loss": 0.9904, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8802648782730103, |
|
"learning_rate": 8.188460569412952e-06, |
|
"loss": 0.9765, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8487595915794373, |
|
"learning_rate": 8.16326530612245e-06, |
|
"loss": 1.0184, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.9282223582267761, |
|
"learning_rate": 8.138070042831948e-06, |
|
"loss": 1.0047, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 1.0162054300308228, |
|
"learning_rate": 8.112874779541447e-06, |
|
"loss": 0.9909, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8241488337516785, |
|
"learning_rate": 8.087679516250947e-06, |
|
"loss": 0.9976, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.069097638130188, |
|
"learning_rate": 8.062484252960444e-06, |
|
"loss": 0.9874, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.9084659218788147, |
|
"learning_rate": 8.037288989669942e-06, |
|
"loss": 0.9751, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.9863806962966919, |
|
"learning_rate": 8.012093726379441e-06, |
|
"loss": 0.9815, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.9838799834251404, |
|
"learning_rate": 7.98689846308894e-06, |
|
"loss": 1.0019, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.0254228115081787, |
|
"learning_rate": 7.961703199798438e-06, |
|
"loss": 0.9918, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.9409896731376648, |
|
"learning_rate": 7.936507936507936e-06, |
|
"loss": 0.9899, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9909108281135559, |
|
"learning_rate": 7.911312673217436e-06, |
|
"loss": 0.9666, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9930652976036072, |
|
"learning_rate": 7.886117409926935e-06, |
|
"loss": 0.9643, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.983561098575592, |
|
"learning_rate": 7.860922146636433e-06, |
|
"loss": 0.9895, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.9257413148880005, |
|
"learning_rate": 7.835726883345932e-06, |
|
"loss": 0.9933, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.9350329041481018, |
|
"learning_rate": 7.81053162005543e-06, |
|
"loss": 0.9725, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 1.0069069862365723, |
|
"learning_rate": 7.78533635676493e-06, |
|
"loss": 0.9739, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.9367712736129761, |
|
"learning_rate": 7.760141093474427e-06, |
|
"loss": 0.9835, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.867667555809021, |
|
"learning_rate": 7.734945830183926e-06, |
|
"loss": 0.9937, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.9225870370864868, |
|
"learning_rate": 7.709750566893424e-06, |
|
"loss": 0.9771, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.9846392869949341, |
|
"learning_rate": 7.684555303602923e-06, |
|
"loss": 0.9988, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8313792943954468, |
|
"learning_rate": 7.659360040312423e-06, |
|
"loss": 0.9594, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.9007426500320435, |
|
"learning_rate": 7.63416477702192e-06, |
|
"loss": 0.9641, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.9949538111686707, |
|
"learning_rate": 7.608969513731419e-06, |
|
"loss": 0.9981, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.0600723028182983, |
|
"learning_rate": 7.583774250440918e-06, |
|
"loss": 0.9799, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.124242901802063, |
|
"learning_rate": 7.558578987150417e-06, |
|
"loss": 0.9923, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 1.1153829097747803, |
|
"learning_rate": 7.533383723859915e-06, |
|
"loss": 0.976, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.996911346912384, |
|
"learning_rate": 7.508188460569413e-06, |
|
"loss": 0.9952, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.866919219493866, |
|
"learning_rate": 7.482993197278913e-06, |
|
"loss": 0.9708, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.0361676216125488, |
|
"learning_rate": 7.457797933988411e-06, |
|
"loss": 0.9757, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.0245063304901123, |
|
"learning_rate": 7.432602670697909e-06, |
|
"loss": 0.9879, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.9136086106300354, |
|
"learning_rate": 7.4074074074074075e-06, |
|
"loss": 0.962, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.1000611782073975, |
|
"learning_rate": 7.382212144116907e-06, |
|
"loss": 0.9743, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.0377795696258545, |
|
"learning_rate": 7.3570168808264054e-06, |
|
"loss": 0.9392, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.9714633226394653, |
|
"learning_rate": 7.331821617535904e-06, |
|
"loss": 0.964, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 1.027772068977356, |
|
"learning_rate": 7.306626354245403e-06, |
|
"loss": 0.9784, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9620116949081421, |
|
"learning_rate": 7.281431090954901e-06, |
|
"loss": 0.9803, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.0313563346862793, |
|
"learning_rate": 7.2562358276644e-06, |
|
"loss": 0.9816, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 1.058566927909851, |
|
"learning_rate": 7.231040564373898e-06, |
|
"loss": 0.9692, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9281766414642334, |
|
"learning_rate": 7.205845301083398e-06, |
|
"loss": 0.9508, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.960678219795227, |
|
"learning_rate": 7.180650037792895e-06, |
|
"loss": 0.9437, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.9975042343139648, |
|
"learning_rate": 7.155454774502394e-06, |
|
"loss": 0.97, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.9827629327774048, |
|
"learning_rate": 7.130259511211892e-06, |
|
"loss": 0.9709, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.084519386291504, |
|
"learning_rate": 7.105064247921392e-06, |
|
"loss": 0.9736, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 1.1209474802017212, |
|
"learning_rate": 7.0798689846308895e-06, |
|
"loss": 0.9722, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 1.0013864040374756, |
|
"learning_rate": 7.054673721340388e-06, |
|
"loss": 0.9671, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.9566646218299866, |
|
"learning_rate": 7.0294784580498875e-06, |
|
"loss": 0.9767, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.9742752909660339, |
|
"learning_rate": 7.004283194759386e-06, |
|
"loss": 0.9703, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.1095930337905884, |
|
"learning_rate": 6.979087931468885e-06, |
|
"loss": 0.9817, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.0295912027359009, |
|
"learning_rate": 6.953892668178382e-06, |
|
"loss": 0.9756, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.0642352104187012, |
|
"learning_rate": 6.928697404887882e-06, |
|
"loss": 0.9503, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.157819390296936, |
|
"learning_rate": 6.90350214159738e-06, |
|
"loss": 0.9765, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.0977922677993774, |
|
"learning_rate": 6.878306878306879e-06, |
|
"loss": 0.9706, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.9640308022499084, |
|
"learning_rate": 6.853111615016378e-06, |
|
"loss": 0.9578, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 1.0732492208480835, |
|
"learning_rate": 6.827916351725876e-06, |
|
"loss": 0.9455, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.9846688508987427, |
|
"learning_rate": 6.8027210884353745e-06, |
|
"loss": 0.9764, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.8813304901123047, |
|
"learning_rate": 6.777525825144873e-06, |
|
"loss": 0.9842, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.9812163710594177, |
|
"learning_rate": 6.752330561854372e-06, |
|
"loss": 0.9593, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.0121666193008423, |
|
"learning_rate": 6.72713529856387e-06, |
|
"loss": 0.9623, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 1.0399516820907593, |
|
"learning_rate": 6.701940035273369e-06, |
|
"loss": 0.9665, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.9701886773109436, |
|
"learning_rate": 6.676744771982868e-06, |
|
"loss": 0.978, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.0398330688476562, |
|
"learning_rate": 6.651549508692367e-06, |
|
"loss": 0.9515, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.038935661315918, |
|
"learning_rate": 6.626354245401865e-06, |
|
"loss": 0.958, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 1.0119668245315552, |
|
"learning_rate": 6.601158982111363e-06, |
|
"loss": 0.9347, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.9660181403160095, |
|
"learning_rate": 6.575963718820862e-06, |
|
"loss": 0.9577, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.9453220963478088, |
|
"learning_rate": 6.550768455530361e-06, |
|
"loss": 0.9375, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.9495384693145752, |
|
"learning_rate": 6.525573192239859e-06, |
|
"loss": 0.9311, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 1.1077016592025757, |
|
"learning_rate": 6.500377928949359e-06, |
|
"loss": 0.9489, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 1.0165854692459106, |
|
"learning_rate": 6.4751826656588565e-06, |
|
"loss": 0.958, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 1.3001985549926758, |
|
"learning_rate": 6.449987402368355e-06, |
|
"loss": 0.9325, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 1.0839056968688965, |
|
"learning_rate": 6.424792139077854e-06, |
|
"loss": 0.9769, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 1.0319207906723022, |
|
"learning_rate": 6.399596875787353e-06, |
|
"loss": 0.9574, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 1.0123823881149292, |
|
"learning_rate": 6.374401612496851e-06, |
|
"loss": 0.9315, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 1.1985503435134888, |
|
"learning_rate": 6.349206349206349e-06, |
|
"loss": 0.9354, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 1.1550285816192627, |
|
"learning_rate": 6.324011085915849e-06, |
|
"loss": 0.9344, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 1.0712833404541016, |
|
"learning_rate": 6.298815822625347e-06, |
|
"loss": 0.9444, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 1.0778782367706299, |
|
"learning_rate": 6.273620559334846e-06, |
|
"loss": 0.935, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.0286056995391846, |
|
"learning_rate": 6.2484252960443435e-06, |
|
"loss": 0.9606, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.1475175619125366, |
|
"learning_rate": 6.223230032753843e-06, |
|
"loss": 0.9354, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 1.1092008352279663, |
|
"learning_rate": 6.1980347694633414e-06, |
|
"loss": 0.9556, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 1.0508161783218384, |
|
"learning_rate": 6.17283950617284e-06, |
|
"loss": 0.9388, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.0714831352233887, |
|
"learning_rate": 6.147644242882339e-06, |
|
"loss": 0.9334, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.0550581216812134, |
|
"learning_rate": 6.122448979591837e-06, |
|
"loss": 0.9398, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 1.1077096462249756, |
|
"learning_rate": 6.097253716301336e-06, |
|
"loss": 0.9486, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.0329755544662476, |
|
"learning_rate": 6.072058453010834e-06, |
|
"loss": 0.9635, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 1.085105061531067, |
|
"learning_rate": 6.046863189720334e-06, |
|
"loss": 0.9394, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 1.0389838218688965, |
|
"learning_rate": 6.021667926429831e-06, |
|
"loss": 0.9474, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.1028958559036255, |
|
"learning_rate": 5.99647266313933e-06, |
|
"loss": 0.9364, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 1.0558817386627197, |
|
"learning_rate": 5.971277399848829e-06, |
|
"loss": 0.9276, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 1.0861153602600098, |
|
"learning_rate": 5.946082136558328e-06, |
|
"loss": 0.9375, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 1.1104577779769897, |
|
"learning_rate": 5.920886873267826e-06, |
|
"loss": 0.948, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 1.0775477886199951, |
|
"learning_rate": 5.895691609977324e-06, |
|
"loss": 0.9334, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.0878546237945557, |
|
"learning_rate": 5.8704963466868235e-06, |
|
"loss": 0.952, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.0840810537338257, |
|
"learning_rate": 5.845301083396322e-06, |
|
"loss": 0.9312, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.2151501178741455, |
|
"learning_rate": 5.820105820105821e-06, |
|
"loss": 0.9549, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 1.0833278894424438, |
|
"learning_rate": 5.79491055681532e-06, |
|
"loss": 0.9557, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 1.0161901712417603, |
|
"learning_rate": 5.769715293524818e-06, |
|
"loss": 0.9312, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 1.075750708580017, |
|
"learning_rate": 5.744520030234316e-06, |
|
"loss": 0.9501, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 1.042380928993225, |
|
"learning_rate": 5.719324766943815e-06, |
|
"loss": 0.9381, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 1.009562373161316, |
|
"learning_rate": 5.694129503653314e-06, |
|
"loss": 0.9305, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 1.1201280355453491, |
|
"learning_rate": 5.668934240362812e-06, |
|
"loss": 0.9318, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 1.1156421899795532, |
|
"learning_rate": 5.6437389770723105e-06, |
|
"loss": 0.9445, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 1.0913002490997314, |
|
"learning_rate": 5.61854371378181e-06, |
|
"loss": 0.9537, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 1.041400671005249, |
|
"learning_rate": 5.593348450491308e-06, |
|
"loss": 0.9464, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 1.0327285528182983, |
|
"learning_rate": 5.568153187200807e-06, |
|
"loss": 0.9361, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 1.1421902179718018, |
|
"learning_rate": 5.542957923910305e-06, |
|
"loss": 0.9764, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 1.202650785446167, |
|
"learning_rate": 5.517762660619804e-06, |
|
"loss": 0.9553, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 1.0642273426055908, |
|
"learning_rate": 5.492567397329303e-06, |
|
"loss": 0.9485, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 1.0749919414520264, |
|
"learning_rate": 5.467372134038801e-06, |
|
"loss": 0.9418, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 1.0974488258361816, |
|
"learning_rate": 5.442176870748301e-06, |
|
"loss": 0.9539, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 1.1204264163970947, |
|
"learning_rate": 5.416981607457798e-06, |
|
"loss": 0.9357, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 1.268764853477478, |
|
"learning_rate": 5.391786344167297e-06, |
|
"loss": 0.9512, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 1.159050703048706, |
|
"learning_rate": 5.366591080876795e-06, |
|
"loss": 0.9402, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 1.2613215446472168, |
|
"learning_rate": 5.341395817586295e-06, |
|
"loss": 0.9411, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 1.1373072862625122, |
|
"learning_rate": 5.3162005542957925e-06, |
|
"loss": 0.9752, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 1.144242763519287, |
|
"learning_rate": 5.291005291005291e-06, |
|
"loss": 0.9471, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 1.0989221334457397, |
|
"learning_rate": 5.2658100277147905e-06, |
|
"loss": 0.9348, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 1.0864170789718628, |
|
"learning_rate": 5.240614764424289e-06, |
|
"loss": 0.9435, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 1.1140406131744385, |
|
"learning_rate": 5.2154195011337876e-06, |
|
"loss": 0.9541, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 1.1856478452682495, |
|
"learning_rate": 5.190224237843285e-06, |
|
"loss": 0.9151, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 1.2324665784835815, |
|
"learning_rate": 5.165028974552785e-06, |
|
"loss": 0.9411, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 1.038913369178772, |
|
"learning_rate": 5.139833711262283e-06, |
|
"loss": 0.9487, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.040578842163086, |
|
"learning_rate": 5.114638447971782e-06, |
|
"loss": 0.943, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.289865493774414, |
|
"learning_rate": 5.089443184681281e-06, |
|
"loss": 0.9382, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 1.1952651739120483, |
|
"learning_rate": 5.064247921390779e-06, |
|
"loss": 0.9453, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 1.1922310590744019, |
|
"learning_rate": 5.0390526581002774e-06, |
|
"loss": 0.9242, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 1.1691875457763672, |
|
"learning_rate": 5.013857394809776e-06, |
|
"loss": 0.932, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 1.254301905632019, |
|
"learning_rate": 4.9886621315192745e-06, |
|
"loss": 0.9341, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 1.0548466444015503, |
|
"learning_rate": 4.963466868228773e-06, |
|
"loss": 0.9496, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 1.2470088005065918, |
|
"learning_rate": 4.938271604938272e-06, |
|
"loss": 0.9488, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 1.2311805486679077, |
|
"learning_rate": 4.91307634164777e-06, |
|
"loss": 0.942, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 1.1870156526565552, |
|
"learning_rate": 4.88788107835727e-06, |
|
"loss": 0.94, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 1.060556173324585, |
|
"learning_rate": 4.862685815066767e-06, |
|
"loss": 0.9223, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 1.1396002769470215, |
|
"learning_rate": 4.837490551776267e-06, |
|
"loss": 0.9588, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 1.2042090892791748, |
|
"learning_rate": 4.812295288485765e-06, |
|
"loss": 0.9534, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 1.1195515394210815, |
|
"learning_rate": 4.787100025195264e-06, |
|
"loss": 0.9528, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 1.290823221206665, |
|
"learning_rate": 4.761904761904762e-06, |
|
"loss": 0.9514, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 1.0271083116531372, |
|
"learning_rate": 4.736709498614261e-06, |
|
"loss": 0.9699, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.2110365629196167, |
|
"learning_rate": 4.7115142353237595e-06, |
|
"loss": 0.9358, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.1250057220458984, |
|
"learning_rate": 4.686318972033258e-06, |
|
"loss": 0.9491, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.1912841796875, |
|
"learning_rate": 4.661123708742757e-06, |
|
"loss": 0.9332, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 1.0299352407455444, |
|
"learning_rate": 4.635928445452255e-06, |
|
"loss": 0.9577, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 1.0603456497192383, |
|
"learning_rate": 4.610733182161754e-06, |
|
"loss": 0.9332, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.1462222337722778, |
|
"learning_rate": 4.585537918871252e-06, |
|
"loss": 0.9343, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.0181183815002441, |
|
"learning_rate": 4.560342655580751e-06, |
|
"loss": 0.9365, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 1.2271602153778076, |
|
"learning_rate": 4.53514739229025e-06, |
|
"loss": 0.9171, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 1.1205991506576538, |
|
"learning_rate": 4.509952128999748e-06, |
|
"loss": 0.9299, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.1556931734085083, |
|
"learning_rate": 4.484756865709247e-06, |
|
"loss": 0.9652, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.0930700302124023, |
|
"learning_rate": 4.459561602418746e-06, |
|
"loss": 0.9259, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 1.207714319229126, |
|
"learning_rate": 4.434366339128244e-06, |
|
"loss": 0.9529, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 1.1223605871200562, |
|
"learning_rate": 4.409171075837743e-06, |
|
"loss": 0.9397, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 1.2369412183761597, |
|
"learning_rate": 4.3839758125472415e-06, |
|
"loss": 0.9246, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 1.232971429824829, |
|
"learning_rate": 4.35878054925674e-06, |
|
"loss": 0.9349, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 1.1932899951934814, |
|
"learning_rate": 4.333585285966239e-06, |
|
"loss": 0.9495, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 1.2230783700942993, |
|
"learning_rate": 4.308390022675737e-06, |
|
"loss": 0.9205, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.2365373373031616, |
|
"learning_rate": 4.283194759385236e-06, |
|
"loss": 0.9214, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.132912278175354, |
|
"learning_rate": 4.257999496094734e-06, |
|
"loss": 0.9154, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 1.2143046855926514, |
|
"learning_rate": 4.232804232804233e-06, |
|
"loss": 0.9485, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 1.2785279750823975, |
|
"learning_rate": 4.207608969513731e-06, |
|
"loss": 0.9462, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.3493692874908447, |
|
"learning_rate": 4.182413706223231e-06, |
|
"loss": 0.9382, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.1612365245819092, |
|
"learning_rate": 4.1572184429327285e-06, |
|
"loss": 0.9407, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 1.2210137844085693, |
|
"learning_rate": 4.132023179642228e-06, |
|
"loss": 0.9196, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 1.1422805786132812, |
|
"learning_rate": 4.1068279163517265e-06, |
|
"loss": 0.9426, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 1.1776541471481323, |
|
"learning_rate": 4.081632653061225e-06, |
|
"loss": 0.9436, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 1.2455426454544067, |
|
"learning_rate": 4.0564373897707236e-06, |
|
"loss": 0.9299, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 1.146103858947754, |
|
"learning_rate": 4.031242126480222e-06, |
|
"loss": 0.9353, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 1.1490944623947144, |
|
"learning_rate": 4.006046863189721e-06, |
|
"loss": 0.9373, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 1.1841084957122803, |
|
"learning_rate": 3.980851599899219e-06, |
|
"loss": 0.9214, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 1.360938310623169, |
|
"learning_rate": 3.955656336608718e-06, |
|
"loss": 0.9233, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 1.1501877307891846, |
|
"learning_rate": 3.930461073318216e-06, |
|
"loss": 0.9149, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 1.3008352518081665, |
|
"learning_rate": 3.905265810027715e-06, |
|
"loss": 0.9592, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 1.2323644161224365, |
|
"learning_rate": 3.8800705467372134e-06, |
|
"loss": 0.9295, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 1.1721792221069336, |
|
"learning_rate": 3.854875283446712e-06, |
|
"loss": 0.9431, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.2879486083984375, |
|
"learning_rate": 3.829680020156211e-06, |
|
"loss": 0.9463, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 1.1511075496673584, |
|
"learning_rate": 3.8044847568657095e-06, |
|
"loss": 0.9664, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 1.3236780166625977, |
|
"learning_rate": 3.7792894935752085e-06, |
|
"loss": 0.9368, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 1.1940069198608398, |
|
"learning_rate": 3.7540942302847066e-06, |
|
"loss": 0.9548, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 1.1677494049072266, |
|
"learning_rate": 3.7288989669942056e-06, |
|
"loss": 0.9592, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 1.1518640518188477, |
|
"learning_rate": 3.7037037037037037e-06, |
|
"loss": 0.9497, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 1.159209966659546, |
|
"learning_rate": 3.6785084404132027e-06, |
|
"loss": 0.9368, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 1.2133903503417969, |
|
"learning_rate": 3.6533131771227017e-06, |
|
"loss": 0.9257, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 1.329829454421997, |
|
"learning_rate": 3.6281179138322e-06, |
|
"loss": 0.9544, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 1.264733910560608, |
|
"learning_rate": 3.602922650541699e-06, |
|
"loss": 0.9525, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.2355499267578125, |
|
"learning_rate": 3.577727387251197e-06, |
|
"loss": 0.95, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.1711816787719727, |
|
"learning_rate": 3.552532123960696e-06, |
|
"loss": 0.9443, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 1.1221129894256592, |
|
"learning_rate": 3.527336860670194e-06, |
|
"loss": 0.9422, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 1.154335379600525, |
|
"learning_rate": 3.502141597379693e-06, |
|
"loss": 0.9462, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 1.2431056499481201, |
|
"learning_rate": 3.476946334089191e-06, |
|
"loss": 0.9305, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.1966361999511719, |
|
"learning_rate": 3.45175107079869e-06, |
|
"loss": 0.9537, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.2401050329208374, |
|
"learning_rate": 3.426555807508189e-06, |
|
"loss": 0.9262, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 1.1483770608901978, |
|
"learning_rate": 3.4013605442176872e-06, |
|
"loss": 0.9515, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 1.1387736797332764, |
|
"learning_rate": 3.376165280927186e-06, |
|
"loss": 0.9293, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 1.2311975955963135, |
|
"learning_rate": 3.3509700176366843e-06, |
|
"loss": 0.9201, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 1.1818970441818237, |
|
"learning_rate": 3.3257747543461833e-06, |
|
"loss": 0.9439, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 1.3574014902114868, |
|
"learning_rate": 3.3005794910556814e-06, |
|
"loss": 0.9235, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 1.1797312498092651, |
|
"learning_rate": 3.2753842277651804e-06, |
|
"loss": 0.9235, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 1.225716233253479, |
|
"learning_rate": 3.2501889644746794e-06, |
|
"loss": 0.9363, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 1.1875022649765015, |
|
"learning_rate": 3.2249937011841775e-06, |
|
"loss": 0.9155, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 1.2818158864974976, |
|
"learning_rate": 3.1997984378936765e-06, |
|
"loss": 0.9189, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 1.0735368728637695, |
|
"learning_rate": 3.1746031746031746e-06, |
|
"loss": 0.9183, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 1.2235422134399414, |
|
"learning_rate": 3.1494079113126736e-06, |
|
"loss": 0.9383, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 1.2761168479919434, |
|
"learning_rate": 3.1242126480221717e-06, |
|
"loss": 0.9161, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 1.1925853490829468, |
|
"learning_rate": 3.0990173847316707e-06, |
|
"loss": 0.9133, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 1.1860272884368896, |
|
"learning_rate": 3.0738221214411697e-06, |
|
"loss": 0.9317, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 1.285198450088501, |
|
"learning_rate": 3.048626858150668e-06, |
|
"loss": 0.9332, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 1.220641851425171, |
|
"learning_rate": 3.023431594860167e-06, |
|
"loss": 0.9464, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 1.31607186794281, |
|
"learning_rate": 2.998236331569665e-06, |
|
"loss": 0.9233, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 1.2238725423812866, |
|
"learning_rate": 2.973041068279164e-06, |
|
"loss": 0.9359, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 1.1632503271102905, |
|
"learning_rate": 2.947845804988662e-06, |
|
"loss": 0.9117, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 1.3732521533966064, |
|
"learning_rate": 2.922650541698161e-06, |
|
"loss": 0.9068, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 1.2329943180084229, |
|
"learning_rate": 2.89745527840766e-06, |
|
"loss": 0.9208, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 1.331434726715088, |
|
"learning_rate": 2.872260015117158e-06, |
|
"loss": 0.9353, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.3423517942428589, |
|
"learning_rate": 2.847064751826657e-06, |
|
"loss": 0.9216, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.2505260705947876, |
|
"learning_rate": 2.8218694885361552e-06, |
|
"loss": 0.9267, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 1.1538835763931274, |
|
"learning_rate": 2.796674225245654e-06, |
|
"loss": 0.9195, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 1.3361774682998657, |
|
"learning_rate": 2.7714789619551523e-06, |
|
"loss": 0.9252, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 1.2870272397994995, |
|
"learning_rate": 2.7462836986646513e-06, |
|
"loss": 0.9011, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 1.238709807395935, |
|
"learning_rate": 2.7210884353741503e-06, |
|
"loss": 0.9203, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 1.1434613466262817, |
|
"learning_rate": 2.6958931720836484e-06, |
|
"loss": 0.9263, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 1.2599645853042603, |
|
"learning_rate": 2.6706979087931474e-06, |
|
"loss": 0.9224, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 1.3294990062713623, |
|
"learning_rate": 2.6455026455026455e-06, |
|
"loss": 0.9116, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 1.2516047954559326, |
|
"learning_rate": 2.6203073822121445e-06, |
|
"loss": 0.9287, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 1.2115424871444702, |
|
"learning_rate": 2.5951121189216426e-06, |
|
"loss": 0.9384, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 1.1633473634719849, |
|
"learning_rate": 2.5699168556311416e-06, |
|
"loss": 0.9211, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 1.181342363357544, |
|
"learning_rate": 2.5447215923406406e-06, |
|
"loss": 0.9122, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 1.2719746828079224, |
|
"learning_rate": 2.5195263290501387e-06, |
|
"loss": 0.9302, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 1.292691707611084, |
|
"learning_rate": 2.4943310657596373e-06, |
|
"loss": 0.9226, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 1.1643900871276855, |
|
"learning_rate": 2.469135802469136e-06, |
|
"loss": 0.903, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 1.2362549304962158, |
|
"learning_rate": 2.443940539178635e-06, |
|
"loss": 0.9238, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 1.1967755556106567, |
|
"learning_rate": 2.4187452758881334e-06, |
|
"loss": 0.9372, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 1.1664007902145386, |
|
"learning_rate": 2.393550012597632e-06, |
|
"loss": 0.9249, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 1.39642333984375, |
|
"learning_rate": 2.3683547493071305e-06, |
|
"loss": 0.9381, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 1.2738609313964844, |
|
"learning_rate": 2.343159486016629e-06, |
|
"loss": 0.9345, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 1.2925204038619995, |
|
"learning_rate": 2.3179642227261276e-06, |
|
"loss": 0.9156, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 1.470628261566162, |
|
"learning_rate": 2.292768959435626e-06, |
|
"loss": 0.9128, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 1.280967354774475, |
|
"learning_rate": 2.267573696145125e-06, |
|
"loss": 0.9225, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 1.3174177408218384, |
|
"learning_rate": 2.2423784328546237e-06, |
|
"loss": 0.9138, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 1.3123756647109985, |
|
"learning_rate": 2.217183169564122e-06, |
|
"loss": 0.9284, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 1.232643723487854, |
|
"learning_rate": 2.1919879062736208e-06, |
|
"loss": 0.9201, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 1.207700490951538, |
|
"learning_rate": 2.1667926429831193e-06, |
|
"loss": 0.9061, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 1.2038023471832275, |
|
"learning_rate": 2.141597379692618e-06, |
|
"loss": 0.9437, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 1.4127525091171265, |
|
"learning_rate": 2.1164021164021164e-06, |
|
"loss": 0.9142, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 1.3315386772155762, |
|
"learning_rate": 2.0912068531116154e-06, |
|
"loss": 0.9217, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 1.3743528127670288, |
|
"learning_rate": 2.066011589821114e-06, |
|
"loss": 0.8931, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 1.2260786294937134, |
|
"learning_rate": 2.0408163265306125e-06, |
|
"loss": 0.9224, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.3002612590789795, |
|
"learning_rate": 2.015621063240111e-06, |
|
"loss": 0.9371, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.1567788124084473, |
|
"learning_rate": 1.9904257999496096e-06, |
|
"loss": 0.9244, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 1.2505953311920166, |
|
"learning_rate": 1.965230536659108e-06, |
|
"loss": 0.8985, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 1.3488541841506958, |
|
"learning_rate": 1.9400352733686067e-06, |
|
"loss": 0.9111, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 1.371126651763916, |
|
"learning_rate": 1.9148400100781057e-06, |
|
"loss": 0.9094, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 1.3531619310379028, |
|
"learning_rate": 1.8896447467876043e-06, |
|
"loss": 0.9353, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 1.261234998703003, |
|
"learning_rate": 1.8644494834971028e-06, |
|
"loss": 0.9635, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 1.2285544872283936, |
|
"learning_rate": 1.8392542202066014e-06, |
|
"loss": 0.9104, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 1.2340446710586548, |
|
"learning_rate": 1.8140589569161e-06, |
|
"loss": 0.9275, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 1.1888856887817383, |
|
"learning_rate": 1.7888636936255985e-06, |
|
"loss": 0.8929, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 1.3445130586624146, |
|
"learning_rate": 1.763668430335097e-06, |
|
"loss": 0.9233, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 1.2736765146255493, |
|
"learning_rate": 1.7384731670445956e-06, |
|
"loss": 0.9143, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.1950249671936035, |
|
"learning_rate": 1.7132779037540945e-06, |
|
"loss": 0.8981, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.3372317552566528, |
|
"learning_rate": 1.688082640463593e-06, |
|
"loss": 0.9149, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 1.2550235986709595, |
|
"learning_rate": 1.6628873771730917e-06, |
|
"loss": 0.9448, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 1.3083131313323975, |
|
"learning_rate": 1.6376921138825902e-06, |
|
"loss": 0.926, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 1.2111276388168335, |
|
"learning_rate": 1.6124968505920888e-06, |
|
"loss": 0.9237, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 1.1683611869812012, |
|
"learning_rate": 1.5873015873015873e-06, |
|
"loss": 0.9265, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 1.2934691905975342, |
|
"learning_rate": 1.5621063240110859e-06, |
|
"loss": 0.8935, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 1.2122243642807007, |
|
"learning_rate": 1.5369110607205848e-06, |
|
"loss": 0.9184, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.318337321281433, |
|
"learning_rate": 1.5117157974300834e-06, |
|
"loss": 0.919, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.2197314500808716, |
|
"learning_rate": 1.486520534139582e-06, |
|
"loss": 0.9118, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 1.1137281656265259, |
|
"learning_rate": 1.4613252708490805e-06, |
|
"loss": 0.9312, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 1.2075839042663574, |
|
"learning_rate": 1.436130007558579e-06, |
|
"loss": 0.9258, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 1.379468321800232, |
|
"learning_rate": 1.4109347442680776e-06, |
|
"loss": 0.9163, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 1.3524144887924194, |
|
"learning_rate": 1.3857394809775762e-06, |
|
"loss": 0.9211, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 1.342460036277771, |
|
"learning_rate": 1.3605442176870751e-06, |
|
"loss": 0.9107, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 1.3853424787521362, |
|
"learning_rate": 1.3353489543965737e-06, |
|
"loss": 0.9367, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 1.397850513458252, |
|
"learning_rate": 1.3101536911060723e-06, |
|
"loss": 0.9173, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 1.439322590827942, |
|
"learning_rate": 1.2849584278155708e-06, |
|
"loss": 0.9158, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 1.199365258216858, |
|
"learning_rate": 1.2597631645250694e-06, |
|
"loss": 0.9167, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 1.323708415031433, |
|
"learning_rate": 1.234567901234568e-06, |
|
"loss": 0.9087, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 1.323588252067566, |
|
"learning_rate": 1.2093726379440667e-06, |
|
"loss": 0.9217, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 1.3864338397979736, |
|
"learning_rate": 1.1841773746535652e-06, |
|
"loss": 0.9202, |
|
"step": 3500 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 3969, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 3.1042768180419625e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|