|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9988358556461003, |
|
"eval_steps": 500, |
|
"global_step": 3864, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 8.620689655172414e-08, |
|
"loss": 3.4377, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 4.3103448275862073e-07, |
|
"loss": 3.4043, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 8.620689655172415e-07, |
|
"loss": 3.4252, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 1.2931034482758623e-06, |
|
"loss": 3.3862, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 1.724137931034483e-06, |
|
"loss": 3.3863, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 2.1551724137931035e-06, |
|
"loss": 3.3899, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.546875, |
|
"learning_rate": 2.5862068965517246e-06, |
|
"loss": 3.3691, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 3.017241379310345e-06, |
|
"loss": 3.3793, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 3.448275862068966e-06, |
|
"loss": 3.3317, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 3.8793103448275865e-06, |
|
"loss": 3.3268, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.1484375, |
|
"learning_rate": 4.310344827586207e-06, |
|
"loss": 3.291, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0546875, |
|
"learning_rate": 4.741379310344828e-06, |
|
"loss": 3.2941, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.99609375, |
|
"learning_rate": 5.172413793103449e-06, |
|
"loss": 3.2778, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.91796875, |
|
"learning_rate": 5.603448275862069e-06, |
|
"loss": 3.2182, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.8203125, |
|
"learning_rate": 6.03448275862069e-06, |
|
"loss": 3.2205, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.80078125, |
|
"learning_rate": 6.465517241379311e-06, |
|
"loss": 3.1764, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.76953125, |
|
"learning_rate": 6.896551724137932e-06, |
|
"loss": 3.1814, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.76953125, |
|
"learning_rate": 7.327586206896552e-06, |
|
"loss": 3.1258, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.7265625, |
|
"learning_rate": 7.758620689655173e-06, |
|
"loss": 3.0397, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.7421875, |
|
"learning_rate": 8.189655172413794e-06, |
|
"loss": 3.0763, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.73828125, |
|
"learning_rate": 8.620689655172414e-06, |
|
"loss": 3.0838, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.80078125, |
|
"learning_rate": 9.051724137931036e-06, |
|
"loss": 2.9988, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.72265625, |
|
"learning_rate": 9.482758620689655e-06, |
|
"loss": 2.9653, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.7265625, |
|
"learning_rate": 9.913793103448277e-06, |
|
"loss": 2.9536, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.67578125, |
|
"learning_rate": 9.999971896515836e-06, |
|
"loss": 2.8924, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.65234375, |
|
"learning_rate": 9.999857726652867e-06, |
|
"loss": 2.8778, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.60546875, |
|
"learning_rate": 9.999655735947143e-06, |
|
"loss": 2.8204, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 9.999365927946568e-06, |
|
"loss": 2.8143, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 9.998988307741521e-06, |
|
"loss": 2.8075, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 9.998522881964777e-06, |
|
"loss": 2.7778, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.997969658791384e-06, |
|
"loss": 2.6992, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.997328647938524e-06, |
|
"loss": 2.7048, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 9.996599860665342e-06, |
|
"loss": 2.7525, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 9.995783309772743e-06, |
|
"loss": 2.7152, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.99487900960317e-06, |
|
"loss": 2.6601, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 9.99388697604036e-06, |
|
"loss": 2.6125, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.99280722650905e-06, |
|
"loss": 2.6056, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 9.991639779974688e-06, |
|
"loss": 2.6514, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 9.99038465694308e-06, |
|
"loss": 2.641, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 9.989041879460046e-06, |
|
"loss": 2.6054, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 9.987611471111027e-06, |
|
"loss": 2.596, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 9.986093457020676e-06, |
|
"loss": 2.5794, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 9.984487863852401e-06, |
|
"loss": 2.5771, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 9.982794719807916e-06, |
|
"loss": 2.54, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 9.981014054626737e-06, |
|
"loss": 2.5216, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.979145899585653e-06, |
|
"loss": 2.5178, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 9.977190287498191e-06, |
|
"loss": 2.5356, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 9.975147252714025e-06, |
|
"loss": 2.5361, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.973016831118389e-06, |
|
"loss": 2.5285, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 9.970799060131428e-06, |
|
"loss": 2.4848, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 9.968493978707555e-06, |
|
"loss": 2.4364, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 9.966101627334758e-06, |
|
"loss": 2.4305, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.9636220480339e-06, |
|
"loss": 2.4391, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.961055284357965e-06, |
|
"loss": 2.4398, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.958401381391306e-06, |
|
"loss": 2.5026, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 9.955660385748853e-06, |
|
"loss": 2.4484, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 9.952832345575283e-06, |
|
"loss": 2.4266, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 9.94991731054418e-06, |
|
"loss": 2.4405, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 9.946915331857172e-06, |
|
"loss": 2.4422, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.943826462243019e-06, |
|
"loss": 2.4495, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.940650755956686e-06, |
|
"loss": 2.3972, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.937388268778409e-06, |
|
"loss": 2.4367, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 9.934039058012685e-06, |
|
"loss": 2.3607, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 9.930603182487297e-06, |
|
"loss": 2.393, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 9.927080702552256e-06, |
|
"loss": 2.3503, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.923471680078753e-06, |
|
"loss": 2.392, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 9.919776178458072e-06, |
|
"loss": 2.3518, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 9.915994262600475e-06, |
|
"loss": 2.3383, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.912125998934055e-06, |
|
"loss": 2.3516, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 9.908171455403586e-06, |
|
"loss": 2.3174, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 9.90413070146931e-06, |
|
"loss": 2.3188, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.900003808105729e-06, |
|
"loss": 2.3135, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.375, |
|
"learning_rate": 9.895790847800361e-06, |
|
"loss": 2.3136, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 9.891491894552452e-06, |
|
"loss": 2.3092, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 9.887107023871692e-06, |
|
"loss": 2.2987, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 9.882636312776878e-06, |
|
"loss": 2.3219, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 9.878079839794572e-06, |
|
"loss": 2.2945, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 9.873437684957705e-06, |
|
"loss": 2.3071, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 9.868709929804193e-06, |
|
"loss": 2.3112, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 9.863896657375485e-06, |
|
"loss": 2.2789, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 9.858997952215112e-06, |
|
"loss": 2.2764, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.85401390036721e-06, |
|
"loss": 2.3129, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 9.848944589374993e-06, |
|
"loss": 2.3237, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 9.84379010827923e-06, |
|
"loss": 2.2283, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.375, |
|
"learning_rate": 9.838550547616671e-06, |
|
"loss": 2.2523, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 9.833225999418461e-06, |
|
"loss": 2.2641, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.827816557208523e-06, |
|
"loss": 2.2737, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.822322316001917e-06, |
|
"loss": 2.2647, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.816743372303166e-06, |
|
"loss": 2.2409, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 9.811079824104566e-06, |
|
"loss": 2.2018, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 9.805331770884462e-06, |
|
"loss": 2.2343, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.799499313605506e-06, |
|
"loss": 2.2398, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.793582554712873e-06, |
|
"loss": 2.2474, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.787581598132469e-06, |
|
"loss": 2.2416, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.781496549269108e-06, |
|
"loss": 2.2005, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 9.775327515004656e-06, |
|
"loss": 2.2498, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 9.769074603696153e-06, |
|
"loss": 2.2456, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.762737925173914e-06, |
|
"loss": 2.2294, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.756317590739592e-06, |
|
"loss": 2.1614, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.749813713164233e-06, |
|
"loss": 2.1949, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 9.743226406686293e-06, |
|
"loss": 2.2037, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 9.736555787009622e-06, |
|
"loss": 2.2386, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 9.729801971301443e-06, |
|
"loss": 2.1857, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 9.72296507819029e-06, |
|
"loss": 2.2382, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.716045227763923e-06, |
|
"loss": 2.1812, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 9.70904254156722e-06, |
|
"loss": 2.1713, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.70195714260004e-06, |
|
"loss": 2.1758, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 9.694789155315067e-06, |
|
"loss": 2.1967, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.68753870561562e-06, |
|
"loss": 2.2178, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.680205920853442e-06, |
|
"loss": 2.1908, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 9.67279092982647e-06, |
|
"loss": 2.1777, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.665293862776557e-06, |
|
"loss": 2.1835, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 9.657714851387204e-06, |
|
"loss": 2.1737, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 9.650054028781226e-06, |
|
"loss": 2.1535, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 9.642311529518438e-06, |
|
"loss": 2.1533, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 9.634487489593268e-06, |
|
"loss": 2.1289, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 9.626582046432384e-06, |
|
"loss": 2.1073, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 9.618595338892272e-06, |
|
"loss": 2.1438, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 9.610527507256803e-06, |
|
"loss": 2.1822, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 9.602378693234762e-06, |
|
"loss": 2.1667, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 9.594149039957366e-06, |
|
"loss": 2.1701, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 9.585838691975744e-06, |
|
"loss": 2.1239, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 9.577447795258403e-06, |
|
"loss": 2.19, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 9.56897649718866e-06, |
|
"loss": 2.1609, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.56042494656206e-06, |
|
"loss": 2.1666, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 9.551793293583749e-06, |
|
"loss": 2.1561, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 9.54308168986585e-06, |
|
"loss": 2.1633, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 9.534290288424795e-06, |
|
"loss": 2.1065, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.525419243678633e-06, |
|
"loss": 2.1084, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 9.516468711444323e-06, |
|
"loss": 2.1653, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.507438848934996e-06, |
|
"loss": 2.1056, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 9.49832981475719e-06, |
|
"loss": 2.1215, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 9.48914176890807e-06, |
|
"loss": 2.1386, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.479874872772618e-06, |
|
"loss": 2.1141, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 9.470529289120786e-06, |
|
"loss": 2.151, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 9.461105182104655e-06, |
|
"loss": 2.0897, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.375, |
|
"learning_rate": 9.451602717255536e-06, |
|
"loss": 2.1514, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 9.442022061481076e-06, |
|
"loss": 2.1311, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 9.432363383062314e-06, |
|
"loss": 2.1416, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 9.422626851650733e-06, |
|
"loss": 2.1198, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 9.412812638265279e-06, |
|
"loss": 2.1094, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 9.402920915289355e-06, |
|
"loss": 2.1226, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 9.392951856467795e-06, |
|
"loss": 2.1302, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 9.382905636903809e-06, |
|
"loss": 2.0773, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.372782433055915e-06, |
|
"loss": 2.1493, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 9.362582422734826e-06, |
|
"loss": 2.1009, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 9.352305785100348e-06, |
|
"loss": 2.1114, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 9.341952700658205e-06, |
|
"loss": 2.1021, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.331523351256898e-06, |
|
"loss": 2.0768, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 9.321017920084485e-06, |
|
"loss": 2.0724, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.310436591665383e-06, |
|
"loss": 2.085, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.299779551857118e-06, |
|
"loss": 2.0719, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 9.289046987847058e-06, |
|
"loss": 2.0916, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 9.278239088149128e-06, |
|
"loss": 2.061, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.267356042600506e-06, |
|
"loss": 2.0851, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 9.256398042358274e-06, |
|
"loss": 2.1034, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.245365279896077e-06, |
|
"loss": 2.0681, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.234257949000723e-06, |
|
"loss": 2.072, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 9.223076244768797e-06, |
|
"loss": 2.064, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 9.211820363603226e-06, |
|
"loss": 2.0746, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 9.200490503209831e-06, |
|
"loss": 2.1001, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 9.18908686259385e-06, |
|
"loss": 2.1237, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 9.17760964205645e-06, |
|
"loss": 2.0758, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.166059043191204e-06, |
|
"loss": 2.0421, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 9.154435268880547e-06, |
|
"loss": 2.0514, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 9.142738523292222e-06, |
|
"loss": 2.081, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 9.130969011875686e-06, |
|
"loss": 2.0762, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 9.119126941358502e-06, |
|
"loss": 2.0405, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 9.107212519742714e-06, |
|
"loss": 2.0439, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.095225956301186e-06, |
|
"loss": 2.0508, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 9.083167461573926e-06, |
|
"loss": 2.038, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.071037247364396e-06, |
|
"loss": 2.0457, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 9.058835526735788e-06, |
|
"loss": 2.0627, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.046562514007275e-06, |
|
"loss": 2.0369, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 9.034218424750258e-06, |
|
"loss": 2.0828, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 9.02180347578457e-06, |
|
"loss": 2.0686, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 9.009317885174672e-06, |
|
"loss": 2.0532, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 8.99676187222582e-06, |
|
"loss": 2.0406, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 8.98413565748022e-06, |
|
"loss": 2.0875, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 8.971439462713145e-06, |
|
"loss": 2.0521, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.958673510929046e-06, |
|
"loss": 2.0569, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 8.945838026357633e-06, |
|
"loss": 2.0385, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 8.932933234449933e-06, |
|
"loss": 2.0561, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 8.91995936187434e-06, |
|
"loss": 2.0482, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 8.906916636512618e-06, |
|
"loss": 2.0214, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 8.89380528745592e-06, |
|
"loss": 2.0194, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 8.880625545000733e-06, |
|
"loss": 2.0503, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 8.86737764064487e-06, |
|
"loss": 2.0215, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.854061807083376e-06, |
|
"loss": 2.0278, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 8.84067827820445e-06, |
|
"loss": 2.0763, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 8.827227289085338e-06, |
|
"loss": 2.0128, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 8.813709075988205e-06, |
|
"loss": 2.0198, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 8.800123876355976e-06, |
|
"loss": 2.0678, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 8.786471928808182e-06, |
|
"loss": 1.9936, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 8.772753473136751e-06, |
|
"loss": 2.0451, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 8.758968750301808e-06, |
|
"loss": 1.9825, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 8.74511800242744e-06, |
|
"loss": 2.0519, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 8.731201472797436e-06, |
|
"loss": 2.0726, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 8.717219405851024e-06, |
|
"loss": 2.0264, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 8.703172047178576e-06, |
|
"loss": 2.0735, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 8.689059643517285e-06, |
|
"loss": 2.0678, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 8.674882442746844e-06, |
|
"loss": 2.0188, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 8.66064069388508e-06, |
|
"loss": 2.0203, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 8.646334647083587e-06, |
|
"loss": 2.0728, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 8.631964553623336e-06, |
|
"loss": 2.038, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 8.617530665910253e-06, |
|
"loss": 2.0447, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 8.603033237470783e-06, |
|
"loss": 2.046, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 8.588472522947456e-06, |
|
"loss": 2.0006, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 8.57384877809439e-06, |
|
"loss": 2.0695, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 8.559162259772811e-06, |
|
"loss": 2.0498, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.544413225946547e-06, |
|
"loss": 2.0221, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 8.529601935677481e-06, |
|
"loss": 1.9977, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 8.514728649121017e-06, |
|
"loss": 2.0469, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 8.499793627521501e-06, |
|
"loss": 2.0298, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 8.484797133207634e-06, |
|
"loss": 2.0354, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.469739429587865e-06, |
|
"loss": 2.0555, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 8.454620781145761e-06, |
|
"loss": 2.0557, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 8.439441453435371e-06, |
|
"loss": 2.0416, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.424201713076553e-06, |
|
"loss": 2.0333, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.408901827750289e-06, |
|
"loss": 2.0203, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 8.393542066193994e-06, |
|
"loss": 2.0092, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 8.37812269819678e-06, |
|
"loss": 2.0201, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 8.36264399459474e-06, |
|
"loss": 1.9957, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.347106227266161e-06, |
|
"loss": 2.02, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 8.331509669126778e-06, |
|
"loss": 2.0649, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 8.31585459412496e-06, |
|
"loss": 2.0384, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 8.30014127723691e-06, |
|
"loss": 2.0024, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 8.284369994461825e-06, |
|
"loss": 2.0536, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 8.268541022817058e-06, |
|
"loss": 2.0044, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 8.252654640333246e-06, |
|
"loss": 2.0584, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 8.236711126049426e-06, |
|
"loss": 2.0532, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 8.220710760008143e-06, |
|
"loss": 1.9918, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 8.204653823250516e-06, |
|
"loss": 2.0174, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.188540597811316e-06, |
|
"loss": 1.9979, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 8.172371366714004e-06, |
|
"loss": 2.0073, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 8.156146413965763e-06, |
|
"loss": 2.0156, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 8.1398660245525e-06, |
|
"loss": 1.985, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 8.123530484433867e-06, |
|
"loss": 2.0104, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 8.107140080538197e-06, |
|
"loss": 1.9925, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 8.090695100757505e-06, |
|
"loss": 2.0538, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 8.074195833942405e-06, |
|
"loss": 2.0415, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 8.057642569897049e-06, |
|
"loss": 2.0044, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 8.041035599374026e-06, |
|
"loss": 2.0371, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 8.024375214069269e-06, |
|
"loss": 2.017, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 8.007661706616919e-06, |
|
"loss": 2.0202, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 7.99089537058419e-06, |
|
"loss": 2.0547, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 7.974076500466215e-06, |
|
"loss": 2.005, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 7.957205391680872e-06, |
|
"loss": 2.012, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.940282340563586e-06, |
|
"loss": 2.0097, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 7.92330764436214e-06, |
|
"loss": 2.0385, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 7.906281601231449e-06, |
|
"loss": 2.0071, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 7.889204510228304e-06, |
|
"loss": 2.0146, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 7.87207667130615e-06, |
|
"loss": 1.9922, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 7.854898385309797e-06, |
|
"loss": 1.9942, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 7.83766995397014e-06, |
|
"loss": 2.0493, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 7.820391679898863e-06, |
|
"loss": 2.021, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 7.803063866583119e-06, |
|
"loss": 2.0249, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 7.7856868183802e-06, |
|
"loss": 1.9945, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.9995732307434082, |
|
"eval_runtime": 249.4314, |
|
"eval_samples_per_second": 9.197, |
|
"eval_steps_per_second": 4.598, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 7.7682608405122e-06, |
|
"loss": 2.0239, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 7.750786239060635e-06, |
|
"loss": 2.0123, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 7.733263320961087e-06, |
|
"loss": 1.977, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 7.715692393997802e-06, |
|
"loss": 2.0119, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 7.698073766798281e-06, |
|
"loss": 2.0071, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.680407748827869e-06, |
|
"loss": 1.9962, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 7.662694650384315e-06, |
|
"loss": 1.9844, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 7.644934782592315e-06, |
|
"loss": 2.0186, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.627128457398061e-06, |
|
"loss": 2.0253, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 7.609275987563745e-06, |
|
"loss": 2.0297, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 7.591377686662081e-06, |
|
"loss": 1.9942, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 7.573433869070788e-06, |
|
"loss": 1.9983, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 7.555444849967073e-06, |
|
"loss": 1.9873, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 7.537410945322085e-06, |
|
"loss": 2.045, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 7.519332471895384e-06, |
|
"loss": 2.0481, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.501209747229356e-06, |
|
"loss": 1.9862, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.483043089643654e-06, |
|
"loss": 2.0116, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 7.464832818229586e-06, |
|
"loss": 2.0211, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 7.446579252844536e-06, |
|
"loss": 1.9864, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 7.428282714106324e-06, |
|
"loss": 2.0303, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 7.409943523387586e-06, |
|
"loss": 1.9974, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 7.391562002810123e-06, |
|
"loss": 1.9797, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 7.37313847523925e-06, |
|
"loss": 1.9493, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 7.354673264278115e-06, |
|
"loss": 2.0172, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.336166694262028e-06, |
|
"loss": 2.0123, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 7.3176190902527526e-06, |
|
"loss": 1.9659, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 7.299030778032799e-06, |
|
"loss": 1.9849, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.280402084099711e-06, |
|
"loss": 1.9803, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 7.261733335660317e-06, |
|
"loss": 2.024, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 7.243024860624991e-06, |
|
"loss": 1.9692, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 7.224276987601895e-06, |
|
"loss": 2.0185, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 7.205490045891198e-06, |
|
"loss": 2.0265, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 7.186664365479301e-06, |
|
"loss": 2.0226, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 7.167800277033039e-06, |
|
"loss": 2.0106, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 7.148898111893867e-06, |
|
"loss": 2.0205, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.129958202072047e-06, |
|
"loss": 1.9522, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.110980880240814e-06, |
|
"loss": 1.9759, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 7.091966479730532e-06, |
|
"loss": 2.02, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 7.07291533452284e-06, |
|
"loss": 2.0119, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.053827779244784e-06, |
|
"loss": 2.0199, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 7.034704149162944e-06, |
|
"loss": 2.0191, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.015544780177539e-06, |
|
"loss": 1.9632, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 6.996350008816532e-06, |
|
"loss": 2.0271, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 6.977120172229713e-06, |
|
"loss": 1.9492, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.957855608182788e-06, |
|
"loss": 1.9785, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 6.938556655051432e-06, |
|
"loss": 1.991, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.9192236518153566e-06, |
|
"loss": 1.9875, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 6.8998569380523535e-06, |
|
"loss": 1.9848, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 6.880456853932327e-06, |
|
"loss": 2.0381, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 6.861023740211318e-06, |
|
"loss": 1.9968, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 6.841557938225527e-06, |
|
"loss": 1.9984, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.822059789885308e-06, |
|
"loss": 1.9702, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.802529637669171e-06, |
|
"loss": 1.9838, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 6.782967824617762e-06, |
|
"loss": 1.9962, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 6.76337469432784e-06, |
|
"loss": 2.0012, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 6.743750590946239e-06, |
|
"loss": 1.9384, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 6.72409585916383e-06, |
|
"loss": 1.9947, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 6.704410844209453e-06, |
|
"loss": 2.0145, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 6.684695891843871e-06, |
|
"loss": 2.0012, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 6.664951348353682e-06, |
|
"loss": 1.9855, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.645177560545245e-06, |
|
"loss": 2.0137, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 6.625374875738585e-06, |
|
"loss": 2.0052, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 6.605543641761293e-06, |
|
"loss": 1.9763, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 6.585684206942416e-06, |
|
"loss": 2.0089, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.56579692010634e-06, |
|
"loss": 1.9824, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 6.545882130566663e-06, |
|
"loss": 1.9818, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 6.525940188120059e-06, |
|
"loss": 2.0417, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 6.505971443040132e-06, |
|
"loss": 1.9811, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 6.485976246071268e-06, |
|
"loss": 2.0328, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 6.4659549484224704e-06, |
|
"loss": 1.9734, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 6.445907901761189e-06, |
|
"loss": 2.0472, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.425835458207155e-06, |
|
"loss": 2.0085, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.405737970326179e-06, |
|
"loss": 2.022, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 6.3856157911239714e-06, |
|
"loss": 2.0213, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.365469274039936e-06, |
|
"loss": 1.9837, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 6.345298772940962e-06, |
|
"loss": 1.999, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 6.325104642115214e-06, |
|
"loss": 1.9988, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.304887236265902e-06, |
|
"loss": 1.9888, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 6.2846469105050545e-06, |
|
"loss": 1.9944, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.264384020347281e-06, |
|
"loss": 1.975, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 6.244098921703524e-06, |
|
"loss": 2.0254, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 6.223791970874817e-06, |
|
"loss": 1.9662, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 6.203463524546017e-06, |
|
"loss": 1.9826, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 6.183113939779539e-06, |
|
"loss": 2.0226, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 6.162743574009094e-06, |
|
"loss": 1.9639, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.142352785033395e-06, |
|
"loss": 2.0038, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.121941931009894e-06, |
|
"loss": 1.9956, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.101511370448468e-06, |
|
"loss": 2.025, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 6.08106146220514e-06, |
|
"loss": 1.9768, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.060592565475765e-06, |
|
"loss": 2.0067, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 6.040105039789726e-06, |
|
"loss": 2.0106, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.019599245003616e-06, |
|
"loss": 1.9991, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 5.999075541294921e-06, |
|
"loss": 2.0292, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 5.978534289155689e-06, |
|
"loss": 2.0486, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 5.957975849386202e-06, |
|
"loss": 2.0203, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 5.937400583088636e-06, |
|
"loss": 2.0101, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 5.916808851660718e-06, |
|
"loss": 2.0079, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.896201016789385e-06, |
|
"loss": 1.9835, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 5.8755774404444175e-06, |
|
"loss": 1.9988, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 5.8549384848720965e-06, |
|
"loss": 2.0291, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 5.834284512588832e-06, |
|
"loss": 1.9633, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 5.813615886374795e-06, |
|
"loss": 1.9922, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 5.792932969267553e-06, |
|
"loss": 1.9726, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 5.772236124555684e-06, |
|
"loss": 2.0007, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 5.751525715772401e-06, |
|
"loss": 1.978, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 5.730802106689168e-06, |
|
"loss": 2.0138, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 5.7100656613093005e-06, |
|
"loss": 1.9792, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.689316743861589e-06, |
|
"loss": 2.0005, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 5.668555718793884e-06, |
|
"loss": 2.0156, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 5.647782950766709e-06, |
|
"loss": 1.9823, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.626998804646842e-06, |
|
"loss": 1.9878, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 5.606203645500916e-06, |
|
"loss": 2.0163, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 5.5853978385890054e-06, |
|
"loss": 2.0128, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 5.564581749358209e-06, |
|
"loss": 2.0038, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 5.543755743436231e-06, |
|
"loss": 2.0005, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 5.522920186624958e-06, |
|
"loss": 1.9923, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 5.502075444894035e-06, |
|
"loss": 2.0108, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 5.481221884374435e-06, |
|
"loss": 1.9527, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 5.4603598713520354e-06, |
|
"loss": 1.9828, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5.439489772261176e-06, |
|
"loss": 1.9947, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.418611953678224e-06, |
|
"loss": 2.0235, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 5.39772678231514e-06, |
|
"loss": 1.9974, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 5.376834625013031e-06, |
|
"loss": 2.012, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 5.355935848735712e-06, |
|
"loss": 1.9656, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 5.335030820563258e-06, |
|
"loss": 1.9831, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.5, |
|
"learning_rate": 5.3141199076855546e-06, |
|
"loss": 2.0297, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 5.293203477395851e-06, |
|
"loss": 1.9985, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 5.272281897084307e-06, |
|
"loss": 2.0078, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 5.251355534231546e-06, |
|
"loss": 1.9816, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 5.2304247564021886e-06, |
|
"loss": 2.0208, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 5.209489931238405e-06, |
|
"loss": 2.0368, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 5.188551426453458e-06, |
|
"loss": 1.9596, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 5.167609609825238e-06, |
|
"loss": 2.038, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.14666484918981e-06, |
|
"loss": 1.9991, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 5.125717512434947e-06, |
|
"loss": 2.0086, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 5.104767967493675e-06, |
|
"loss": 1.9864, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 5.083816582337799e-06, |
|
"loss": 1.9859, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 5.062863724971453e-06, |
|
"loss": 1.9661, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 5.041909763424625e-06, |
|
"loss": 2.0011, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 5.020955065746702e-06, |
|
"loss": 2.0104, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5e-06, |
|
"loss": 1.9524, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 4.979044934253299e-06, |
|
"loss": 1.9903, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 4.9580902365753765e-06, |
|
"loss": 1.9945, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 4.937136275028549e-06, |
|
"loss": 1.9889, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 4.916183417662203e-06, |
|
"loss": 1.9972, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 4.895232032506326e-06, |
|
"loss": 2.0227, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 4.874282487565053e-06, |
|
"loss": 2.0109, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 4.853335150810191e-06, |
|
"loss": 1.9937, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 4.832390390174763e-06, |
|
"loss": 2.0196, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 4.811448573546543e-06, |
|
"loss": 2.0225, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 4.790510068761596e-06, |
|
"loss": 2.0174, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 4.769575243597815e-06, |
|
"loss": 2.0172, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 4.7486444657684575e-06, |
|
"loss": 2.0014, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 4.727718102915694e-06, |
|
"loss": 1.9513, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 4.706796522604152e-06, |
|
"loss": 1.987, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.5, |
|
"learning_rate": 4.685880092314448e-06, |
|
"loss": 1.9774, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 4.664969179436744e-06, |
|
"loss": 1.9865, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 4.644064151264289e-06, |
|
"loss": 1.9413, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 4.623165374986971e-06, |
|
"loss": 2.0014, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 4.602273217684861e-06, |
|
"loss": 2.0078, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 4.581388046321777e-06, |
|
"loss": 1.9833, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 4.560510227738825e-06, |
|
"loss": 1.9535, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 4.539640128647965e-06, |
|
"loss": 2.0031, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.518778115625566e-06, |
|
"loss": 2.023, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 4.497924555105966e-06, |
|
"loss": 1.9782, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 4.477079813375043e-06, |
|
"loss": 1.9828, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 4.456244256563769e-06, |
|
"loss": 1.9677, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 4.435418250641791e-06, |
|
"loss": 1.9976, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 4.414602161410995e-06, |
|
"loss": 2.0058, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 4.393796354499086e-06, |
|
"loss": 1.9918, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 4.373001195353159e-06, |
|
"loss": 1.9882, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 4.352217049233294e-06, |
|
"loss": 1.9824, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 4.331444281206118e-06, |
|
"loss": 2.0098, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 4.310683256138414e-06, |
|
"loss": 2.0122, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 4.2899343386907e-06, |
|
"loss": 2.0138, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 4.269197893310834e-06, |
|
"loss": 1.9788, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 4.2484742842276e-06, |
|
"loss": 2.0198, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 4.227763875444317e-06, |
|
"loss": 1.9988, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 4.207067030732449e-06, |
|
"loss": 1.983, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 4.186384113625207e-06, |
|
"loss": 1.9983, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 4.165715487411169e-06, |
|
"loss": 1.9702, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 4.145061515127905e-06, |
|
"loss": 1.9954, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 4.124422559555584e-06, |
|
"loss": 2.0175, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.5, |
|
"learning_rate": 4.103798983210617e-06, |
|
"loss": 1.9919, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 4.083191148339283e-06, |
|
"loss": 1.9898, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 4.0625994169113645e-06, |
|
"loss": 2.017, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 4.042024150613798e-06, |
|
"loss": 1.9761, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 4.021465710844311e-06, |
|
"loss": 1.972, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 4.000924458705079e-06, |
|
"loss": 1.9848, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 3.980400754996384e-06, |
|
"loss": 2.0117, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 3.959894960210275e-06, |
|
"loss": 2.0132, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 3.939407434524238e-06, |
|
"loss": 2.0503, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 3.918938537794863e-06, |
|
"loss": 2.0558, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 3.898488629551535e-06, |
|
"loss": 1.9943, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 3.87805806899011e-06, |
|
"loss": 1.9821, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 3.857647214966606e-06, |
|
"loss": 1.9503, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 3.837256425990909e-06, |
|
"loss": 1.9906, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 3.816886060220462e-06, |
|
"loss": 1.9814, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 3.7965364754539845e-06, |
|
"loss": 2.0166, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 3.7762080291251836e-06, |
|
"loss": 1.9964, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 3.755901078296478e-06, |
|
"loss": 2.0175, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 3.7356159796527214e-06, |
|
"loss": 2.0271, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 3.7153530894949476e-06, |
|
"loss": 2.0338, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 3.6951127637340995e-06, |
|
"loss": 1.9582, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 3.6748953578847875e-06, |
|
"loss": 1.9524, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 3.6547012270590397e-06, |
|
"loss": 1.9826, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 3.6345307259600657e-06, |
|
"loss": 1.9812, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 3.6143842088760294e-06, |
|
"loss": 1.9708, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 3.5942620296738218e-06, |
|
"loss": 2.01, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 3.5741645417928457e-06, |
|
"loss": 1.9714, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 3.554092098238811e-06, |
|
"loss": 1.9949, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 3.5340450515775316e-06, |
|
"loss": 1.9974, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 3.514023753928734e-06, |
|
"loss": 2.0229, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 3.49402855695987e-06, |
|
"loss": 1.9732, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 3.474059811879944e-06, |
|
"loss": 2.0332, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 3.45411786943334e-06, |
|
"loss": 1.9856, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 3.434203079893662e-06, |
|
"loss": 1.9775, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 3.414315793057586e-06, |
|
"loss": 1.9561, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 3.3944563582387084e-06, |
|
"loss": 1.9733, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 3.374625124261416e-06, |
|
"loss": 2.0123, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 3.3548224394547557e-06, |
|
"loss": 2.0205, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 3.3350486516463194e-06, |
|
"loss": 1.9915, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 3.31530410815613e-06, |
|
"loss": 2.0599, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 3.2955891557905477e-06, |
|
"loss": 1.9845, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 3.275904140836172e-06, |
|
"loss": 1.9529, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 3.256249409053761e-06, |
|
"loss": 1.9569, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 3.2366253056721607e-06, |
|
"loss": 2.0147, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.5, |
|
"learning_rate": 3.2170321753822374e-06, |
|
"loss": 2.0061, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 3.1974703623308288e-06, |
|
"loss": 2.0166, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 3.1779402101146917e-06, |
|
"loss": 2.0089, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 3.1584420617744737e-06, |
|
"loss": 2.0264, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 3.138976259788682e-06, |
|
"loss": 1.9822, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 3.119543146067675e-06, |
|
"loss": 1.9841, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 3.100143061947648e-06, |
|
"loss": 2.0002, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 3.0807763481846455e-06, |
|
"loss": 1.9902, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 3.0614433449485716e-06, |
|
"loss": 1.967, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 3.0421443918172155e-06, |
|
"loss": 1.8929, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 3.022879827770288e-06, |
|
"loss": 2.0185, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 3.00364999118347e-06, |
|
"loss": 2.0332, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 2.984455219822462e-06, |
|
"loss": 1.9836, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 2.965295850837057e-06, |
|
"loss": 1.9414, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.946172220755217e-06, |
|
"loss": 1.9639, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 2.927084665477162e-06, |
|
"loss": 1.9881, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 2.90803352026947e-06, |
|
"loss": 2.0282, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 2.889019119759187e-06, |
|
"loss": 1.9554, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 2.8700417979279555e-06, |
|
"loss": 2.0012, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.8511018881061347e-06, |
|
"loss": 1.9745, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 2.8321997229669616e-06, |
|
"loss": 2.0432, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.5, |
|
"learning_rate": 2.8133356345206996e-06, |
|
"loss": 1.9733, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 2.7945099541088026e-06, |
|
"loss": 1.9746, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 2.775723012398107e-06, |
|
"loss": 2.0081, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 2.7569751393750088e-06, |
|
"loss": 2.0195, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 2.7382666643396825e-06, |
|
"loss": 1.9804, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 2.7195979159002918e-06, |
|
"loss": 1.971, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 2.7009692219672025e-06, |
|
"loss": 2.0013, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 2.6823809097472495e-06, |
|
"loss": 1.9895, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 2.6638333057379743e-06, |
|
"loss": 2.0207, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 2.645326735721886e-06, |
|
"loss": 2.0215, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.9850621223449707, |
|
"eval_runtime": 247.3914, |
|
"eval_samples_per_second": 9.273, |
|
"eval_steps_per_second": 4.636, |
|
"step": 2577 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 2.6268615247607533e-06, |
|
"loss": 1.9937, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 2.608437997189878e-06, |
|
"loss": 2.0004, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 2.5900564766124163e-06, |
|
"loss": 1.9831, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 2.5717172858936767e-06, |
|
"loss": 1.9916, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.5534207471554644e-06, |
|
"loss": 1.9886, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 2.5351671817704156e-06, |
|
"loss": 2.0097, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 2.5169569103563484e-06, |
|
"loss": 2.0168, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 2.498790252770645e-06, |
|
"loss": 1.9968, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 0.5, |
|
"learning_rate": 2.480667528104617e-06, |
|
"loss": 1.9686, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 2.462589054677914e-06, |
|
"loss": 2.0257, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 2.444555150032929e-06, |
|
"loss": 1.9721, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 2.4265661309292116e-06, |
|
"loss": 1.9975, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 2.40862231333792e-06, |
|
"loss": 2.0092, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 2.390724012436255e-06, |
|
"loss": 2.0166, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 2.372871542601939e-06, |
|
"loss": 2.0051, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 2.3550652174076867e-06, |
|
"loss": 1.9946, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 2.3373053496156865e-06, |
|
"loss": 1.9957, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 2.319592251172133e-06, |
|
"loss": 1.9535, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 2.301926233201721e-06, |
|
"loss": 1.9934, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 2.2843076060022003e-06, |
|
"loss": 2.0027, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 2.2667366790389152e-06, |
|
"loss": 1.9822, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 2.2492137609393663e-06, |
|
"loss": 1.9879, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 2.231739159487803e-06, |
|
"loss": 1.9868, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 2.2143131816198007e-06, |
|
"loss": 2.0242, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 2.196936133416882e-06, |
|
"loss": 1.9868, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 2.1796083201011387e-06, |
|
"loss": 1.9965, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 2.16233004602986e-06, |
|
"loss": 1.9783, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 2.145101614690205e-06, |
|
"loss": 1.9721, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 2.1279233286938503e-06, |
|
"loss": 2.0304, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 2.1107954897716976e-06, |
|
"loss": 1.964, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 2.093718398768553e-06, |
|
"loss": 1.9907, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 2.0766923556378576e-06, |
|
"loss": 2.0063, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.059717659436415e-06, |
|
"loss": 1.9807, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 2.0427946083191295e-06, |
|
"loss": 1.9811, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 2.0259234995337864e-06, |
|
"loss": 1.9759, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.009104629415813e-06, |
|
"loss": 2.0222, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 1.9923382933830836e-06, |
|
"loss": 2.0119, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 1.9756247859307336e-06, |
|
"loss": 1.9132, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 1.9589644006259754e-06, |
|
"loss": 1.9956, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 1.9423574301029524e-06, |
|
"loss": 1.9716, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.5, |
|
"learning_rate": 1.925804166057596e-06, |
|
"loss": 2.0107, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 1.9093048992424964e-06, |
|
"loss": 1.9849, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 1.8928599194618052e-06, |
|
"loss": 2.0232, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 1.876469515566136e-06, |
|
"loss": 1.976, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.8601339754475007e-06, |
|
"loss": 2.0444, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.8438535860342398e-06, |
|
"loss": 1.9898, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 1.827628633285996e-06, |
|
"loss": 1.9559, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 1.8114594021886845e-06, |
|
"loss": 1.9971, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 1.795346176749484e-06, |
|
"loss": 1.9682, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 1.7792892399918587e-06, |
|
"loss": 2.0249, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 1.7632888739505744e-06, |
|
"loss": 1.9901, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.7473453596667544e-06, |
|
"loss": 1.9835, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 1.7314589771829426e-06, |
|
"loss": 1.9695, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 1.715630005538174e-06, |
|
"loss": 1.9969, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.6998587227630903e-06, |
|
"loss": 2.0026, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 1.684145405875039e-06, |
|
"loss": 2.0031, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 1.668490330873223e-06, |
|
"loss": 2.0011, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 1.6528937727338411e-06, |
|
"loss": 1.9748, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 1.6373560054052629e-06, |
|
"loss": 1.9881, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 1.6218773018032212e-06, |
|
"loss": 1.9932, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 1.6064579338060088e-06, |
|
"loss": 2.0219, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 1.5910981722497115e-06, |
|
"loss": 2.0025, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 1.575798286923449e-06, |
|
"loss": 2.0118, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 1.5605585465646295e-06, |
|
"loss": 1.9888, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.545379218854241e-06, |
|
"loss": 1.9833, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 1.5302605704121375e-06, |
|
"loss": 2.0166, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 1.5152028667923669e-06, |
|
"loss": 2.0144, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 1.5002063724785004e-06, |
|
"loss": 1.9796, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 1.4852713508789835e-06, |
|
"loss": 2.0341, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 1.4703980643225208e-06, |
|
"loss": 2.0127, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.4555867740534552e-06, |
|
"loss": 1.9815, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 1.4408377402271907e-06, |
|
"loss": 2.0077, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 0.5, |
|
"learning_rate": 1.4261512219056118e-06, |
|
"loss": 2.008, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 1.411527477052544e-06, |
|
"loss": 1.9747, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.3969667625292165e-06, |
|
"loss": 1.9639, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 1.3824693340897483e-06, |
|
"loss": 1.9885, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 1.3680354463766642e-06, |
|
"loss": 1.9922, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 1.3536653529164146e-06, |
|
"loss": 2.0102, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 1.3393593061149223e-06, |
|
"loss": 1.998, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 1.3251175572531583e-06, |
|
"loss": 1.9725, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 1.3109403564827155e-06, |
|
"loss": 1.9735, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 1.2968279528214246e-06, |
|
"loss": 1.9715, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 1.2827805941489767e-06, |
|
"loss": 1.9592, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.5, |
|
"learning_rate": 1.2687985272025655e-06, |
|
"loss": 1.9479, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 1.2548819975725624e-06, |
|
"loss": 2.0103, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 1.2410312496981923e-06, |
|
"loss": 1.972, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 1.2272465268632512e-06, |
|
"loss": 1.9831, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 1.2135280711918197e-06, |
|
"loss": 1.9965, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 1.1998761236440248e-06, |
|
"loss": 2.0397, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 1.1862909240117982e-06, |
|
"loss": 1.9739, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 1.1727727109146632e-06, |
|
"loss": 1.9653, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 1.1593217217955522e-06, |
|
"loss": 2.0289, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 1.1459381929166251e-06, |
|
"loss": 1.9883, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.1326223593551294e-06, |
|
"loss": 1.9846, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 1.1193744549992675e-06, |
|
"loss": 2.0217, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 1.1061947125440818e-06, |
|
"loss": 2.0168, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 1.0930833634873811e-06, |
|
"loss": 1.981, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 1.0800406381256616e-06, |
|
"loss": 1.9489, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 1.0670667655500665e-06, |
|
"loss": 1.9975, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 1.0541619736423686e-06, |
|
"loss": 1.969, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 1.0413264890709546e-06, |
|
"loss": 2.017, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 1.0285605372868567e-06, |
|
"loss": 2.0058, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 1.0158643425197817e-06, |
|
"loss": 2.0093, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 1.003238127774181e-06, |
|
"loss": 1.9601, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 9.906821148253303e-07, |
|
"loss": 1.997, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 9.781965242154307e-07, |
|
"loss": 2.0081, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.65781575249743e-07, |
|
"loss": 1.9958, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 9.534374859927248e-07, |
|
"loss": 1.9913, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 9.411644732642122e-07, |
|
"loss": 2.0226, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 9.289627526356048e-07, |
|
"loss": 1.9847, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 9.168325384260751e-07, |
|
"loss": 1.9623, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 9.047740436988168e-07, |
|
"loss": 1.9876, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 8.927874802572861e-07, |
|
"loss": 1.9982, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 8.808730586414982e-07, |
|
"loss": 1.9776, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 8.690309881243147e-07, |
|
"loss": 2.0388, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 8.57261476707778e-07, |
|
"loss": 2.0119, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.455647311194537e-07, |
|
"loss": 2.0039, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 8.339409568087981e-07, |
|
"loss": 1.9728, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 8.223903579435499e-07, |
|
"loss": 1.9684, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 8.109131374061508e-07, |
|
"loss": 1.9851, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.995094967901701e-07, |
|
"loss": 1.9839, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.881796363967753e-07, |
|
"loss": 2.0393, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 7.769237552312048e-07, |
|
"loss": 1.9957, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.657420509992785e-07, |
|
"loss": 1.9819, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 7.546347201039255e-07, |
|
"loss": 2.0461, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 7.436019576417258e-07, |
|
"loss": 2.0057, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 7.326439573994953e-07, |
|
"loss": 1.9784, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 7.217609118508722e-07, |
|
"loss": 1.9894, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 7.109530121529439e-07, |
|
"loss": 2.0161, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.002204481428826e-07, |
|
"loss": 1.9951, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 6.895634083346159e-07, |
|
"loss": 1.9595, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 6.789820799155155e-07, |
|
"loss": 1.9779, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 6.684766487431027e-07, |
|
"loss": 1.9571, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.580472993417952e-07, |
|
"loss": 2.0053, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.476942148996529e-07, |
|
"loss": 1.9901, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.374175772651725e-07, |
|
"loss": 2.011, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 6.272175669440861e-07, |
|
"loss": 1.9691, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.170943630961918e-07, |
|
"loss": 1.9929, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 6.070481435322062e-07, |
|
"loss": 2.0098, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 5.970790847106461e-07, |
|
"loss": 1.9893, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 5.871873617347218e-07, |
|
"loss": 2.0239, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.773731483492684e-07, |
|
"loss": 1.9829, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 5.676366169376873e-07, |
|
"loss": 2.0305, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 5.579779385189261e-07, |
|
"loss": 2.0101, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 5.483972827444645e-07, |
|
"loss": 2.0039, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 5.388948178953462e-07, |
|
"loss": 1.9907, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 5.294707108792146e-07, |
|
"loss": 1.9854, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.201251272273833e-07, |
|
"loss": 1.9777, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 5.108582310919302e-07, |
|
"loss": 1.9771, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 5.016701852428113e-07, |
|
"loss": 1.9773, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 4.925611510650058e-07, |
|
"loss": 2.0213, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 4.835312885556781e-07, |
|
"loss": 2.045, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 4.745807563213678e-07, |
|
"loss": 2.0107, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 4.6570971157520574e-07, |
|
"loss": 1.9895, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 4.569183101341501e-07, |
|
"loss": 2.0058, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 4.482067064162521e-07, |
|
"loss": 1.9324, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 4.3957505343794115e-07, |
|
"loss": 1.9871, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 4.310235028113402e-07, |
|
"loss": 2.0171, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 4.225522047415992e-07, |
|
"loss": 1.978, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 4.141613080242579e-07, |
|
"loss": 2.018, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 4.058509600426358e-07, |
|
"loss": 1.9975, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.5, |
|
"learning_rate": 3.976213067652396e-07, |
|
"loss": 2.0103, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 3.894724927431981e-07, |
|
"loss": 1.9982, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 3.8140466110772875e-07, |
|
"loss": 2.0159, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 3.734179535676169e-07, |
|
"loss": 1.9809, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 3.655125104067325e-07, |
|
"loss": 1.997, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 3.576884704815631e-07, |
|
"loss": 1.9811, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 3.499459712187742e-07, |
|
"loss": 2.0204, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 3.422851486127987e-07, |
|
"loss": 1.9684, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 3.3470613722344335e-07, |
|
"loss": 2.0207, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 3.272090701735314e-07, |
|
"loss": 1.9406, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 3.1979407914655756e-07, |
|
"loss": 2.0232, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 3.1246129438438076e-07, |
|
"loss": 1.9855, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 3.0521084468493446e-07, |
|
"loss": 2.0071, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 2.980428573999605e-07, |
|
"loss": 1.9886, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 2.909574584327812e-07, |
|
"loss": 2.0096, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 2.839547722360769e-07, |
|
"loss": 1.9977, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 2.770349218097096e-07, |
|
"loss": 1.956, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 2.7019802869855783e-07, |
|
"loss": 2.0202, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 2.6344421299037915e-07, |
|
"loss": 2.0191, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 2.5677359331370834e-07, |
|
"loss": 2.0047, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 2.5018628683576796e-07, |
|
"loss": 1.9828, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 0.5, |
|
"learning_rate": 2.436824092604101e-07, |
|
"loss": 2.0087, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 2.3726207482608842e-07, |
|
"loss": 1.9827, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 2.309253963038477e-07, |
|
"loss": 1.9635, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 0.5, |
|
"learning_rate": 2.246724849953452e-07, |
|
"loss": 2.0159, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 2.1850345073089253e-07, |
|
"loss": 2.009, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 2.1241840186753238e-07, |
|
"loss": 1.9898, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.0641744528712925e-07, |
|
"loss": 2.0171, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 2.0050068639449472e-07, |
|
"loss": 2.006, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 1.9466822911553772e-07, |
|
"loss": 2.0039, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 1.889201758954351e-07, |
|
"loss": 2.0155, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 1.832566276968345e-07, |
|
"loss": 2.0047, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 1.7767768399808372e-07, |
|
"loss": 2.0088, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 1.7218344279147702e-07, |
|
"loss": 1.978, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 1.6677400058153993e-07, |
|
"loss": 1.9719, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 1.6144945238332987e-07, |
|
"loss": 1.9975, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 1.5620989172077106e-07, |
|
"loss": 1.9942, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 1.5105541062500838e-07, |
|
"loss": 1.9904, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 1.4598609963279164e-07, |
|
"loss": 2.0009, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 1.4100204778488947e-07, |
|
"loss": 2.0399, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 1.361033426245173e-07, |
|
"loss": 1.9531, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 1.3129007019580752e-07, |
|
"loss": 1.9986, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.265623150422951e-07, |
|
"loss": 1.9936, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 1.2192016020542986e-07, |
|
"loss": 2.03, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 1.1736368722312263e-07, |
|
"loss": 1.9686, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 1.1289297612830984e-07, |
|
"loss": 2.0041, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.0850810544754897e-07, |
|
"loss": 2.0213, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 1.0420915219964023e-07, |
|
"loss": 1.9643, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 9.999619189427102e-08, |
|
"loss": 2.0194, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 9.586929853069139e-08, |
|
"loss": 2.0342, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 9.182854459641565e-08, |
|
"loss": 1.9825, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 8.787400106594568e-08, |
|
"loss": 1.9844, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 8.400573739952677e-08, |
|
"loss": 2.0232, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 8.02238215419282e-08, |
|
"loss": 2.0074, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.652831992124799e-08, |
|
"loss": 1.9956, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.291929744774495e-08, |
|
"loss": 1.982, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 6.939681751270356e-08, |
|
"loss": 1.9714, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 6.596094198731473e-08, |
|
"loss": 1.9725, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 6.261173122159292e-08, |
|
"loss": 1.9973, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 5.934924404331355e-08, |
|
"loss": 1.9777, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 5.6173537756982756e-08, |
|
"loss": 1.9547, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 5.30846681428282e-08, |
|
"loss": 1.9781, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 5.00826894558204e-08, |
|
"loss": 1.9667, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 0.5, |
|
"learning_rate": 4.716765442471849e-08, |
|
"loss": 2.0228, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 4.433961425114708e-08, |
|
"loss": 2.0139, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 4.1598618608693074e-08, |
|
"loss": 2.0082, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 3.894471564203639e-08, |
|
"loss": 1.9803, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 3.637795196610228e-08, |
|
"loss": 1.9956, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 3.3898372665243096e-08, |
|
"loss": 2.0058, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 3.1506021292447285e-08, |
|
"loss": 1.9566, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 2.9200939868573308e-08, |
|
"loss": 2.0039, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 2.6983168881611897e-08, |
|
"loss": 1.9406, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 2.485274728597442e-08, |
|
"loss": 2.0118, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 2.2809712501810632e-08, |
|
"loss": 2.0342, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 2.08541004143481e-08, |
|
"loss": 1.983, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 1.898594537326437e-08, |
|
"loss": 2.0319, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 1.7205280192084117e-08, |
|
"loss": 1.9244, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 1.5512136147600167e-08, |
|
"loss": 1.9726, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 1.3906542979326143e-08, |
|
"loss": 2.0218, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 1.2388528888973017e-08, |
|
"loss": 2.0047, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.0958120539955596e-08, |
|
"loss": 1.977, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 9.615343056922355e-09, |
|
"loss": 1.9978, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 8.360220025313005e-09, |
|
"loss": 1.9994, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 7.1927734909488235e-09, |
|
"loss": 2.0006, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 6.113023959640751e-09, |
|
"loss": 1.9753, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.120990396830228e-09, |
|
"loss": 1.9885, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 4.216690227258902e-09, |
|
"loss": 2.0332, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 3.400139334658881e-09, |
|
"loss": 2.0213, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 2.6713520614757295e-09, |
|
"loss": 1.9891, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 2.0303412086164487e-09, |
|
"loss": 2.0187, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.477118035223546e-09, |
|
"loss": 1.9992, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 1.01169225847908e-09, |
|
"loss": 1.9885, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 6.340720534325772e-10, |
|
"loss": 1.9889, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 3.4426405285725715e-10, |
|
"loss": 1.9909, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.4227334713456943e-10, |
|
"loss": 1.9585, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 2.8103484164820894e-11, |
|
"loss": 1.9766, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.9850057363510132, |
|
"eval_runtime": 246.6106, |
|
"eval_samples_per_second": 9.302, |
|
"eval_steps_per_second": 4.651, |
|
"step": 3864 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 3864, |
|
"total_flos": 8.050802347506401e+17, |
|
"train_loss": 2.091761199090298, |
|
"train_runtime": 21448.6469, |
|
"train_samples_per_second": 2.883, |
|
"train_steps_per_second": 0.18 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 3864, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 1000, |
|
"total_flos": 8.050802347506401e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|