diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6322 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 15.0, + "eval_steps": 500, + "global_step": 4395, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0034129692832764505, + "grad_norm": 1.8359375, + "learning_rate": 4.545454545454545e-07, + "loss": 3.0499, + "step": 1 + }, + { + "epoch": 0.017064846416382253, + "grad_norm": 2.234375, + "learning_rate": 2.2727272727272728e-06, + "loss": 3.0434, + "step": 5 + }, + { + "epoch": 0.034129692832764506, + "grad_norm": 2.078125, + "learning_rate": 4.5454545454545455e-06, + "loss": 3.0699, + "step": 10 + }, + { + "epoch": 0.051194539249146756, + "grad_norm": 1.8515625, + "learning_rate": 6.818181818181818e-06, + "loss": 3.0656, + "step": 15 + }, + { + "epoch": 0.06825938566552901, + "grad_norm": 3.0, + "learning_rate": 9.090909090909091e-06, + "loss": 3.0526, + "step": 20 + }, + { + "epoch": 0.08532423208191127, + "grad_norm": 2.015625, + "learning_rate": 1.1363636363636365e-05, + "loss": 3.0382, + "step": 25 + }, + { + "epoch": 0.10238907849829351, + "grad_norm": 17.125, + "learning_rate": 1.3636363636363637e-05, + "loss": 2.982, + "step": 30 + }, + { + "epoch": 0.11945392491467577, + "grad_norm": 2.03125, + "learning_rate": 1.590909090909091e-05, + "loss": 2.9332, + "step": 35 + }, + { + "epoch": 0.13651877133105803, + "grad_norm": 3.140625, + "learning_rate": 1.8181818181818182e-05, + "loss": 2.8934, + "step": 40 + }, + { + "epoch": 0.15358361774744028, + "grad_norm": 3.25, + "learning_rate": 2.0454545454545457e-05, + "loss": 2.7804, + "step": 45 + }, + { + "epoch": 0.17064846416382254, + "grad_norm": 1.3984375, + "learning_rate": 2.272727272727273e-05, + "loss": 2.7194, + "step": 50 + }, + { + "epoch": 0.18771331058020477, + "grad_norm": 1.296875, + "learning_rate": 2.5e-05, + "loss": 2.5961, + "step": 55 + }, + { + "epoch": 0.20477815699658702, + "grad_norm": 1.6640625, + "learning_rate": 2.7272727272727273e-05, + "loss": 2.5046, + "step": 60 + }, + { + "epoch": 0.22184300341296928, + "grad_norm": 1.0234375, + "learning_rate": 2.954545454545455e-05, + "loss": 2.3975, + "step": 65 + }, + { + "epoch": 0.23890784982935154, + "grad_norm": 1.390625, + "learning_rate": 3.181818181818182e-05, + "loss": 2.3091, + "step": 70 + }, + { + "epoch": 0.25597269624573377, + "grad_norm": 1.1171875, + "learning_rate": 3.409090909090909e-05, + "loss": 2.2036, + "step": 75 + }, + { + "epoch": 0.27303754266211605, + "grad_norm": 0.84375, + "learning_rate": 3.6363636363636364e-05, + "loss": 2.113, + "step": 80 + }, + { + "epoch": 0.2901023890784983, + "grad_norm": 1.7421875, + "learning_rate": 3.8636363636363636e-05, + "loss": 2.043, + "step": 85 + }, + { + "epoch": 0.30716723549488056, + "grad_norm": 6.375, + "learning_rate": 4.0909090909090915e-05, + "loss": 1.9568, + "step": 90 + }, + { + "epoch": 0.3242320819112628, + "grad_norm": 0.71484375, + "learning_rate": 4.318181818181819e-05, + "loss": 1.8927, + "step": 95 + }, + { + "epoch": 0.3412969283276451, + "grad_norm": 1.5078125, + "learning_rate": 4.545454545454546e-05, + "loss": 1.8394, + "step": 100 + }, + { + "epoch": 0.3583617747440273, + "grad_norm": 25.625, + "learning_rate": 4.772727272727273e-05, + "loss": 1.7808, + "step": 105 + }, + { + "epoch": 0.37542662116040953, + "grad_norm": 0.54296875, + "learning_rate": 5e-05, + "loss": 1.7467, + "step": 110 + }, + { + "epoch": 0.3924914675767918, + "grad_norm": 0.8671875, + "learning_rate": 5.2272727272727274e-05, + "loss": 1.6988, + "step": 115 + }, + { + "epoch": 0.40955631399317405, + "grad_norm": 0.609375, + "learning_rate": 5.4545454545454546e-05, + "loss": 1.6442, + "step": 120 + }, + { + "epoch": 0.42662116040955633, + "grad_norm": 0.5078125, + "learning_rate": 5.6818181818181825e-05, + "loss": 1.5875, + "step": 125 + }, + { + "epoch": 0.44368600682593856, + "grad_norm": 0.890625, + "learning_rate": 5.90909090909091e-05, + "loss": 1.5646, + "step": 130 + }, + { + "epoch": 0.46075085324232085, + "grad_norm": 0.53125, + "learning_rate": 6.136363636363636e-05, + "loss": 1.5244, + "step": 135 + }, + { + "epoch": 0.4778156996587031, + "grad_norm": 0.39453125, + "learning_rate": 6.363636363636364e-05, + "loss": 1.4945, + "step": 140 + }, + { + "epoch": 0.4948805460750853, + "grad_norm": 0.447265625, + "learning_rate": 6.59090909090909e-05, + "loss": 1.469, + "step": 145 + }, + { + "epoch": 0.5119453924914675, + "grad_norm": 0.75390625, + "learning_rate": 6.818181818181818e-05, + "loss": 1.4478, + "step": 150 + }, + { + "epoch": 0.5290102389078498, + "grad_norm": 0.392578125, + "learning_rate": 7.045454545454546e-05, + "loss": 1.4291, + "step": 155 + }, + { + "epoch": 0.5460750853242321, + "grad_norm": 0.314453125, + "learning_rate": 7.272727272727273e-05, + "loss": 1.4056, + "step": 160 + }, + { + "epoch": 0.5631399317406144, + "grad_norm": 0.326171875, + "learning_rate": 7.500000000000001e-05, + "loss": 1.3839, + "step": 165 + }, + { + "epoch": 0.5802047781569966, + "grad_norm": 0.283203125, + "learning_rate": 7.727272727272727e-05, + "loss": 1.3664, + "step": 170 + }, + { + "epoch": 0.5972696245733788, + "grad_norm": 0.39453125, + "learning_rate": 7.954545454545455e-05, + "loss": 1.3557, + "step": 175 + }, + { + "epoch": 0.6143344709897611, + "grad_norm": 0.33984375, + "learning_rate": 8.181818181818183e-05, + "loss": 1.3317, + "step": 180 + }, + { + "epoch": 0.6313993174061433, + "grad_norm": 0.3125, + "learning_rate": 8.40909090909091e-05, + "loss": 1.3323, + "step": 185 + }, + { + "epoch": 0.6484641638225256, + "grad_norm": 0.384765625, + "learning_rate": 8.636363636363637e-05, + "loss": 1.3129, + "step": 190 + }, + { + "epoch": 0.6655290102389079, + "grad_norm": 0.435546875, + "learning_rate": 8.863636363636364e-05, + "loss": 1.3194, + "step": 195 + }, + { + "epoch": 0.6825938566552902, + "grad_norm": 0.4140625, + "learning_rate": 9.090909090909092e-05, + "loss": 1.2992, + "step": 200 + }, + { + "epoch": 0.6996587030716723, + "grad_norm": 0.296875, + "learning_rate": 9.318181818181818e-05, + "loss": 1.2934, + "step": 205 + }, + { + "epoch": 0.7167235494880546, + "grad_norm": 0.515625, + "learning_rate": 9.545454545454546e-05, + "loss": 1.2759, + "step": 210 + }, + { + "epoch": 0.7337883959044369, + "grad_norm": 0.52734375, + "learning_rate": 9.772727272727274e-05, + "loss": 1.2775, + "step": 215 + }, + { + "epoch": 0.7508532423208191, + "grad_norm": 0.357421875, + "learning_rate": 0.0001, + "loss": 1.2696, + "step": 220 + }, + { + "epoch": 0.7679180887372014, + "grad_norm": 0.29296875, + "learning_rate": 0.00010227272727272727, + "loss": 1.2621, + "step": 225 + }, + { + "epoch": 0.7849829351535836, + "grad_norm": 0.3359375, + "learning_rate": 0.00010454545454545455, + "loss": 1.251, + "step": 230 + }, + { + "epoch": 0.8020477815699659, + "grad_norm": 0.419921875, + "learning_rate": 0.00010681818181818181, + "loss": 1.2544, + "step": 235 + }, + { + "epoch": 0.8191126279863481, + "grad_norm": 0.48046875, + "learning_rate": 0.00010909090909090909, + "loss": 1.2528, + "step": 240 + }, + { + "epoch": 0.8361774744027304, + "grad_norm": 0.5234375, + "learning_rate": 0.00011136363636363636, + "loss": 1.2459, + "step": 245 + }, + { + "epoch": 0.8532423208191127, + "grad_norm": 0.455078125, + "learning_rate": 0.00011363636363636365, + "loss": 1.2322, + "step": 250 + }, + { + "epoch": 0.8703071672354948, + "grad_norm": 0.451171875, + "learning_rate": 0.00011590909090909093, + "loss": 1.2154, + "step": 255 + }, + { + "epoch": 0.8873720136518771, + "grad_norm": 0.44140625, + "learning_rate": 0.0001181818181818182, + "loss": 1.2258, + "step": 260 + }, + { + "epoch": 0.9044368600682594, + "grad_norm": 0.56640625, + "learning_rate": 0.00012045454545454546, + "loss": 1.213, + "step": 265 + }, + { + "epoch": 0.9215017064846417, + "grad_norm": 0.46875, + "learning_rate": 0.00012272727272727272, + "loss": 1.224, + "step": 270 + }, + { + "epoch": 0.9385665529010239, + "grad_norm": 0.51171875, + "learning_rate": 0.000125, + "loss": 1.2093, + "step": 275 + }, + { + "epoch": 0.9556313993174061, + "grad_norm": 0.90234375, + "learning_rate": 0.00012727272727272728, + "loss": 1.2132, + "step": 280 + }, + { + "epoch": 0.9726962457337884, + "grad_norm": 0.63671875, + "learning_rate": 0.00012954545454545456, + "loss": 1.2083, + "step": 285 + }, + { + "epoch": 0.9897610921501706, + "grad_norm": 0.671875, + "learning_rate": 0.0001318181818181818, + "loss": 1.2085, + "step": 290 + }, + { + "epoch": 1.0, + "eval_loss": 2.486323833465576, + "eval_runtime": 0.5451, + "eval_samples_per_second": 18.345, + "eval_steps_per_second": 1.834, + "step": 293 + }, + { + "epoch": 1.006825938566553, + "grad_norm": 0.52734375, + "learning_rate": 0.0001340909090909091, + "loss": 1.1892, + "step": 295 + }, + { + "epoch": 1.023890784982935, + "grad_norm": 0.4296875, + "learning_rate": 0.00013636363636363637, + "loss": 1.191, + "step": 300 + }, + { + "epoch": 1.0409556313993173, + "grad_norm": 0.5390625, + "learning_rate": 0.00013863636363636365, + "loss": 1.18, + "step": 305 + }, + { + "epoch": 1.0580204778156996, + "grad_norm": 0.703125, + "learning_rate": 0.00014090909090909093, + "loss": 1.1964, + "step": 310 + }, + { + "epoch": 1.075085324232082, + "grad_norm": 0.435546875, + "learning_rate": 0.0001431818181818182, + "loss": 1.1877, + "step": 315 + }, + { + "epoch": 1.0921501706484642, + "grad_norm": 0.59375, + "learning_rate": 0.00014545454545454546, + "loss": 1.1846, + "step": 320 + }, + { + "epoch": 1.1092150170648465, + "grad_norm": 0.8125, + "learning_rate": 0.00014772727272727274, + "loss": 1.1833, + "step": 325 + }, + { + "epoch": 1.1262798634812285, + "grad_norm": 0.9296875, + "learning_rate": 0.00015000000000000001, + "loss": 1.1704, + "step": 330 + }, + { + "epoch": 1.1433447098976108, + "grad_norm": 1.5703125, + "learning_rate": 0.00015227272727272727, + "loss": 1.1886, + "step": 335 + }, + { + "epoch": 1.1604095563139931, + "grad_norm": 0.4609375, + "learning_rate": 0.00015454545454545454, + "loss": 1.1759, + "step": 340 + }, + { + "epoch": 1.1774744027303754, + "grad_norm": 0.337890625, + "learning_rate": 0.00015681818181818182, + "loss": 1.1712, + "step": 345 + }, + { + "epoch": 1.1945392491467577, + "grad_norm": 0.48046875, + "learning_rate": 0.0001590909090909091, + "loss": 1.1637, + "step": 350 + }, + { + "epoch": 1.21160409556314, + "grad_norm": 0.58203125, + "learning_rate": 0.00016136363636363635, + "loss": 1.1657, + "step": 355 + }, + { + "epoch": 1.2286689419795223, + "grad_norm": 0.54296875, + "learning_rate": 0.00016363636363636366, + "loss": 1.1745, + "step": 360 + }, + { + "epoch": 1.2457337883959045, + "grad_norm": 0.421875, + "learning_rate": 0.00016590909090909094, + "loss": 1.1496, + "step": 365 + }, + { + "epoch": 1.2627986348122868, + "grad_norm": 0.546875, + "learning_rate": 0.0001681818181818182, + "loss": 1.1653, + "step": 370 + }, + { + "epoch": 1.2798634812286689, + "grad_norm": 0.5078125, + "learning_rate": 0.00017045454545454547, + "loss": 1.1702, + "step": 375 + }, + { + "epoch": 1.2969283276450512, + "grad_norm": 0.53125, + "learning_rate": 0.00017272727272727275, + "loss": 1.15, + "step": 380 + }, + { + "epoch": 1.3139931740614335, + "grad_norm": 1.2421875, + "learning_rate": 0.000175, + "loss": 1.1615, + "step": 385 + }, + { + "epoch": 1.3310580204778157, + "grad_norm": 1.625, + "learning_rate": 0.00017727272727272728, + "loss": 1.1662, + "step": 390 + }, + { + "epoch": 1.348122866894198, + "grad_norm": 0.5, + "learning_rate": 0.00017954545454545456, + "loss": 1.1579, + "step": 395 + }, + { + "epoch": 1.36518771331058, + "grad_norm": 0.90234375, + "learning_rate": 0.00018181818181818183, + "loss": 1.1628, + "step": 400 + }, + { + "epoch": 1.3822525597269624, + "grad_norm": 0.3515625, + "learning_rate": 0.00018409090909090909, + "loss": 1.1521, + "step": 405 + }, + { + "epoch": 1.3993174061433447, + "grad_norm": 0.455078125, + "learning_rate": 0.00018636363636363636, + "loss": 1.1422, + "step": 410 + }, + { + "epoch": 1.416382252559727, + "grad_norm": 0.52734375, + "learning_rate": 0.00018863636363636364, + "loss": 1.1408, + "step": 415 + }, + { + "epoch": 1.4334470989761092, + "grad_norm": 0.53515625, + "learning_rate": 0.00019090909090909092, + "loss": 1.1356, + "step": 420 + }, + { + "epoch": 1.4505119453924915, + "grad_norm": 0.46875, + "learning_rate": 0.0001931818181818182, + "loss": 1.1497, + "step": 425 + }, + { + "epoch": 1.4675767918088738, + "grad_norm": 0.47265625, + "learning_rate": 0.00019545454545454548, + "loss": 1.1437, + "step": 430 + }, + { + "epoch": 1.484641638225256, + "grad_norm": 0.42578125, + "learning_rate": 0.00019772727272727273, + "loss": 1.1518, + "step": 435 + }, + { + "epoch": 1.5017064846416384, + "grad_norm": 0.38671875, + "learning_rate": 0.0002, + "loss": 1.1518, + "step": 440 + }, + { + "epoch": 1.5187713310580204, + "grad_norm": 0.3984375, + "learning_rate": 0.000199999211292062, + "loss": 1.1498, + "step": 445 + }, + { + "epoch": 1.5358361774744027, + "grad_norm": 0.388671875, + "learning_rate": 0.00019999684518068916, + "loss": 1.1378, + "step": 450 + }, + { + "epoch": 1.552901023890785, + "grad_norm": 0.87109375, + "learning_rate": 0.00019999290170320485, + "loss": 1.1434, + "step": 455 + }, + { + "epoch": 1.5699658703071673, + "grad_norm": 0.6953125, + "learning_rate": 0.00019998738092181421, + "loss": 1.1417, + "step": 460 + }, + { + "epoch": 1.5870307167235493, + "grad_norm": 0.79296875, + "learning_rate": 0.00019998028292360286, + "loss": 1.1329, + "step": 465 + }, + { + "epoch": 1.6040955631399316, + "grad_norm": 0.5625, + "learning_rate": 0.00019997160782053578, + "loss": 1.1339, + "step": 470 + }, + { + "epoch": 1.621160409556314, + "grad_norm": 0.384765625, + "learning_rate": 0.00019996135574945544, + "loss": 1.1273, + "step": 475 + }, + { + "epoch": 1.6382252559726962, + "grad_norm": 0.455078125, + "learning_rate": 0.00019994952687207954, + "loss": 1.1343, + "step": 480 + }, + { + "epoch": 1.6552901023890785, + "grad_norm": 0.69921875, + "learning_rate": 0.00019993612137499876, + "loss": 1.1374, + "step": 485 + }, + { + "epoch": 1.6723549488054608, + "grad_norm": 0.88671875, + "learning_rate": 0.00019992113946967353, + "loss": 1.1368, + "step": 490 + }, + { + "epoch": 1.689419795221843, + "grad_norm": 0.490234375, + "learning_rate": 0.00019990458139243077, + "loss": 1.1289, + "step": 495 + }, + { + "epoch": 1.7064846416382253, + "grad_norm": 0.53515625, + "learning_rate": 0.00019988644740446022, + "loss": 1.1255, + "step": 500 + }, + { + "epoch": 1.7235494880546076, + "grad_norm": 0.5234375, + "learning_rate": 0.00019986673779181033, + "loss": 1.1149, + "step": 505 + }, + { + "epoch": 1.74061433447099, + "grad_norm": 0.41015625, + "learning_rate": 0.0001998454528653836, + "loss": 1.1241, + "step": 510 + }, + { + "epoch": 1.757679180887372, + "grad_norm": 0.41796875, + "learning_rate": 0.0001998225929609319, + "loss": 1.1252, + "step": 515 + }, + { + "epoch": 1.7747440273037542, + "grad_norm": 0.458984375, + "learning_rate": 0.00019979815843905097, + "loss": 1.1292, + "step": 520 + }, + { + "epoch": 1.7918088737201365, + "grad_norm": 0.400390625, + "learning_rate": 0.0001997721496851748, + "loss": 1.1147, + "step": 525 + }, + { + "epoch": 1.8088737201365188, + "grad_norm": 0.53125, + "learning_rate": 0.00019974456710956964, + "loss": 1.1155, + "step": 530 + }, + { + "epoch": 1.8259385665529009, + "grad_norm": 0.546875, + "learning_rate": 0.00019971541114732741, + "loss": 1.1213, + "step": 535 + }, + { + "epoch": 1.8430034129692832, + "grad_norm": 0.40234375, + "learning_rate": 0.0001996846822583589, + "loss": 1.1257, + "step": 540 + }, + { + "epoch": 1.8600682593856654, + "grad_norm": 0.38671875, + "learning_rate": 0.00019965238092738643, + "loss": 1.1217, + "step": 545 + }, + { + "epoch": 1.8771331058020477, + "grad_norm": 0.5390625, + "learning_rate": 0.0001996185076639364, + "loss": 1.122, + "step": 550 + }, + { + "epoch": 1.89419795221843, + "grad_norm": 0.390625, + "learning_rate": 0.00019958306300233098, + "loss": 1.1236, + "step": 555 + }, + { + "epoch": 1.9112627986348123, + "grad_norm": 0.5390625, + "learning_rate": 0.00019954604750167993, + "loss": 1.122, + "step": 560 + }, + { + "epoch": 1.9283276450511946, + "grad_norm": 0.66796875, + "learning_rate": 0.00019950746174587163, + "loss": 1.1271, + "step": 565 + }, + { + "epoch": 1.9453924914675769, + "grad_norm": 0.47265625, + "learning_rate": 0.0001994673063435639, + "loss": 1.1064, + "step": 570 + }, + { + "epoch": 1.9624573378839592, + "grad_norm": 0.3359375, + "learning_rate": 0.0001994255819281744, + "loss": 1.1186, + "step": 575 + }, + { + "epoch": 1.9795221843003414, + "grad_norm": 0.63671875, + "learning_rate": 0.0001993822891578708, + "loss": 1.1054, + "step": 580 + }, + { + "epoch": 1.9965870307167235, + "grad_norm": 0.68359375, + "learning_rate": 0.00019933742871556, + "loss": 1.1135, + "step": 585 + }, + { + "epoch": 2.0, + "eval_loss": 2.4516425132751465, + "eval_runtime": 0.5387, + "eval_samples_per_second": 18.563, + "eval_steps_per_second": 1.856, + "step": 586 + }, + { + "epoch": 2.013651877133106, + "grad_norm": 0.66015625, + "learning_rate": 0.00019929100130887782, + "loss": 1.1079, + "step": 590 + }, + { + "epoch": 2.030716723549488, + "grad_norm": 0.94140625, + "learning_rate": 0.0001992430076701775, + "loss": 1.088, + "step": 595 + }, + { + "epoch": 2.04778156996587, + "grad_norm": 0.400390625, + "learning_rate": 0.00019919344855651833, + "loss": 1.0921, + "step": 600 + }, + { + "epoch": 2.0648464163822524, + "grad_norm": 0.59375, + "learning_rate": 0.00019914232474965365, + "loss": 1.0909, + "step": 605 + }, + { + "epoch": 2.0819112627986347, + "grad_norm": 0.42578125, + "learning_rate": 0.00019908963705601846, + "loss": 1.0986, + "step": 610 + }, + { + "epoch": 2.098976109215017, + "grad_norm": 0.435546875, + "learning_rate": 0.0001990353863067169, + "loss": 1.0925, + "step": 615 + }, + { + "epoch": 2.1160409556313993, + "grad_norm": 0.640625, + "learning_rate": 0.00019897957335750878, + "loss": 1.0887, + "step": 620 + }, + { + "epoch": 2.1331058020477816, + "grad_norm": 0.5078125, + "learning_rate": 0.00019892219908879653, + "loss": 1.0991, + "step": 625 + }, + { + "epoch": 2.150170648464164, + "grad_norm": 0.416015625, + "learning_rate": 0.00019886326440561093, + "loss": 1.0949, + "step": 630 + }, + { + "epoch": 2.167235494880546, + "grad_norm": 0.373046875, + "learning_rate": 0.00019880277023759702, + "loss": 1.0841, + "step": 635 + }, + { + "epoch": 2.1843003412969284, + "grad_norm": 0.78515625, + "learning_rate": 0.0001987407175389994, + "loss": 1.0947, + "step": 640 + }, + { + "epoch": 2.2013651877133107, + "grad_norm": 0.42578125, + "learning_rate": 0.0001986771072886472, + "loss": 1.1026, + "step": 645 + }, + { + "epoch": 2.218430034129693, + "grad_norm": 0.392578125, + "learning_rate": 0.00019861194048993863, + "loss": 1.0918, + "step": 650 + }, + { + "epoch": 2.2354948805460753, + "grad_norm": 0.41015625, + "learning_rate": 0.0001985452181708251, + "loss": 1.0903, + "step": 655 + }, + { + "epoch": 2.252559726962457, + "grad_norm": 0.7109375, + "learning_rate": 0.00019847694138379506, + "loss": 1.0978, + "step": 660 + }, + { + "epoch": 2.26962457337884, + "grad_norm": 0.4609375, + "learning_rate": 0.0001984071112058574, + "loss": 1.0864, + "step": 665 + }, + { + "epoch": 2.2866894197952217, + "grad_norm": 0.341796875, + "learning_rate": 0.00019833572873852444, + "loss": 1.0896, + "step": 670 + }, + { + "epoch": 2.303754266211604, + "grad_norm": 0.53125, + "learning_rate": 0.00019826279510779454, + "loss": 1.0962, + "step": 675 + }, + { + "epoch": 2.3208191126279862, + "grad_norm": 0.54296875, + "learning_rate": 0.00019818831146413434, + "loss": 1.0766, + "step": 680 + }, + { + "epoch": 2.3378839590443685, + "grad_norm": 0.337890625, + "learning_rate": 0.0001981122789824607, + "loss": 1.0853, + "step": 685 + }, + { + "epoch": 2.354948805460751, + "grad_norm": 0.69140625, + "learning_rate": 0.0001980346988621221, + "loss": 1.0788, + "step": 690 + }, + { + "epoch": 2.372013651877133, + "grad_norm": 0.96875, + "learning_rate": 0.00019795557232687956, + "loss": 1.0804, + "step": 695 + }, + { + "epoch": 2.3890784982935154, + "grad_norm": 0.470703125, + "learning_rate": 0.0001978749006248877, + "loss": 1.0674, + "step": 700 + }, + { + "epoch": 2.4061433447098977, + "grad_norm": 0.326171875, + "learning_rate": 0.00019779268502867473, + "loss": 1.0931, + "step": 705 + }, + { + "epoch": 2.42320819112628, + "grad_norm": 0.458984375, + "learning_rate": 0.0001977089268351225, + "loss": 1.0854, + "step": 710 + }, + { + "epoch": 2.4402730375426622, + "grad_norm": 0.43359375, + "learning_rate": 0.00019762362736544607, + "loss": 1.0858, + "step": 715 + }, + { + "epoch": 2.4573378839590445, + "grad_norm": 0.396484375, + "learning_rate": 0.00019753678796517282, + "loss": 1.0835, + "step": 720 + }, + { + "epoch": 2.474402730375427, + "grad_norm": 0.59375, + "learning_rate": 0.00019744841000412123, + "loss": 1.0881, + "step": 725 + }, + { + "epoch": 2.491467576791809, + "grad_norm": 0.6171875, + "learning_rate": 0.00019735849487637929, + "loss": 1.091, + "step": 730 + }, + { + "epoch": 2.508532423208191, + "grad_norm": 0.5625, + "learning_rate": 0.0001972670440002825, + "loss": 1.0877, + "step": 735 + }, + { + "epoch": 2.5255972696245736, + "grad_norm": 0.419921875, + "learning_rate": 0.00019717405881839145, + "loss": 1.0777, + "step": 740 + }, + { + "epoch": 2.5426621160409555, + "grad_norm": 0.380859375, + "learning_rate": 0.00019707954079746927, + "loss": 1.0934, + "step": 745 + }, + { + "epoch": 2.5597269624573378, + "grad_norm": 0.439453125, + "learning_rate": 0.00019698349142845814, + "loss": 1.085, + "step": 750 + }, + { + "epoch": 2.57679180887372, + "grad_norm": 0.38671875, + "learning_rate": 0.00019688591222645607, + "loss": 1.0744, + "step": 755 + }, + { + "epoch": 2.5938566552901023, + "grad_norm": 0.4375, + "learning_rate": 0.00019678680473069293, + "loss": 1.0818, + "step": 760 + }, + { + "epoch": 2.6109215017064846, + "grad_norm": 0.3984375, + "learning_rate": 0.00019668617050450603, + "loss": 1.0824, + "step": 765 + }, + { + "epoch": 2.627986348122867, + "grad_norm": 0.4921875, + "learning_rate": 0.00019658401113531565, + "loss": 1.0828, + "step": 770 + }, + { + "epoch": 2.645051194539249, + "grad_norm": 1.09375, + "learning_rate": 0.00019648032823459994, + "loss": 1.0884, + "step": 775 + }, + { + "epoch": 2.6621160409556315, + "grad_norm": 0.55859375, + "learning_rate": 0.00019637512343786937, + "loss": 1.0835, + "step": 780 + }, + { + "epoch": 2.6791808873720138, + "grad_norm": 0.484375, + "learning_rate": 0.00019626839840464119, + "loss": 1.0828, + "step": 785 + }, + { + "epoch": 2.696245733788396, + "grad_norm": 0.376953125, + "learning_rate": 0.0001961601548184129, + "loss": 1.0881, + "step": 790 + }, + { + "epoch": 2.7133105802047783, + "grad_norm": 0.35546875, + "learning_rate": 0.00019605039438663614, + "loss": 1.0772, + "step": 795 + }, + { + "epoch": 2.73037542662116, + "grad_norm": 0.349609375, + "learning_rate": 0.0001959391188406893, + "loss": 1.0677, + "step": 800 + }, + { + "epoch": 2.747440273037543, + "grad_norm": 0.486328125, + "learning_rate": 0.00019582632993585052, + "loss": 1.0815, + "step": 805 + }, + { + "epoch": 2.7645051194539247, + "grad_norm": 0.470703125, + "learning_rate": 0.00019571202945126994, + "loss": 1.0763, + "step": 810 + }, + { + "epoch": 2.781569965870307, + "grad_norm": 0.396484375, + "learning_rate": 0.0001955962191899415, + "loss": 1.0684, + "step": 815 + }, + { + "epoch": 2.7986348122866893, + "grad_norm": 0.373046875, + "learning_rate": 0.00019547890097867468, + "loss": 1.0847, + "step": 820 + }, + { + "epoch": 2.8156996587030716, + "grad_norm": 0.474609375, + "learning_rate": 0.00019536007666806556, + "loss": 1.071, + "step": 825 + }, + { + "epoch": 2.832764505119454, + "grad_norm": 0.380859375, + "learning_rate": 0.00019523974813246767, + "loss": 1.0873, + "step": 830 + }, + { + "epoch": 2.849829351535836, + "grad_norm": 0.40234375, + "learning_rate": 0.00019511791726996243, + "loss": 1.0676, + "step": 835 + }, + { + "epoch": 2.8668941979522184, + "grad_norm": 0.51953125, + "learning_rate": 0.0001949945860023292, + "loss": 1.0748, + "step": 840 + }, + { + "epoch": 2.8839590443686007, + "grad_norm": 0.384765625, + "learning_rate": 0.00019486975627501502, + "loss": 1.0716, + "step": 845 + }, + { + "epoch": 2.901023890784983, + "grad_norm": 0.38671875, + "learning_rate": 0.0001947434300571038, + "loss": 1.0777, + "step": 850 + }, + { + "epoch": 2.9180887372013653, + "grad_norm": 0.365234375, + "learning_rate": 0.00019461560934128533, + "loss": 1.0733, + "step": 855 + }, + { + "epoch": 2.9351535836177476, + "grad_norm": 0.42578125, + "learning_rate": 0.0001944862961438239, + "loss": 1.0582, + "step": 860 + }, + { + "epoch": 2.9522184300341294, + "grad_norm": 0.462890625, + "learning_rate": 0.00019435549250452645, + "loss": 1.0657, + "step": 865 + }, + { + "epoch": 2.969283276450512, + "grad_norm": 1.1171875, + "learning_rate": 0.0001942232004867103, + "loss": 1.0746, + "step": 870 + }, + { + "epoch": 2.986348122866894, + "grad_norm": 0.466796875, + "learning_rate": 0.0001940894221771708, + "loss": 1.0715, + "step": 875 + }, + { + "epoch": 3.0, + "eval_loss": 2.447284698486328, + "eval_runtime": 0.553, + "eval_samples_per_second": 18.083, + "eval_steps_per_second": 1.808, + "step": 879 + }, + { + "epoch": 3.0034129692832763, + "grad_norm": 0.80859375, + "learning_rate": 0.00019395415968614813, + "loss": 1.0736, + "step": 880 + }, + { + "epoch": 3.0204778156996586, + "grad_norm": 0.47265625, + "learning_rate": 0.00019381741514729443, + "loss": 1.0618, + "step": 885 + }, + { + "epoch": 3.037542662116041, + "grad_norm": 0.390625, + "learning_rate": 0.0001936791907176397, + "loss": 1.0571, + "step": 890 + }, + { + "epoch": 3.054607508532423, + "grad_norm": 0.46484375, + "learning_rate": 0.00019353948857755803, + "loss": 1.0626, + "step": 895 + }, + { + "epoch": 3.0716723549488054, + "grad_norm": 0.357421875, + "learning_rate": 0.00019339831093073318, + "loss": 1.053, + "step": 900 + }, + { + "epoch": 3.0887372013651877, + "grad_norm": 0.380859375, + "learning_rate": 0.00019325566000412376, + "loss": 1.06, + "step": 905 + }, + { + "epoch": 3.10580204778157, + "grad_norm": 0.38671875, + "learning_rate": 0.0001931115380479281, + "loss": 1.0452, + "step": 910 + }, + { + "epoch": 3.1228668941979523, + "grad_norm": 0.515625, + "learning_rate": 0.00019296594733554892, + "loss": 1.0642, + "step": 915 + }, + { + "epoch": 3.1399317406143346, + "grad_norm": 0.5, + "learning_rate": 0.0001928188901635571, + "loss": 1.0474, + "step": 920 + }, + { + "epoch": 3.156996587030717, + "grad_norm": 0.380859375, + "learning_rate": 0.00019267036885165588, + "loss": 1.0526, + "step": 925 + }, + { + "epoch": 3.174061433447099, + "grad_norm": 0.4296875, + "learning_rate": 0.00019252038574264405, + "loss": 1.061, + "step": 930 + }, + { + "epoch": 3.1911262798634814, + "grad_norm": 0.443359375, + "learning_rate": 0.00019236894320237894, + "loss": 1.0519, + "step": 935 + }, + { + "epoch": 3.2081911262798632, + "grad_norm": 0.458984375, + "learning_rate": 0.00019221604361973919, + "loss": 1.0479, + "step": 940 + }, + { + "epoch": 3.2252559726962455, + "grad_norm": 0.50390625, + "learning_rate": 0.00019206168940658712, + "loss": 1.049, + "step": 945 + }, + { + "epoch": 3.242320819112628, + "grad_norm": 0.462890625, + "learning_rate": 0.00019190588299773062, + "loss": 1.0474, + "step": 950 + }, + { + "epoch": 3.25938566552901, + "grad_norm": 0.462890625, + "learning_rate": 0.00019174862685088472, + "loss": 1.06, + "step": 955 + }, + { + "epoch": 3.2764505119453924, + "grad_norm": 0.373046875, + "learning_rate": 0.0001915899234466328, + "loss": 1.0464, + "step": 960 + }, + { + "epoch": 3.2935153583617747, + "grad_norm": 0.48046875, + "learning_rate": 0.00019142977528838762, + "loss": 1.0531, + "step": 965 + }, + { + "epoch": 3.310580204778157, + "grad_norm": 0.380859375, + "learning_rate": 0.0001912681849023516, + "loss": 1.0518, + "step": 970 + }, + { + "epoch": 3.3276450511945392, + "grad_norm": 0.447265625, + "learning_rate": 0.00019110515483747716, + "loss": 1.0535, + "step": 975 + }, + { + "epoch": 3.3447098976109215, + "grad_norm": 0.625, + "learning_rate": 0.0001909406876654264, + "loss": 1.0559, + "step": 980 + }, + { + "epoch": 3.361774744027304, + "grad_norm": 0.51953125, + "learning_rate": 0.00019077478598053063, + "loss": 1.0528, + "step": 985 + }, + { + "epoch": 3.378839590443686, + "grad_norm": 0.46875, + "learning_rate": 0.00019060745239974936, + "loss": 1.0431, + "step": 990 + }, + { + "epoch": 3.3959044368600684, + "grad_norm": 0.63671875, + "learning_rate": 0.0001904386895626291, + "loss": 1.0456, + "step": 995 + }, + { + "epoch": 3.4129692832764507, + "grad_norm": 0.48828125, + "learning_rate": 0.00019026850013126157, + "loss": 1.0579, + "step": 1000 + }, + { + "epoch": 3.430034129692833, + "grad_norm": 0.625, + "learning_rate": 0.0001900968867902419, + "loss": 1.0592, + "step": 1005 + }, + { + "epoch": 3.4470989761092152, + "grad_norm": 0.51171875, + "learning_rate": 0.00018992385224662623, + "loss": 1.0476, + "step": 1010 + }, + { + "epoch": 3.464163822525597, + "grad_norm": 0.470703125, + "learning_rate": 0.00018974939922988883, + "loss": 1.0517, + "step": 1015 + }, + { + "epoch": 3.4812286689419794, + "grad_norm": 0.423828125, + "learning_rate": 0.00018957353049187936, + "loss": 1.0607, + "step": 1020 + }, + { + "epoch": 3.4982935153583616, + "grad_norm": 0.4765625, + "learning_rate": 0.00018939624880677918, + "loss": 1.0502, + "step": 1025 + }, + { + "epoch": 3.515358361774744, + "grad_norm": 0.3671875, + "learning_rate": 0.0001892175569710577, + "loss": 1.041, + "step": 1030 + }, + { + "epoch": 3.532423208191126, + "grad_norm": 0.52734375, + "learning_rate": 0.00018903745780342839, + "loss": 1.0382, + "step": 1035 + }, + { + "epoch": 3.5494880546075085, + "grad_norm": 0.3984375, + "learning_rate": 0.00018885595414480405, + "loss": 1.0426, + "step": 1040 + }, + { + "epoch": 3.5665529010238908, + "grad_norm": 0.400390625, + "learning_rate": 0.0001886730488582522, + "loss": 1.0524, + "step": 1045 + }, + { + "epoch": 3.583617747440273, + "grad_norm": 0.58203125, + "learning_rate": 0.00018848874482894993, + "loss": 1.0371, + "step": 1050 + }, + { + "epoch": 3.6006825938566553, + "grad_norm": 0.412109375, + "learning_rate": 0.00018830304496413822, + "loss": 1.0571, + "step": 1055 + }, + { + "epoch": 3.6177474402730376, + "grad_norm": 0.33984375, + "learning_rate": 0.00018811595219307622, + "loss": 1.0458, + "step": 1060 + }, + { + "epoch": 3.63481228668942, + "grad_norm": 0.455078125, + "learning_rate": 0.000187927469466995, + "loss": 1.0474, + "step": 1065 + }, + { + "epoch": 3.651877133105802, + "grad_norm": 0.37109375, + "learning_rate": 0.00018773759975905098, + "loss": 1.0438, + "step": 1070 + }, + { + "epoch": 3.6689419795221845, + "grad_norm": 0.384765625, + "learning_rate": 0.00018754634606427914, + "loss": 1.0577, + "step": 1075 + }, + { + "epoch": 3.6860068259385663, + "grad_norm": 0.435546875, + "learning_rate": 0.00018735371139954558, + "loss": 1.0522, + "step": 1080 + }, + { + "epoch": 3.703071672354949, + "grad_norm": 0.55859375, + "learning_rate": 0.0001871596988035001, + "loss": 1.0622, + "step": 1085 + }, + { + "epoch": 3.720136518771331, + "grad_norm": 0.53125, + "learning_rate": 0.00018696431133652817, + "loss": 1.0404, + "step": 1090 + }, + { + "epoch": 3.737201365187713, + "grad_norm": 0.41796875, + "learning_rate": 0.00018676755208070275, + "loss": 1.0576, + "step": 1095 + }, + { + "epoch": 3.7542662116040955, + "grad_norm": 0.396484375, + "learning_rate": 0.00018656942413973555, + "loss": 1.0525, + "step": 1100 + }, + { + "epoch": 3.7713310580204777, + "grad_norm": 0.392578125, + "learning_rate": 0.0001863699306389282, + "loss": 1.047, + "step": 1105 + }, + { + "epoch": 3.78839590443686, + "grad_norm": 0.54296875, + "learning_rate": 0.0001861690747251228, + "loss": 1.0547, + "step": 1110 + }, + { + "epoch": 3.8054607508532423, + "grad_norm": 0.455078125, + "learning_rate": 0.00018596685956665245, + "loss": 1.0366, + "step": 1115 + }, + { + "epoch": 3.8225255972696246, + "grad_norm": 0.373046875, + "learning_rate": 0.00018576328835329117, + "loss": 1.0444, + "step": 1120 + }, + { + "epoch": 3.839590443686007, + "grad_norm": 0.498046875, + "learning_rate": 0.00018555836429620358, + "loss": 1.0428, + "step": 1125 + }, + { + "epoch": 3.856655290102389, + "grad_norm": 0.4453125, + "learning_rate": 0.00018535209062789433, + "loss": 1.0425, + "step": 1130 + }, + { + "epoch": 3.8737201365187715, + "grad_norm": 0.392578125, + "learning_rate": 0.00018514447060215698, + "loss": 1.0503, + "step": 1135 + }, + { + "epoch": 3.8907849829351537, + "grad_norm": 0.384765625, + "learning_rate": 0.00018493550749402278, + "loss": 1.0376, + "step": 1140 + }, + { + "epoch": 3.9078498293515356, + "grad_norm": 0.3984375, + "learning_rate": 0.00018472520459970898, + "loss": 1.054, + "step": 1145 + }, + { + "epoch": 3.9249146757679183, + "grad_norm": 0.44921875, + "learning_rate": 0.0001845135652365668, + "loss": 1.0491, + "step": 1150 + }, + { + "epoch": 3.9419795221843, + "grad_norm": 0.37890625, + "learning_rate": 0.00018430059274302917, + "loss": 1.0454, + "step": 1155 + }, + { + "epoch": 3.9590443686006824, + "grad_norm": 0.365234375, + "learning_rate": 0.00018408629047855804, + "loss": 1.0466, + "step": 1160 + }, + { + "epoch": 3.9761092150170647, + "grad_norm": 0.34765625, + "learning_rate": 0.00018387066182359133, + "loss": 1.0356, + "step": 1165 + }, + { + "epoch": 3.993174061433447, + "grad_norm": 0.357421875, + "learning_rate": 0.00018365371017948964, + "loss": 1.0471, + "step": 1170 + }, + { + "epoch": 4.0, + "eval_loss": 2.452413558959961, + "eval_runtime": 0.5427, + "eval_samples_per_second": 18.427, + "eval_steps_per_second": 1.843, + "step": 1172 + }, + { + "epoch": 4.010238907849829, + "grad_norm": 0.47265625, + "learning_rate": 0.00018343543896848273, + "loss": 1.0282, + "step": 1175 + }, + { + "epoch": 4.027303754266212, + "grad_norm": 0.41796875, + "learning_rate": 0.00018321585163361527, + "loss": 1.0262, + "step": 1180 + }, + { + "epoch": 4.044368600682594, + "grad_norm": 0.365234375, + "learning_rate": 0.00018299495163869275, + "loss": 1.0263, + "step": 1185 + }, + { + "epoch": 4.061433447098976, + "grad_norm": 0.359375, + "learning_rate": 0.0001827727424682268, + "loss": 1.0265, + "step": 1190 + }, + { + "epoch": 4.078498293515358, + "grad_norm": 0.375, + "learning_rate": 0.00018254922762738008, + "loss": 1.0266, + "step": 1195 + }, + { + "epoch": 4.09556313993174, + "grad_norm": 0.3828125, + "learning_rate": 0.00018232441064191125, + "loss": 1.0326, + "step": 1200 + }, + { + "epoch": 4.112627986348123, + "grad_norm": 0.3828125, + "learning_rate": 0.0001820982950581191, + "loss": 1.0278, + "step": 1205 + }, + { + "epoch": 4.129692832764505, + "grad_norm": 0.46484375, + "learning_rate": 0.00018187088444278674, + "loss": 1.0206, + "step": 1210 + }, + { + "epoch": 4.146757679180888, + "grad_norm": 0.4140625, + "learning_rate": 0.00018164218238312535, + "loss": 1.037, + "step": 1215 + }, + { + "epoch": 4.163822525597269, + "grad_norm": 0.3671875, + "learning_rate": 0.00018141219248671745, + "loss": 1.0229, + "step": 1220 + }, + { + "epoch": 4.180887372013652, + "grad_norm": 0.376953125, + "learning_rate": 0.00018118091838146029, + "loss": 1.0223, + "step": 1225 + }, + { + "epoch": 4.197952218430034, + "grad_norm": 0.373046875, + "learning_rate": 0.00018094836371550824, + "loss": 1.0175, + "step": 1230 + }, + { + "epoch": 4.215017064846417, + "grad_norm": 0.380859375, + "learning_rate": 0.00018071453215721554, + "loss": 1.0369, + "step": 1235 + }, + { + "epoch": 4.2320819112627985, + "grad_norm": 0.41015625, + "learning_rate": 0.00018047942739507836, + "loss": 1.0182, + "step": 1240 + }, + { + "epoch": 4.249146757679181, + "grad_norm": 0.421875, + "learning_rate": 0.00018024305313767646, + "loss": 1.0192, + "step": 1245 + }, + { + "epoch": 4.266211604095563, + "grad_norm": 0.40625, + "learning_rate": 0.000180005413113615, + "loss": 1.0427, + "step": 1250 + }, + { + "epoch": 4.283276450511945, + "grad_norm": 0.42578125, + "learning_rate": 0.00017976651107146533, + "loss": 1.0313, + "step": 1255 + }, + { + "epoch": 4.300341296928328, + "grad_norm": 0.359375, + "learning_rate": 0.0001795263507797063, + "loss": 1.0195, + "step": 1260 + }, + { + "epoch": 4.3174061433447095, + "grad_norm": 0.453125, + "learning_rate": 0.00017928493602666445, + "loss": 1.0222, + "step": 1265 + }, + { + "epoch": 4.334470989761092, + "grad_norm": 0.5546875, + "learning_rate": 0.00017904227062045437, + "loss": 1.0183, + "step": 1270 + }, + { + "epoch": 4.351535836177474, + "grad_norm": 0.6328125, + "learning_rate": 0.00017879835838891875, + "loss": 1.0321, + "step": 1275 + }, + { + "epoch": 4.368600682593857, + "grad_norm": 0.7265625, + "learning_rate": 0.00017855320317956784, + "loss": 1.0241, + "step": 1280 + }, + { + "epoch": 4.385665529010239, + "grad_norm": 0.380859375, + "learning_rate": 0.00017830680885951887, + "loss": 1.019, + "step": 1285 + }, + { + "epoch": 4.402730375426621, + "grad_norm": 0.7265625, + "learning_rate": 0.00017805917931543492, + "loss": 1.0291, + "step": 1290 + }, + { + "epoch": 4.419795221843003, + "grad_norm": 0.8671875, + "learning_rate": 0.00017781031845346375, + "loss": 1.0254, + "step": 1295 + }, + { + "epoch": 4.436860068259386, + "grad_norm": 0.38671875, + "learning_rate": 0.00017756023019917607, + "loss": 1.0232, + "step": 1300 + }, + { + "epoch": 4.453924914675768, + "grad_norm": 0.384765625, + "learning_rate": 0.00017730891849750377, + "loss": 1.0267, + "step": 1305 + }, + { + "epoch": 4.4709897610921505, + "grad_norm": 0.38671875, + "learning_rate": 0.0001770563873126775, + "loss": 1.0282, + "step": 1310 + }, + { + "epoch": 4.488054607508532, + "grad_norm": 0.357421875, + "learning_rate": 0.0001768026406281642, + "loss": 1.0384, + "step": 1315 + }, + { + "epoch": 4.505119453924914, + "grad_norm": 0.37109375, + "learning_rate": 0.00017654768244660448, + "loss": 1.0197, + "step": 1320 + }, + { + "epoch": 4.522184300341297, + "grad_norm": 0.458984375, + "learning_rate": 0.00017629151678974907, + "loss": 1.023, + "step": 1325 + }, + { + "epoch": 4.53924914675768, + "grad_norm": 0.359375, + "learning_rate": 0.00017603414769839577, + "loss": 1.0289, + "step": 1330 + }, + { + "epoch": 4.5563139931740615, + "grad_norm": 0.72265625, + "learning_rate": 0.00017577557923232546, + "loss": 1.0222, + "step": 1335 + }, + { + "epoch": 4.573378839590443, + "grad_norm": 0.5, + "learning_rate": 0.00017551581547023819, + "loss": 1.0285, + "step": 1340 + }, + { + "epoch": 4.590443686006826, + "grad_norm": 0.392578125, + "learning_rate": 0.00017525486050968875, + "loss": 1.0288, + "step": 1345 + }, + { + "epoch": 4.607508532423208, + "grad_norm": 0.37890625, + "learning_rate": 0.00017499271846702213, + "loss": 1.0302, + "step": 1350 + }, + { + "epoch": 4.624573378839591, + "grad_norm": 0.419921875, + "learning_rate": 0.00017472939347730856, + "loss": 1.0358, + "step": 1355 + }, + { + "epoch": 4.6416382252559725, + "grad_norm": 0.451171875, + "learning_rate": 0.0001744648896942782, + "loss": 1.0278, + "step": 1360 + }, + { + "epoch": 4.658703071672355, + "grad_norm": 0.38671875, + "learning_rate": 0.00017419921129025576, + "loss": 1.0171, + "step": 1365 + }, + { + "epoch": 4.675767918088737, + "grad_norm": 0.376953125, + "learning_rate": 0.0001739323624560945, + "loss": 1.0152, + "step": 1370 + }, + { + "epoch": 4.69283276450512, + "grad_norm": 0.384765625, + "learning_rate": 0.00017366434740111037, + "loss": 1.0247, + "step": 1375 + }, + { + "epoch": 4.709897610921502, + "grad_norm": 0.431640625, + "learning_rate": 0.00017339517035301532, + "loss": 1.0212, + "step": 1380 + }, + { + "epoch": 4.726962457337884, + "grad_norm": 0.3828125, + "learning_rate": 0.00017312483555785086, + "loss": 1.0309, + "step": 1385 + }, + { + "epoch": 4.744027303754266, + "grad_norm": 0.353515625, + "learning_rate": 0.000172853347279921, + "loss": 1.0298, + "step": 1390 + }, + { + "epoch": 4.761092150170649, + "grad_norm": 0.373046875, + "learning_rate": 0.00017258070980172494, + "loss": 1.0215, + "step": 1395 + }, + { + "epoch": 4.778156996587031, + "grad_norm": 0.4453125, + "learning_rate": 0.0001723069274238895, + "loss": 1.0249, + "step": 1400 + }, + { + "epoch": 4.795221843003413, + "grad_norm": 0.4921875, + "learning_rate": 0.0001720320044651014, + "loss": 1.0259, + "step": 1405 + }, + { + "epoch": 4.812286689419795, + "grad_norm": 0.380859375, + "learning_rate": 0.00017175594526203905, + "loss": 1.0215, + "step": 1410 + }, + { + "epoch": 4.829351535836177, + "grad_norm": 0.42578125, + "learning_rate": 0.00017147875416930416, + "loss": 1.0272, + "step": 1415 + }, + { + "epoch": 4.84641638225256, + "grad_norm": 0.34765625, + "learning_rate": 0.00017120043555935298, + "loss": 1.0365, + "step": 1420 + }, + { + "epoch": 4.863481228668942, + "grad_norm": 0.36328125, + "learning_rate": 0.00017092099382242748, + "loss": 1.02, + "step": 1425 + }, + { + "epoch": 4.8805460750853245, + "grad_norm": 0.455078125, + "learning_rate": 0.00017064043336648599, + "loss": 1.021, + "step": 1430 + }, + { + "epoch": 4.897610921501706, + "grad_norm": 0.400390625, + "learning_rate": 0.0001703587586171337, + "loss": 1.0156, + "step": 1435 + }, + { + "epoch": 4.914675767918089, + "grad_norm": 0.375, + "learning_rate": 0.00017007597401755276, + "loss": 1.0283, + "step": 1440 + }, + { + "epoch": 4.931740614334471, + "grad_norm": 0.443359375, + "learning_rate": 0.00016979208402843237, + "loss": 1.0194, + "step": 1445 + }, + { + "epoch": 4.948805460750854, + "grad_norm": 0.57421875, + "learning_rate": 0.00016950709312789833, + "loss": 1.0198, + "step": 1450 + }, + { + "epoch": 4.965870307167235, + "grad_norm": 0.37890625, + "learning_rate": 0.00016922100581144228, + "loss": 1.028, + "step": 1455 + }, + { + "epoch": 4.982935153583618, + "grad_norm": 0.4765625, + "learning_rate": 0.00016893382659185105, + "loss": 1.0157, + "step": 1460 + }, + { + "epoch": 5.0, + "grad_norm": 0.416015625, + "learning_rate": 0.00016864555999913518, + "loss": 1.0357, + "step": 1465 + }, + { + "epoch": 5.0, + "eval_loss": 2.468480110168457, + "eval_runtime": 0.549, + "eval_samples_per_second": 18.214, + "eval_steps_per_second": 1.821, + "step": 1465 + }, + { + "epoch": 5.017064846416382, + "grad_norm": 0.380859375, + "learning_rate": 0.0001683562105804577, + "loss": 1.0001, + "step": 1470 + }, + { + "epoch": 5.034129692832765, + "grad_norm": 0.5078125, + "learning_rate": 0.00016806578290006225, + "loss": 0.9998, + "step": 1475 + }, + { + "epoch": 5.051194539249146, + "grad_norm": 0.400390625, + "learning_rate": 0.0001677742815392012, + "loss": 0.9999, + "step": 1480 + }, + { + "epoch": 5.068259385665529, + "grad_norm": 0.43359375, + "learning_rate": 0.00016748171109606328, + "loss": 1.0085, + "step": 1485 + }, + { + "epoch": 5.085324232081911, + "grad_norm": 0.416015625, + "learning_rate": 0.00016718807618570106, + "loss": 1.0018, + "step": 1490 + }, + { + "epoch": 5.102389078498294, + "grad_norm": 0.453125, + "learning_rate": 0.00016689338143995833, + "loss": 0.9997, + "step": 1495 + }, + { + "epoch": 5.1194539249146755, + "grad_norm": 0.4453125, + "learning_rate": 0.00016659763150739677, + "loss": 1.009, + "step": 1500 + }, + { + "epoch": 5.136518771331058, + "grad_norm": 0.357421875, + "learning_rate": 0.00016630083105322266, + "loss": 1.0047, + "step": 1505 + }, + { + "epoch": 5.15358361774744, + "grad_norm": 0.33984375, + "learning_rate": 0.00016600298475921365, + "loss": 1.004, + "step": 1510 + }, + { + "epoch": 5.170648464163823, + "grad_norm": 0.400390625, + "learning_rate": 0.00016570409732364437, + "loss": 1.0022, + "step": 1515 + }, + { + "epoch": 5.187713310580205, + "grad_norm": 0.427734375, + "learning_rate": 0.0001654041734612127, + "loss": 1.0113, + "step": 1520 + }, + { + "epoch": 5.204778156996587, + "grad_norm": 0.3828125, + "learning_rate": 0.00016510321790296525, + "loss": 1.0171, + "step": 1525 + }, + { + "epoch": 5.221843003412969, + "grad_norm": 0.462890625, + "learning_rate": 0.00016480123539622281, + "loss": 1.0146, + "step": 1530 + }, + { + "epoch": 5.238907849829351, + "grad_norm": 0.38671875, + "learning_rate": 0.00016449823070450531, + "loss": 1.0005, + "step": 1535 + }, + { + "epoch": 5.255972696245734, + "grad_norm": 0.3515625, + "learning_rate": 0.00016419420860745699, + "loss": 1.0093, + "step": 1540 + }, + { + "epoch": 5.273037542662116, + "grad_norm": 0.439453125, + "learning_rate": 0.00016388917390077054, + "loss": 0.9965, + "step": 1545 + }, + { + "epoch": 5.290102389078498, + "grad_norm": 0.466796875, + "learning_rate": 0.00016358313139611195, + "loss": 1.0153, + "step": 1550 + }, + { + "epoch": 5.30716723549488, + "grad_norm": 0.376953125, + "learning_rate": 0.0001632760859210442, + "loss": 1.0094, + "step": 1555 + }, + { + "epoch": 5.324232081911263, + "grad_norm": 0.5234375, + "learning_rate": 0.00016296804231895142, + "loss": 0.9984, + "step": 1560 + }, + { + "epoch": 5.341296928327645, + "grad_norm": 0.37109375, + "learning_rate": 0.00016265900544896225, + "loss": 1.0066, + "step": 1565 + }, + { + "epoch": 5.3583617747440275, + "grad_norm": 0.470703125, + "learning_rate": 0.00016234898018587337, + "loss": 1.0027, + "step": 1570 + }, + { + "epoch": 5.375426621160409, + "grad_norm": 0.470703125, + "learning_rate": 0.0001620379714200725, + "loss": 1.014, + "step": 1575 + }, + { + "epoch": 5.392491467576792, + "grad_norm": 0.39453125, + "learning_rate": 0.00016172598405746124, + "loss": 1.0085, + "step": 1580 + }, + { + "epoch": 5.409556313993174, + "grad_norm": 0.51171875, + "learning_rate": 0.00016141302301937786, + "loss": 0.9999, + "step": 1585 + }, + { + "epoch": 5.426621160409557, + "grad_norm": 0.54296875, + "learning_rate": 0.0001610990932425194, + "loss": 1.0199, + "step": 1590 + }, + { + "epoch": 5.4436860068259385, + "grad_norm": 0.447265625, + "learning_rate": 0.00016078419967886402, + "loss": 1.0137, + "step": 1595 + }, + { + "epoch": 5.460750853242321, + "grad_norm": 0.408203125, + "learning_rate": 0.0001604683472955928, + "loss": 1.0057, + "step": 1600 + }, + { + "epoch": 5.477815699658703, + "grad_norm": 0.419921875, + "learning_rate": 0.00016015154107501133, + "loss": 1.0099, + "step": 1605 + }, + { + "epoch": 5.494880546075085, + "grad_norm": 0.455078125, + "learning_rate": 0.00015983378601447127, + "loss": 1.0066, + "step": 1610 + }, + { + "epoch": 5.511945392491468, + "grad_norm": 0.412109375, + "learning_rate": 0.0001595150871262914, + "loss": 1.0134, + "step": 1615 + }, + { + "epoch": 5.5290102389078495, + "grad_norm": 0.37890625, + "learning_rate": 0.00015919544943767856, + "loss": 1.0108, + "step": 1620 + }, + { + "epoch": 5.546075085324232, + "grad_norm": 0.40625, + "learning_rate": 0.00015887487799064838, + "loss": 1.0229, + "step": 1625 + }, + { + "epoch": 5.563139931740614, + "grad_norm": 0.56640625, + "learning_rate": 0.00015855337784194577, + "loss": 1.0126, + "step": 1630 + }, + { + "epoch": 5.580204778156997, + "grad_norm": 0.37890625, + "learning_rate": 0.00015823095406296514, + "loss": 0.9947, + "step": 1635 + }, + { + "epoch": 5.597269624573379, + "grad_norm": 0.373046875, + "learning_rate": 0.00015790761173967036, + "loss": 1.0063, + "step": 1640 + }, + { + "epoch": 5.614334470989761, + "grad_norm": 0.416015625, + "learning_rate": 0.00015758335597251458, + "loss": 1.0132, + "step": 1645 + }, + { + "epoch": 5.631399317406143, + "grad_norm": 0.4375, + "learning_rate": 0.00015725819187635968, + "loss": 1.0173, + "step": 1650 + }, + { + "epoch": 5.648464163822526, + "grad_norm": 0.388671875, + "learning_rate": 0.00015693212458039584, + "loss": 1.0115, + "step": 1655 + }, + { + "epoch": 5.665529010238908, + "grad_norm": 0.42578125, + "learning_rate": 0.00015660515922806027, + "loss": 0.9966, + "step": 1660 + }, + { + "epoch": 5.6825938566552905, + "grad_norm": 0.349609375, + "learning_rate": 0.00015627730097695638, + "loss": 1.0058, + "step": 1665 + }, + { + "epoch": 5.699658703071672, + "grad_norm": 0.427734375, + "learning_rate": 0.0001559485549987723, + "loss": 1.0143, + "step": 1670 + }, + { + "epoch": 5.716723549488055, + "grad_norm": 0.384765625, + "learning_rate": 0.0001556189264791992, + "loss": 1.0124, + "step": 1675 + }, + { + "epoch": 5.733788395904437, + "grad_norm": 0.40625, + "learning_rate": 0.0001552884206178498, + "loss": 1.0119, + "step": 1680 + }, + { + "epoch": 5.750853242320819, + "grad_norm": 0.412109375, + "learning_rate": 0.00015495704262817597, + "loss": 1.0061, + "step": 1685 + }, + { + "epoch": 5.7679180887372015, + "grad_norm": 0.3828125, + "learning_rate": 0.0001546247977373867, + "loss": 1.0054, + "step": 1690 + }, + { + "epoch": 5.784982935153583, + "grad_norm": 0.37109375, + "learning_rate": 0.00015429169118636566, + "loss": 1.0021, + "step": 1695 + }, + { + "epoch": 5.802047781569966, + "grad_norm": 0.392578125, + "learning_rate": 0.00015395772822958845, + "loss": 1.0037, + "step": 1700 + }, + { + "epoch": 5.819112627986348, + "grad_norm": 0.408203125, + "learning_rate": 0.00015362291413503984, + "loss": 1.0054, + "step": 1705 + }, + { + "epoch": 5.836177474402731, + "grad_norm": 0.345703125, + "learning_rate": 0.00015328725418413045, + "loss": 1.0132, + "step": 1710 + }, + { + "epoch": 5.853242320819112, + "grad_norm": 0.341796875, + "learning_rate": 0.00015295075367161367, + "loss": 1.0041, + "step": 1715 + }, + { + "epoch": 5.870307167235495, + "grad_norm": 0.34375, + "learning_rate": 0.00015261341790550196, + "loss": 1.001, + "step": 1720 + }, + { + "epoch": 5.887372013651877, + "grad_norm": 0.373046875, + "learning_rate": 0.0001522752522069833, + "loss": 1.0102, + "step": 1725 + }, + { + "epoch": 5.90443686006826, + "grad_norm": 0.38671875, + "learning_rate": 0.00015193626191033712, + "loss": 0.996, + "step": 1730 + }, + { + "epoch": 5.921501706484642, + "grad_norm": 0.37109375, + "learning_rate": 0.0001515964523628501, + "loss": 1.0052, + "step": 1735 + }, + { + "epoch": 5.938566552901024, + "grad_norm": 0.3984375, + "learning_rate": 0.00015125582892473204, + "loss": 1.0118, + "step": 1740 + }, + { + "epoch": 5.955631399317406, + "grad_norm": 0.392578125, + "learning_rate": 0.00015091439696903115, + "loss": 0.998, + "step": 1745 + }, + { + "epoch": 5.972696245733788, + "grad_norm": 0.388671875, + "learning_rate": 0.00015057216188154928, + "loss": 0.9925, + "step": 1750 + }, + { + "epoch": 5.989761092150171, + "grad_norm": 0.5, + "learning_rate": 0.00015022912906075702, + "loss": 0.993, + "step": 1755 + }, + { + "epoch": 6.0, + "eval_loss": 2.4702811241149902, + "eval_runtime": 0.5473, + "eval_samples_per_second": 18.272, + "eval_steps_per_second": 1.827, + "step": 1758 + }, + { + "epoch": 6.006825938566553, + "grad_norm": 0.48046875, + "learning_rate": 0.00014988530391770856, + "loss": 0.9939, + "step": 1760 + }, + { + "epoch": 6.023890784982935, + "grad_norm": 0.396484375, + "learning_rate": 0.00014954069187595633, + "loss": 0.9904, + "step": 1765 + }, + { + "epoch": 6.040955631399317, + "grad_norm": 0.396484375, + "learning_rate": 0.00014919529837146528, + "loss": 0.982, + "step": 1770 + }, + { + "epoch": 6.0580204778157, + "grad_norm": 0.421875, + "learning_rate": 0.0001488491288525275, + "loss": 0.9741, + "step": 1775 + }, + { + "epoch": 6.075085324232082, + "grad_norm": 0.439453125, + "learning_rate": 0.0001485021887796759, + "loss": 0.995, + "step": 1780 + }, + { + "epoch": 6.092150170648464, + "grad_norm": 0.39453125, + "learning_rate": 0.00014815448362559826, + "loss": 0.9931, + "step": 1785 + }, + { + "epoch": 6.109215017064846, + "grad_norm": 0.373046875, + "learning_rate": 0.00014780601887505088, + "loss": 1.0001, + "step": 1790 + }, + { + "epoch": 6.126279863481229, + "grad_norm": 0.384765625, + "learning_rate": 0.00014745680002477203, + "loss": 0.9913, + "step": 1795 + }, + { + "epoch": 6.143344709897611, + "grad_norm": 0.36328125, + "learning_rate": 0.00014710683258339536, + "loss": 0.9883, + "step": 1800 + }, + { + "epoch": 6.160409556313994, + "grad_norm": 0.427734375, + "learning_rate": 0.0001467561220713628, + "loss": 0.9835, + "step": 1805 + }, + { + "epoch": 6.177474402730375, + "grad_norm": 0.421875, + "learning_rate": 0.0001464046740208377, + "loss": 0.9894, + "step": 1810 + }, + { + "epoch": 6.194539249146757, + "grad_norm": 0.357421875, + "learning_rate": 0.00014605249397561736, + "loss": 0.9833, + "step": 1815 + }, + { + "epoch": 6.21160409556314, + "grad_norm": 0.4140625, + "learning_rate": 0.00014569958749104575, + "loss": 0.9942, + "step": 1820 + }, + { + "epoch": 6.228668941979522, + "grad_norm": 0.58203125, + "learning_rate": 0.00014534596013392575, + "loss": 0.9937, + "step": 1825 + }, + { + "epoch": 6.2457337883959045, + "grad_norm": 0.6328125, + "learning_rate": 0.00014499161748243147, + "loss": 0.9852, + "step": 1830 + }, + { + "epoch": 6.262798634812286, + "grad_norm": 0.6640625, + "learning_rate": 0.0001446365651260201, + "loss": 0.9886, + "step": 1835 + }, + { + "epoch": 6.279863481228669, + "grad_norm": 0.4375, + "learning_rate": 0.00014428080866534396, + "loss": 0.9893, + "step": 1840 + }, + { + "epoch": 6.296928327645051, + "grad_norm": 0.376953125, + "learning_rate": 0.00014392435371216185, + "loss": 0.9951, + "step": 1845 + }, + { + "epoch": 6.313993174061434, + "grad_norm": 0.34765625, + "learning_rate": 0.0001435672058892509, + "loss": 0.9877, + "step": 1850 + }, + { + "epoch": 6.3310580204778155, + "grad_norm": 0.390625, + "learning_rate": 0.00014320937083031748, + "loss": 0.9922, + "step": 1855 + }, + { + "epoch": 6.348122866894198, + "grad_norm": 0.38671875, + "learning_rate": 0.0001428508541799086, + "loss": 0.9939, + "step": 1860 + }, + { + "epoch": 6.36518771331058, + "grad_norm": 0.3828125, + "learning_rate": 0.0001424916615933229, + "loss": 0.994, + "step": 1865 + }, + { + "epoch": 6.382252559726963, + "grad_norm": 0.404296875, + "learning_rate": 0.00014213179873652127, + "loss": 0.993, + "step": 1870 + }, + { + "epoch": 6.399317406143345, + "grad_norm": 0.408203125, + "learning_rate": 0.00014177127128603745, + "loss": 0.9982, + "step": 1875 + }, + { + "epoch": 6.4163822525597265, + "grad_norm": 0.40625, + "learning_rate": 0.0001414100849288888, + "loss": 0.9926, + "step": 1880 + }, + { + "epoch": 6.433447098976109, + "grad_norm": 0.416015625, + "learning_rate": 0.00014104824536248614, + "loss": 0.995, + "step": 1885 + }, + { + "epoch": 6.450511945392491, + "grad_norm": 0.40625, + "learning_rate": 0.00014068575829454436, + "loss": 0.9894, + "step": 1890 + }, + { + "epoch": 6.467576791808874, + "grad_norm": 0.359375, + "learning_rate": 0.00014032262944299194, + "loss": 0.997, + "step": 1895 + }, + { + "epoch": 6.484641638225256, + "grad_norm": 0.392578125, + "learning_rate": 0.00013995886453588104, + "loss": 0.9861, + "step": 1900 + }, + { + "epoch": 6.501706484641638, + "grad_norm": 0.34765625, + "learning_rate": 0.00013959446931129704, + "loss": 0.9896, + "step": 1905 + }, + { + "epoch": 6.51877133105802, + "grad_norm": 0.380859375, + "learning_rate": 0.0001392294495172681, + "loss": 0.9969, + "step": 1910 + }, + { + "epoch": 6.535836177474403, + "grad_norm": 0.392578125, + "learning_rate": 0.0001388638109116744, + "loss": 0.9902, + "step": 1915 + }, + { + "epoch": 6.552901023890785, + "grad_norm": 0.375, + "learning_rate": 0.00013849755926215735, + "loss": 0.9995, + "step": 1920 + }, + { + "epoch": 6.5699658703071675, + "grad_norm": 0.384765625, + "learning_rate": 0.00013813070034602863, + "loss": 0.9935, + "step": 1925 + }, + { + "epoch": 6.587030716723549, + "grad_norm": 0.466796875, + "learning_rate": 0.00013776323995017898, + "loss": 0.9799, + "step": 1930 + }, + { + "epoch": 6.604095563139932, + "grad_norm": 0.345703125, + "learning_rate": 0.00013739518387098705, + "loss": 0.9959, + "step": 1935 + }, + { + "epoch": 6.621160409556314, + "grad_norm": 0.388671875, + "learning_rate": 0.0001370265379142279, + "loss": 0.9897, + "step": 1940 + }, + { + "epoch": 6.638225255972696, + "grad_norm": 0.443359375, + "learning_rate": 0.0001366573078949813, + "loss": 0.9829, + "step": 1945 + }, + { + "epoch": 6.6552901023890785, + "grad_norm": 0.44921875, + "learning_rate": 0.00013628749963754026, + "loss": 0.9963, + "step": 1950 + }, + { + "epoch": 6.672354948805461, + "grad_norm": 0.52734375, + "learning_rate": 0.0001359171189753189, + "loss": 0.999, + "step": 1955 + }, + { + "epoch": 6.689419795221843, + "grad_norm": 0.6484375, + "learning_rate": 0.00013554617175076062, + "loss": 0.9806, + "step": 1960 + }, + { + "epoch": 6.706484641638225, + "grad_norm": 0.388671875, + "learning_rate": 0.0001351746638152458, + "loss": 0.9903, + "step": 1965 + }, + { + "epoch": 6.723549488054608, + "grad_norm": 0.4765625, + "learning_rate": 0.00013480260102899966, + "loss": 1.0009, + "step": 1970 + }, + { + "epoch": 6.7406143344709895, + "grad_norm": 0.44140625, + "learning_rate": 0.0001344299892609996, + "loss": 0.9879, + "step": 1975 + }, + { + "epoch": 6.757679180887372, + "grad_norm": 0.392578125, + "learning_rate": 0.00013405683438888282, + "loss": 0.9966, + "step": 1980 + }, + { + "epoch": 6.774744027303754, + "grad_norm": 0.4140625, + "learning_rate": 0.00013368314229885347, + "loss": 0.988, + "step": 1985 + }, + { + "epoch": 6.791808873720137, + "grad_norm": 0.36328125, + "learning_rate": 0.00013330891888559002, + "loss": 0.9835, + "step": 1990 + }, + { + "epoch": 6.808873720136519, + "grad_norm": 0.421875, + "learning_rate": 0.00013293417005215188, + "loss": 0.9922, + "step": 1995 + }, + { + "epoch": 6.825938566552901, + "grad_norm": 0.40234375, + "learning_rate": 0.0001325589017098867, + "loss": 0.9893, + "step": 2000 + }, + { + "epoch": 6.843003412969283, + "grad_norm": 0.3828125, + "learning_rate": 0.00013218311977833687, + "loss": 0.9965, + "step": 2005 + }, + { + "epoch": 6.860068259385666, + "grad_norm": 0.365234375, + "learning_rate": 0.0001318068301851463, + "loss": 0.9843, + "step": 2010 + }, + { + "epoch": 6.877133105802048, + "grad_norm": 0.390625, + "learning_rate": 0.00013143003886596669, + "loss": 0.9845, + "step": 2015 + }, + { + "epoch": 6.8941979522184305, + "grad_norm": 0.3515625, + "learning_rate": 0.0001310527517643642, + "loss": 0.9909, + "step": 2020 + }, + { + "epoch": 6.911262798634812, + "grad_norm": 0.359375, + "learning_rate": 0.00013067497483172538, + "loss": 0.9885, + "step": 2025 + }, + { + "epoch": 6.928327645051194, + "grad_norm": 0.375, + "learning_rate": 0.00013029671402716366, + "loss": 0.9879, + "step": 2030 + }, + { + "epoch": 6.945392491467577, + "grad_norm": 0.380859375, + "learning_rate": 0.00012991797531742492, + "loss": 0.9891, + "step": 2035 + }, + { + "epoch": 6.962457337883959, + "grad_norm": 0.34375, + "learning_rate": 0.00012953876467679373, + "loss": 0.9972, + "step": 2040 + }, + { + "epoch": 6.979522184300341, + "grad_norm": 0.369140625, + "learning_rate": 0.00012915908808699893, + "loss": 0.9962, + "step": 2045 + }, + { + "epoch": 6.996587030716723, + "grad_norm": 0.44921875, + "learning_rate": 0.00012877895153711935, + "loss": 0.9941, + "step": 2050 + }, + { + "epoch": 7.0, + "eval_loss": 2.49063777923584, + "eval_runtime": 0.554, + "eval_samples_per_second": 18.051, + "eval_steps_per_second": 1.805, + "step": 2051 + }, + { + "epoch": 7.013651877133106, + "grad_norm": 0.52734375, + "learning_rate": 0.00012839836102348926, + "loss": 0.9759, + "step": 2055 + }, + { + "epoch": 7.030716723549488, + "grad_norm": 0.365234375, + "learning_rate": 0.00012801732254960388, + "loss": 0.9703, + "step": 2060 + }, + { + "epoch": 7.047781569965871, + "grad_norm": 0.375, + "learning_rate": 0.00012763584212602453, + "loss": 0.9643, + "step": 2065 + }, + { + "epoch": 7.064846416382252, + "grad_norm": 0.41796875, + "learning_rate": 0.00012725392577028402, + "loss": 0.9646, + "step": 2070 + }, + { + "epoch": 7.081911262798635, + "grad_norm": 0.400390625, + "learning_rate": 0.0001268715795067916, + "loss": 0.9732, + "step": 2075 + }, + { + "epoch": 7.098976109215017, + "grad_norm": 0.380859375, + "learning_rate": 0.00012648880936673787, + "loss": 0.9786, + "step": 2080 + }, + { + "epoch": 7.1160409556314, + "grad_norm": 0.423828125, + "learning_rate": 0.00012610562138799978, + "loss": 0.9733, + "step": 2085 + }, + { + "epoch": 7.1331058020477816, + "grad_norm": 0.357421875, + "learning_rate": 0.00012572202161504543, + "loss": 0.9808, + "step": 2090 + }, + { + "epoch": 7.150170648464163, + "grad_norm": 0.4609375, + "learning_rate": 0.00012533801609883842, + "loss": 0.9762, + "step": 2095 + }, + { + "epoch": 7.167235494880546, + "grad_norm": 0.38671875, + "learning_rate": 0.00012495361089674285, + "loss": 0.9809, + "step": 2100 + }, + { + "epoch": 7.184300341296928, + "grad_norm": 0.3984375, + "learning_rate": 0.00012456881207242732, + "loss": 0.9821, + "step": 2105 + }, + { + "epoch": 7.201365187713311, + "grad_norm": 0.400390625, + "learning_rate": 0.00012418362569576965, + "loss": 0.9873, + "step": 2110 + }, + { + "epoch": 7.2184300341296925, + "grad_norm": 0.55078125, + "learning_rate": 0.00012379805784276082, + "loss": 0.9727, + "step": 2115 + }, + { + "epoch": 7.235494880546075, + "grad_norm": 0.515625, + "learning_rate": 0.0001234121145954094, + "loss": 0.9827, + "step": 2120 + }, + { + "epoch": 7.252559726962457, + "grad_norm": 0.3828125, + "learning_rate": 0.00012302580204164541, + "loss": 0.9846, + "step": 2125 + }, + { + "epoch": 7.26962457337884, + "grad_norm": 0.42578125, + "learning_rate": 0.0001226391262752245, + "loss": 0.9736, + "step": 2130 + }, + { + "epoch": 7.286689419795222, + "grad_norm": 0.5078125, + "learning_rate": 0.00012225209339563145, + "loss": 0.9743, + "step": 2135 + }, + { + "epoch": 7.303754266211604, + "grad_norm": 0.419921875, + "learning_rate": 0.00012186470950798445, + "loss": 0.9787, + "step": 2140 + }, + { + "epoch": 7.320819112627986, + "grad_norm": 0.490234375, + "learning_rate": 0.00012147698072293842, + "loss": 0.9788, + "step": 2145 + }, + { + "epoch": 7.337883959044369, + "grad_norm": 0.380859375, + "learning_rate": 0.00012108891315658879, + "loss": 0.967, + "step": 2150 + }, + { + "epoch": 7.354948805460751, + "grad_norm": 0.396484375, + "learning_rate": 0.00012070051293037492, + "loss": 0.9792, + "step": 2155 + }, + { + "epoch": 7.372013651877133, + "grad_norm": 0.43359375, + "learning_rate": 0.00012031178617098371, + "loss": 0.9905, + "step": 2160 + }, + { + "epoch": 7.389078498293515, + "grad_norm": 0.400390625, + "learning_rate": 0.00011992273901025269, + "loss": 0.9873, + "step": 2165 + }, + { + "epoch": 7.406143344709897, + "grad_norm": 0.4453125, + "learning_rate": 0.0001195333775850736, + "loss": 0.9872, + "step": 2170 + }, + { + "epoch": 7.42320819112628, + "grad_norm": 0.44140625, + "learning_rate": 0.00011914370803729533, + "loss": 0.98, + "step": 2175 + }, + { + "epoch": 7.440273037542662, + "grad_norm": 0.361328125, + "learning_rate": 0.00011875373651362727, + "loss": 0.9827, + "step": 2180 + }, + { + "epoch": 7.4573378839590445, + "grad_norm": 0.474609375, + "learning_rate": 0.00011836346916554205, + "loss": 0.9738, + "step": 2185 + }, + { + "epoch": 7.474402730375426, + "grad_norm": 0.421875, + "learning_rate": 0.00011797291214917881, + "loss": 0.9762, + "step": 2190 + }, + { + "epoch": 7.491467576791809, + "grad_norm": 0.41796875, + "learning_rate": 0.00011758207162524598, + "loss": 0.9675, + "step": 2195 + }, + { + "epoch": 7.508532423208191, + "grad_norm": 0.384765625, + "learning_rate": 0.00011719095375892396, + "loss": 0.9923, + "step": 2200 + }, + { + "epoch": 7.525597269624574, + "grad_norm": 0.482421875, + "learning_rate": 0.00011679956471976814, + "loss": 0.9818, + "step": 2205 + }, + { + "epoch": 7.5426621160409555, + "grad_norm": 0.3671875, + "learning_rate": 0.0001164079106816113, + "loss": 0.9783, + "step": 2210 + }, + { + "epoch": 7.559726962457338, + "grad_norm": 0.376953125, + "learning_rate": 0.00011601599782246646, + "loss": 0.9735, + "step": 2215 + }, + { + "epoch": 7.57679180887372, + "grad_norm": 0.443359375, + "learning_rate": 0.00011562383232442926, + "loss": 0.9751, + "step": 2220 + }, + { + "epoch": 7.593856655290102, + "grad_norm": 0.3671875, + "learning_rate": 0.0001152314203735805, + "loss": 0.9734, + "step": 2225 + }, + { + "epoch": 7.610921501706485, + "grad_norm": 0.439453125, + "learning_rate": 0.00011483876815988867, + "loss": 0.9706, + "step": 2230 + }, + { + "epoch": 7.627986348122867, + "grad_norm": 0.44140625, + "learning_rate": 0.00011444588187711205, + "loss": 0.9727, + "step": 2235 + }, + { + "epoch": 7.645051194539249, + "grad_norm": 0.41796875, + "learning_rate": 0.00011405276772270126, + "loss": 0.9774, + "step": 2240 + }, + { + "epoch": 7.662116040955631, + "grad_norm": 0.353515625, + "learning_rate": 0.0001136594318977014, + "loss": 0.9815, + "step": 2245 + }, + { + "epoch": 7.679180887372014, + "grad_norm": 0.412109375, + "learning_rate": 0.0001132658806066542, + "loss": 0.9835, + "step": 2250 + }, + { + "epoch": 7.696245733788396, + "grad_norm": 0.384765625, + "learning_rate": 0.00011287212005750024, + "loss": 0.9773, + "step": 2255 + }, + { + "epoch": 7.713310580204778, + "grad_norm": 0.42578125, + "learning_rate": 0.00011247815646148087, + "loss": 0.9835, + "step": 2260 + }, + { + "epoch": 7.73037542662116, + "grad_norm": 0.56640625, + "learning_rate": 0.00011208399603304047, + "loss": 0.9832, + "step": 2265 + }, + { + "epoch": 7.747440273037543, + "grad_norm": 0.38671875, + "learning_rate": 0.00011168964498972818, + "loss": 0.9701, + "step": 2270 + }, + { + "epoch": 7.764505119453925, + "grad_norm": 0.3671875, + "learning_rate": 0.00011129510955209996, + "loss": 0.9832, + "step": 2275 + }, + { + "epoch": 7.7815699658703075, + "grad_norm": 0.546875, + "learning_rate": 0.00011090039594362045, + "loss": 0.9861, + "step": 2280 + }, + { + "epoch": 7.798634812286689, + "grad_norm": 0.5078125, + "learning_rate": 0.00011050551039056479, + "loss": 0.9881, + "step": 2285 + }, + { + "epoch": 7.815699658703072, + "grad_norm": 0.375, + "learning_rate": 0.00011011045912192035, + "loss": 0.9872, + "step": 2290 + }, + { + "epoch": 7.832764505119454, + "grad_norm": 0.373046875, + "learning_rate": 0.0001097152483692886, + "loss": 0.9819, + "step": 2295 + }, + { + "epoch": 7.849829351535837, + "grad_norm": 0.375, + "learning_rate": 0.00010931988436678666, + "loss": 0.9756, + "step": 2300 + }, + { + "epoch": 7.8668941979522184, + "grad_norm": 0.40234375, + "learning_rate": 0.00010892437335094912, + "loss": 0.9662, + "step": 2305 + }, + { + "epoch": 7.8839590443686, + "grad_norm": 0.427734375, + "learning_rate": 0.00010852872156062946, + "loss": 0.9669, + "step": 2310 + }, + { + "epoch": 7.901023890784983, + "grad_norm": 0.388671875, + "learning_rate": 0.00010813293523690191, + "loss": 0.9755, + "step": 2315 + }, + { + "epoch": 7.918088737201365, + "grad_norm": 0.423828125, + "learning_rate": 0.00010773702062296273, + "loss": 0.9916, + "step": 2320 + }, + { + "epoch": 7.935153583617748, + "grad_norm": 0.396484375, + "learning_rate": 0.00010734098396403192, + "loss": 0.9869, + "step": 2325 + }, + { + "epoch": 7.952218430034129, + "grad_norm": 0.447265625, + "learning_rate": 0.00010694483150725458, + "loss": 0.978, + "step": 2330 + }, + { + "epoch": 7.969283276450512, + "grad_norm": 0.451171875, + "learning_rate": 0.00010654856950160253, + "loss": 0.9711, + "step": 2335 + }, + { + "epoch": 7.986348122866894, + "grad_norm": 0.392578125, + "learning_rate": 0.00010615220419777548, + "loss": 0.9844, + "step": 2340 + }, + { + "epoch": 8.0, + "eval_loss": 2.489572525024414, + "eval_runtime": 0.5472, + "eval_samples_per_second": 18.276, + "eval_steps_per_second": 1.828, + "step": 2344 + }, + { + "epoch": 8.003412969283277, + "grad_norm": 0.44140625, + "learning_rate": 0.00010575574184810269, + "loss": 0.9713, + "step": 2345 + }, + { + "epoch": 8.020477815699659, + "grad_norm": 0.3984375, + "learning_rate": 0.0001053591887064442, + "loss": 0.9647, + "step": 2350 + }, + { + "epoch": 8.03754266211604, + "grad_norm": 0.453125, + "learning_rate": 0.00010496255102809223, + "loss": 0.9709, + "step": 2355 + }, + { + "epoch": 8.054607508532424, + "grad_norm": 0.431640625, + "learning_rate": 0.00010456583506967248, + "loss": 0.9701, + "step": 2360 + }, + { + "epoch": 8.071672354948806, + "grad_norm": 0.44921875, + "learning_rate": 0.00010416904708904548, + "loss": 0.9662, + "step": 2365 + }, + { + "epoch": 8.088737201365188, + "grad_norm": 0.46875, + "learning_rate": 0.00010377219334520783, + "loss": 0.9616, + "step": 2370 + }, + { + "epoch": 8.10580204778157, + "grad_norm": 0.4140625, + "learning_rate": 0.00010337528009819344, + "loss": 0.9609, + "step": 2375 + }, + { + "epoch": 8.122866894197951, + "grad_norm": 0.42578125, + "learning_rate": 0.00010297831360897492, + "loss": 0.9714, + "step": 2380 + }, + { + "epoch": 8.139931740614335, + "grad_norm": 0.40234375, + "learning_rate": 0.00010258130013936474, + "loss": 0.9718, + "step": 2385 + }, + { + "epoch": 8.156996587030717, + "grad_norm": 0.4296875, + "learning_rate": 0.00010218424595191631, + "loss": 0.963, + "step": 2390 + }, + { + "epoch": 8.174061433447099, + "grad_norm": 0.361328125, + "learning_rate": 0.00010178715730982549, + "loss": 0.9612, + "step": 2395 + }, + { + "epoch": 8.19112627986348, + "grad_norm": 0.451171875, + "learning_rate": 0.00010139004047683151, + "loss": 0.9757, + "step": 2400 + }, + { + "epoch": 8.208191126279864, + "grad_norm": 0.62890625, + "learning_rate": 0.00010099290171711841, + "loss": 0.961, + "step": 2405 + }, + { + "epoch": 8.225255972696246, + "grad_norm": 0.419921875, + "learning_rate": 0.00010059574729521595, + "loss": 0.962, + "step": 2410 + }, + { + "epoch": 8.242320819112628, + "grad_norm": 0.51171875, + "learning_rate": 0.0001001985834759011, + "loss": 0.9761, + "step": 2415 + }, + { + "epoch": 8.25938566552901, + "grad_norm": 0.390625, + "learning_rate": 9.980141652409895e-05, + "loss": 0.9718, + "step": 2420 + }, + { + "epoch": 8.276450511945393, + "grad_norm": 0.41796875, + "learning_rate": 9.940425270478407e-05, + "loss": 0.9672, + "step": 2425 + }, + { + "epoch": 8.293515358361775, + "grad_norm": 0.431640625, + "learning_rate": 9.900709828288164e-05, + "loss": 0.9658, + "step": 2430 + }, + { + "epoch": 8.310580204778157, + "grad_norm": 0.4140625, + "learning_rate": 9.860995952316851e-05, + "loss": 0.9776, + "step": 2435 + }, + { + "epoch": 8.327645051194539, + "grad_norm": 0.37890625, + "learning_rate": 9.821284269017455e-05, + "loss": 0.9664, + "step": 2440 + }, + { + "epoch": 8.344709897610922, + "grad_norm": 0.380859375, + "learning_rate": 9.781575404808371e-05, + "loss": 0.9672, + "step": 2445 + }, + { + "epoch": 8.361774744027304, + "grad_norm": 0.3828125, + "learning_rate": 9.741869986063526e-05, + "loss": 0.9778, + "step": 2450 + }, + { + "epoch": 8.378839590443686, + "grad_norm": 0.361328125, + "learning_rate": 9.702168639102509e-05, + "loss": 0.9659, + "step": 2455 + }, + { + "epoch": 8.395904436860068, + "grad_norm": 0.392578125, + "learning_rate": 9.662471990180657e-05, + "loss": 0.9623, + "step": 2460 + }, + { + "epoch": 8.41296928327645, + "grad_norm": 0.365234375, + "learning_rate": 9.622780665479222e-05, + "loss": 0.9657, + "step": 2465 + }, + { + "epoch": 8.430034129692833, + "grad_norm": 0.40234375, + "learning_rate": 9.583095291095453e-05, + "loss": 0.9679, + "step": 2470 + }, + { + "epoch": 8.447098976109215, + "grad_norm": 0.443359375, + "learning_rate": 9.543416493032757e-05, + "loss": 0.9686, + "step": 2475 + }, + { + "epoch": 8.464163822525597, + "grad_norm": 0.404296875, + "learning_rate": 9.503744897190778e-05, + "loss": 0.9679, + "step": 2480 + }, + { + "epoch": 8.481228668941979, + "grad_norm": 0.396484375, + "learning_rate": 9.464081129355586e-05, + "loss": 0.9588, + "step": 2485 + }, + { + "epoch": 8.498293515358363, + "grad_norm": 0.431640625, + "learning_rate": 9.424425815189733e-05, + "loss": 0.9775, + "step": 2490 + }, + { + "epoch": 8.515358361774744, + "grad_norm": 0.384765625, + "learning_rate": 9.384779580222453e-05, + "loss": 0.9668, + "step": 2495 + }, + { + "epoch": 8.532423208191126, + "grad_norm": 0.447265625, + "learning_rate": 9.345143049839749e-05, + "loss": 0.9677, + "step": 2500 + }, + { + "epoch": 8.549488054607508, + "grad_norm": 0.48828125, + "learning_rate": 9.305516849274541e-05, + "loss": 0.9603, + "step": 2505 + }, + { + "epoch": 8.56655290102389, + "grad_norm": 0.427734375, + "learning_rate": 9.265901603596811e-05, + "loss": 0.9688, + "step": 2510 + }, + { + "epoch": 8.583617747440274, + "grad_norm": 0.498046875, + "learning_rate": 9.226297937703728e-05, + "loss": 0.9645, + "step": 2515 + }, + { + "epoch": 8.600682593856655, + "grad_norm": 0.431640625, + "learning_rate": 9.186706476309812e-05, + "loss": 0.967, + "step": 2520 + }, + { + "epoch": 8.617747440273037, + "grad_norm": 0.423828125, + "learning_rate": 9.147127843937055e-05, + "loss": 0.9711, + "step": 2525 + }, + { + "epoch": 8.634812286689419, + "grad_norm": 0.455078125, + "learning_rate": 9.107562664905093e-05, + "loss": 0.971, + "step": 2530 + }, + { + "epoch": 8.651877133105803, + "grad_norm": 0.484375, + "learning_rate": 9.068011563321336e-05, + "loss": 0.9722, + "step": 2535 + }, + { + "epoch": 8.668941979522184, + "grad_norm": 0.435546875, + "learning_rate": 9.028475163071141e-05, + "loss": 0.9747, + "step": 2540 + }, + { + "epoch": 8.686006825938566, + "grad_norm": 0.4140625, + "learning_rate": 8.988954087807968e-05, + "loss": 0.9638, + "step": 2545 + }, + { + "epoch": 8.703071672354948, + "grad_norm": 0.400390625, + "learning_rate": 8.949448960943524e-05, + "loss": 0.9625, + "step": 2550 + }, + { + "epoch": 8.720136518771332, + "grad_norm": 0.49609375, + "learning_rate": 8.909960405637958e-05, + "loss": 0.9568, + "step": 2555 + }, + { + "epoch": 8.737201365187714, + "grad_norm": 0.435546875, + "learning_rate": 8.870489044790006e-05, + "loss": 0.9766, + "step": 2560 + }, + { + "epoch": 8.754266211604095, + "grad_norm": 0.41015625, + "learning_rate": 8.831035501027186e-05, + "loss": 0.967, + "step": 2565 + }, + { + "epoch": 8.771331058020477, + "grad_norm": 0.376953125, + "learning_rate": 8.791600396695954e-05, + "loss": 0.9686, + "step": 2570 + }, + { + "epoch": 8.788395904436861, + "grad_norm": 0.373046875, + "learning_rate": 8.752184353851916e-05, + "loss": 0.9684, + "step": 2575 + }, + { + "epoch": 8.805460750853243, + "grad_norm": 0.435546875, + "learning_rate": 8.712787994249979e-05, + "loss": 0.977, + "step": 2580 + }, + { + "epoch": 8.822525597269625, + "grad_norm": 0.419921875, + "learning_rate": 8.673411939334581e-05, + "loss": 0.9712, + "step": 2585 + }, + { + "epoch": 8.839590443686006, + "grad_norm": 0.478515625, + "learning_rate": 8.634056810229862e-05, + "loss": 0.9692, + "step": 2590 + }, + { + "epoch": 8.856655290102388, + "grad_norm": 0.404296875, + "learning_rate": 8.594723227729875e-05, + "loss": 0.9639, + "step": 2595 + }, + { + "epoch": 8.873720136518772, + "grad_norm": 0.447265625, + "learning_rate": 8.555411812288798e-05, + "loss": 0.974, + "step": 2600 + }, + { + "epoch": 8.890784982935154, + "grad_norm": 0.392578125, + "learning_rate": 8.516123184011135e-05, + "loss": 0.9589, + "step": 2605 + }, + { + "epoch": 8.907849829351536, + "grad_norm": 0.43359375, + "learning_rate": 8.47685796264195e-05, + "loss": 0.968, + "step": 2610 + }, + { + "epoch": 8.924914675767917, + "grad_norm": 0.396484375, + "learning_rate": 8.437616767557077e-05, + "loss": 0.9693, + "step": 2615 + }, + { + "epoch": 8.941979522184301, + "grad_norm": 0.5390625, + "learning_rate": 8.398400217753357e-05, + "loss": 0.9727, + "step": 2620 + }, + { + "epoch": 8.959044368600683, + "grad_norm": 0.419921875, + "learning_rate": 8.359208931838871e-05, + "loss": 0.9708, + "step": 2625 + }, + { + "epoch": 8.976109215017065, + "grad_norm": 0.427734375, + "learning_rate": 8.320043528023188e-05, + "loss": 0.9607, + "step": 2630 + }, + { + "epoch": 8.993174061433447, + "grad_norm": 0.455078125, + "learning_rate": 8.280904624107606e-05, + "loss": 0.9779, + "step": 2635 + }, + { + "epoch": 9.0, + "eval_loss": 2.502519130706787, + "eval_runtime": 0.5483, + "eval_samples_per_second": 18.238, + "eval_steps_per_second": 1.824, + "step": 2637 + }, + { + "epoch": 9.01023890784983, + "grad_norm": 0.3828125, + "learning_rate": 8.241792837475405e-05, + "loss": 0.9673, + "step": 2640 + }, + { + "epoch": 9.027303754266212, + "grad_norm": 0.42578125, + "learning_rate": 8.202708785082121e-05, + "loss": 0.9481, + "step": 2645 + }, + { + "epoch": 9.044368600682594, + "grad_norm": 0.39453125, + "learning_rate": 8.163653083445799e-05, + "loss": 0.9694, + "step": 2650 + }, + { + "epoch": 9.061433447098976, + "grad_norm": 0.392578125, + "learning_rate": 8.124626348637279e-05, + "loss": 0.9651, + "step": 2655 + }, + { + "epoch": 9.078498293515358, + "grad_norm": 0.376953125, + "learning_rate": 8.085629196270469e-05, + "loss": 0.9561, + "step": 2660 + }, + { + "epoch": 9.095563139931741, + "grad_norm": 0.408203125, + "learning_rate": 8.046662241492645e-05, + "loss": 0.9617, + "step": 2665 + }, + { + "epoch": 9.112627986348123, + "grad_norm": 0.408203125, + "learning_rate": 8.007726098974734e-05, + "loss": 0.9636, + "step": 2670 + }, + { + "epoch": 9.129692832764505, + "grad_norm": 0.390625, + "learning_rate": 7.96882138290163e-05, + "loss": 0.9661, + "step": 2675 + }, + { + "epoch": 9.146757679180887, + "grad_norm": 0.396484375, + "learning_rate": 7.929948706962508e-05, + "loss": 0.9577, + "step": 2680 + }, + { + "epoch": 9.16382252559727, + "grad_norm": 0.41796875, + "learning_rate": 7.891108684341121e-05, + "loss": 0.961, + "step": 2685 + }, + { + "epoch": 9.180887372013652, + "grad_norm": 0.37109375, + "learning_rate": 7.852301927706159e-05, + "loss": 0.9602, + "step": 2690 + }, + { + "epoch": 9.197952218430034, + "grad_norm": 0.396484375, + "learning_rate": 7.813529049201556e-05, + "loss": 0.9544, + "step": 2695 + }, + { + "epoch": 9.215017064846416, + "grad_norm": 0.470703125, + "learning_rate": 7.774790660436858e-05, + "loss": 0.9569, + "step": 2700 + }, + { + "epoch": 9.2320819112628, + "grad_norm": 0.375, + "learning_rate": 7.736087372477554e-05, + "loss": 0.9636, + "step": 2705 + }, + { + "epoch": 9.249146757679181, + "grad_norm": 0.37109375, + "learning_rate": 7.69741979583546e-05, + "loss": 0.9574, + "step": 2710 + }, + { + "epoch": 9.266211604095563, + "grad_norm": 0.390625, + "learning_rate": 7.658788540459062e-05, + "loss": 0.9536, + "step": 2715 + }, + { + "epoch": 9.283276450511945, + "grad_norm": 0.388671875, + "learning_rate": 7.620194215723919e-05, + "loss": 0.9598, + "step": 2720 + }, + { + "epoch": 9.300341296928327, + "grad_norm": 0.3828125, + "learning_rate": 7.581637430423037e-05, + "loss": 0.9657, + "step": 2725 + }, + { + "epoch": 9.31740614334471, + "grad_norm": 0.435546875, + "learning_rate": 7.543118792757266e-05, + "loss": 0.9639, + "step": 2730 + }, + { + "epoch": 9.334470989761092, + "grad_norm": 0.408203125, + "learning_rate": 7.504638910325717e-05, + "loss": 0.9625, + "step": 2735 + }, + { + "epoch": 9.351535836177474, + "grad_norm": 0.37109375, + "learning_rate": 7.466198390116158e-05, + "loss": 0.9585, + "step": 2740 + }, + { + "epoch": 9.368600682593856, + "grad_norm": 0.447265625, + "learning_rate": 7.427797838495463e-05, + "loss": 0.9634, + "step": 2745 + }, + { + "epoch": 9.38566552901024, + "grad_norm": 0.41796875, + "learning_rate": 7.389437861200024e-05, + "loss": 0.9624, + "step": 2750 + }, + { + "epoch": 9.402730375426621, + "grad_norm": 0.408203125, + "learning_rate": 7.35111906332622e-05, + "loss": 0.9555, + "step": 2755 + }, + { + "epoch": 9.419795221843003, + "grad_norm": 0.435546875, + "learning_rate": 7.312842049320844e-05, + "loss": 0.9575, + "step": 2760 + }, + { + "epoch": 9.436860068259385, + "grad_norm": 0.42578125, + "learning_rate": 7.2746074229716e-05, + "loss": 0.9598, + "step": 2765 + }, + { + "epoch": 9.453924914675769, + "grad_norm": 0.423828125, + "learning_rate": 7.236415787397548e-05, + "loss": 0.9594, + "step": 2770 + }, + { + "epoch": 9.47098976109215, + "grad_norm": 0.408203125, + "learning_rate": 7.198267745039612e-05, + "loss": 0.9571, + "step": 2775 + }, + { + "epoch": 9.488054607508532, + "grad_norm": 0.41015625, + "learning_rate": 7.160163897651075e-05, + "loss": 0.9582, + "step": 2780 + }, + { + "epoch": 9.505119453924914, + "grad_norm": 0.453125, + "learning_rate": 7.122104846288064e-05, + "loss": 0.9583, + "step": 2785 + }, + { + "epoch": 9.522184300341298, + "grad_norm": 0.474609375, + "learning_rate": 7.08409119130011e-05, + "loss": 0.9713, + "step": 2790 + }, + { + "epoch": 9.53924914675768, + "grad_norm": 0.388671875, + "learning_rate": 7.04612353232063e-05, + "loss": 0.9538, + "step": 2795 + }, + { + "epoch": 9.556313993174061, + "grad_norm": 0.41796875, + "learning_rate": 7.008202468257514e-05, + "loss": 0.9572, + "step": 2800 + }, + { + "epoch": 9.573378839590443, + "grad_norm": 0.41015625, + "learning_rate": 6.970328597283637e-05, + "loss": 0.9483, + "step": 2805 + }, + { + "epoch": 9.590443686006825, + "grad_norm": 0.40234375, + "learning_rate": 6.932502516827461e-05, + "loss": 0.9521, + "step": 2810 + }, + { + "epoch": 9.607508532423209, + "grad_norm": 0.38671875, + "learning_rate": 6.894724823563583e-05, + "loss": 0.9534, + "step": 2815 + }, + { + "epoch": 9.62457337883959, + "grad_norm": 0.41015625, + "learning_rate": 6.85699611340333e-05, + "loss": 0.9611, + "step": 2820 + }, + { + "epoch": 9.641638225255972, + "grad_norm": 0.369140625, + "learning_rate": 6.819316981485372e-05, + "loss": 0.9499, + "step": 2825 + }, + { + "epoch": 9.658703071672354, + "grad_norm": 0.361328125, + "learning_rate": 6.781688022166311e-05, + "loss": 0.9689, + "step": 2830 + }, + { + "epoch": 9.675767918088738, + "grad_norm": 0.40234375, + "learning_rate": 6.744109829011332e-05, + "loss": 0.9492, + "step": 2835 + }, + { + "epoch": 9.69283276450512, + "grad_norm": 0.384765625, + "learning_rate": 6.706582994784814e-05, + "loss": 0.9626, + "step": 2840 + }, + { + "epoch": 9.709897610921502, + "grad_norm": 0.408203125, + "learning_rate": 6.669108111441003e-05, + "loss": 0.9641, + "step": 2845 + }, + { + "epoch": 9.726962457337883, + "grad_norm": 0.40234375, + "learning_rate": 6.631685770114654e-05, + "loss": 0.9578, + "step": 2850 + }, + { + "epoch": 9.744027303754265, + "grad_norm": 0.37890625, + "learning_rate": 6.594316561111724e-05, + "loss": 0.9648, + "step": 2855 + }, + { + "epoch": 9.761092150170649, + "grad_norm": 0.390625, + "learning_rate": 6.557001073900044e-05, + "loss": 0.957, + "step": 2860 + }, + { + "epoch": 9.77815699658703, + "grad_norm": 0.375, + "learning_rate": 6.519739897100034e-05, + "loss": 0.9513, + "step": 2865 + }, + { + "epoch": 9.795221843003413, + "grad_norm": 0.453125, + "learning_rate": 6.482533618475422e-05, + "loss": 0.9591, + "step": 2870 + }, + { + "epoch": 9.812286689419794, + "grad_norm": 0.369140625, + "learning_rate": 6.445382824923938e-05, + "loss": 0.9625, + "step": 2875 + }, + { + "epoch": 9.829351535836178, + "grad_norm": 0.37109375, + "learning_rate": 6.408288102468113e-05, + "loss": 0.9606, + "step": 2880 + }, + { + "epoch": 9.84641638225256, + "grad_norm": 0.37890625, + "learning_rate": 6.371250036245976e-05, + "loss": 0.9662, + "step": 2885 + }, + { + "epoch": 9.863481228668942, + "grad_norm": 0.373046875, + "learning_rate": 6.334269210501875e-05, + "loss": 0.9635, + "step": 2890 + }, + { + "epoch": 9.880546075085324, + "grad_norm": 0.365234375, + "learning_rate": 6.297346208577213e-05, + "loss": 0.9649, + "step": 2895 + }, + { + "epoch": 9.897610921501707, + "grad_norm": 0.390625, + "learning_rate": 6.260481612901299e-05, + "loss": 0.9516, + "step": 2900 + }, + { + "epoch": 9.914675767918089, + "grad_norm": 0.3828125, + "learning_rate": 6.223676004982105e-05, + "loss": 0.9601, + "step": 2905 + }, + { + "epoch": 9.93174061433447, + "grad_norm": 0.5625, + "learning_rate": 6.18692996539714e-05, + "loss": 0.9611, + "step": 2910 + }, + { + "epoch": 9.948805460750853, + "grad_norm": 0.39453125, + "learning_rate": 6.150244073784266e-05, + "loss": 0.9742, + "step": 2915 + }, + { + "epoch": 9.965870307167236, + "grad_norm": 0.4296875, + "learning_rate": 6.113618908832561e-05, + "loss": 0.9666, + "step": 2920 + }, + { + "epoch": 9.982935153583618, + "grad_norm": 0.447265625, + "learning_rate": 6.0770550482731924e-05, + "loss": 0.9684, + "step": 2925 + }, + { + "epoch": 10.0, + "grad_norm": 0.41015625, + "learning_rate": 6.0405530688702986e-05, + "loss": 0.9639, + "step": 2930 + }, + { + "epoch": 10.0, + "eval_loss": 2.512617588043213, + "eval_runtime": 0.5446, + "eval_samples_per_second": 18.362, + "eval_steps_per_second": 1.836, + "step": 2930 + }, + { + "epoch": 10.017064846416382, + "grad_norm": 0.427734375, + "learning_rate": 6.0041135464119024e-05, + "loss": 0.9618, + "step": 2935 + }, + { + "epoch": 10.034129692832764, + "grad_norm": 0.384765625, + "learning_rate": 5.9677370557008104e-05, + "loss": 0.9433, + "step": 2940 + }, + { + "epoch": 10.051194539249147, + "grad_norm": 0.478515625, + "learning_rate": 5.9314241705455674e-05, + "loss": 0.9543, + "step": 2945 + }, + { + "epoch": 10.06825938566553, + "grad_norm": 0.408203125, + "learning_rate": 5.895175463751385e-05, + "loss": 0.9579, + "step": 2950 + }, + { + "epoch": 10.085324232081911, + "grad_norm": 0.380859375, + "learning_rate": 5.858991507111122e-05, + "loss": 0.9506, + "step": 2955 + }, + { + "epoch": 10.102389078498293, + "grad_norm": 0.3828125, + "learning_rate": 5.8228728713962543e-05, + "loss": 0.9582, + "step": 2960 + }, + { + "epoch": 10.119453924914676, + "grad_norm": 0.38671875, + "learning_rate": 5.786820126347876e-05, + "loss": 0.9576, + "step": 2965 + }, + { + "epoch": 10.136518771331058, + "grad_norm": 0.4140625, + "learning_rate": 5.750833840667711e-05, + "loss": 0.9506, + "step": 2970 + }, + { + "epoch": 10.15358361774744, + "grad_norm": 0.390625, + "learning_rate": 5.7149145820091385e-05, + "loss": 0.952, + "step": 2975 + }, + { + "epoch": 10.170648464163822, + "grad_norm": 0.38671875, + "learning_rate": 5.6790629169682564e-05, + "loss": 0.9532, + "step": 2980 + }, + { + "epoch": 10.187713310580206, + "grad_norm": 0.396484375, + "learning_rate": 5.6432794110749134e-05, + "loss": 0.9459, + "step": 2985 + }, + { + "epoch": 10.204778156996587, + "grad_norm": 0.490234375, + "learning_rate": 5.607564628783817e-05, + "loss": 0.9513, + "step": 2990 + }, + { + "epoch": 10.22184300341297, + "grad_norm": 0.41796875, + "learning_rate": 5.571919133465605e-05, + "loss": 0.9499, + "step": 2995 + }, + { + "epoch": 10.238907849829351, + "grad_norm": 0.392578125, + "learning_rate": 5.5363434873979903e-05, + "loss": 0.9481, + "step": 3000 + }, + { + "epoch": 10.255972696245733, + "grad_norm": 0.380859375, + "learning_rate": 5.500838251756857e-05, + "loss": 0.9501, + "step": 3005 + }, + { + "epoch": 10.273037542662117, + "grad_norm": 0.3671875, + "learning_rate": 5.465403986607426e-05, + "loss": 0.9498, + "step": 3010 + }, + { + "epoch": 10.290102389078498, + "grad_norm": 0.396484375, + "learning_rate": 5.430041250895428e-05, + "loss": 0.947, + "step": 3015 + }, + { + "epoch": 10.30716723549488, + "grad_norm": 0.42578125, + "learning_rate": 5.3947506024382665e-05, + "loss": 0.9581, + "step": 3020 + }, + { + "epoch": 10.324232081911262, + "grad_norm": 0.408203125, + "learning_rate": 5.359532597916233e-05, + "loss": 0.9549, + "step": 3025 + }, + { + "epoch": 10.341296928327646, + "grad_norm": 0.40625, + "learning_rate": 5.324387792863719e-05, + "loss": 0.968, + "step": 3030 + }, + { + "epoch": 10.358361774744028, + "grad_norm": 0.404296875, + "learning_rate": 5.289316741660466e-05, + "loss": 0.9499, + "step": 3035 + }, + { + "epoch": 10.37542662116041, + "grad_norm": 0.3828125, + "learning_rate": 5.254319997522796e-05, + "loss": 0.9639, + "step": 3040 + }, + { + "epoch": 10.392491467576791, + "grad_norm": 0.404296875, + "learning_rate": 5.21939811249492e-05, + "loss": 0.9555, + "step": 3045 + }, + { + "epoch": 10.409556313993175, + "grad_norm": 0.38671875, + "learning_rate": 5.1845516374401784e-05, + "loss": 0.9533, + "step": 3050 + }, + { + "epoch": 10.426621160409557, + "grad_norm": 0.421875, + "learning_rate": 5.14978112203241e-05, + "loss": 0.9632, + "step": 3055 + }, + { + "epoch": 10.443686006825939, + "grad_norm": 0.380859375, + "learning_rate": 5.11508711474725e-05, + "loss": 0.9596, + "step": 3060 + }, + { + "epoch": 10.46075085324232, + "grad_norm": 0.4140625, + "learning_rate": 5.080470162853472e-05, + "loss": 0.963, + "step": 3065 + }, + { + "epoch": 10.477815699658702, + "grad_norm": 0.412109375, + "learning_rate": 5.0459308124043715e-05, + "loss": 0.9602, + "step": 3070 + }, + { + "epoch": 10.494880546075086, + "grad_norm": 0.4375, + "learning_rate": 5.0114696082291425e-05, + "loss": 0.9429, + "step": 3075 + }, + { + "epoch": 10.511945392491468, + "grad_norm": 0.3828125, + "learning_rate": 4.9770870939242986e-05, + "loss": 0.9569, + "step": 3080 + }, + { + "epoch": 10.52901023890785, + "grad_norm": 0.396484375, + "learning_rate": 4.942783811845074e-05, + "loss": 0.945, + "step": 3085 + }, + { + "epoch": 10.546075085324231, + "grad_norm": 0.38671875, + "learning_rate": 4.908560303096887e-05, + "loss": 0.955, + "step": 3090 + }, + { + "epoch": 10.563139931740615, + "grad_norm": 0.404296875, + "learning_rate": 4.874417107526795e-05, + "loss": 0.9583, + "step": 3095 + }, + { + "epoch": 10.580204778156997, + "grad_norm": 0.38671875, + "learning_rate": 4.840354763714991e-05, + "loss": 0.9499, + "step": 3100 + }, + { + "epoch": 10.597269624573379, + "grad_norm": 0.41015625, + "learning_rate": 4.8063738089662926e-05, + "loss": 0.9528, + "step": 3105 + }, + { + "epoch": 10.61433447098976, + "grad_norm": 0.373046875, + "learning_rate": 4.772474779301669e-05, + "loss": 0.9581, + "step": 3110 + }, + { + "epoch": 10.631399317406144, + "grad_norm": 0.4140625, + "learning_rate": 4.738658209449805e-05, + "loss": 0.9456, + "step": 3115 + }, + { + "epoch": 10.648464163822526, + "grad_norm": 0.384765625, + "learning_rate": 4.704924632838636e-05, + "loss": 0.9507, + "step": 3120 + }, + { + "epoch": 10.665529010238908, + "grad_norm": 0.384765625, + "learning_rate": 4.671274581586958e-05, + "loss": 0.9586, + "step": 3125 + }, + { + "epoch": 10.68259385665529, + "grad_norm": 0.375, + "learning_rate": 4.637708586496018e-05, + "loss": 0.9487, + "step": 3130 + }, + { + "epoch": 10.699658703071673, + "grad_norm": 0.38671875, + "learning_rate": 4.604227177041156e-05, + "loss": 0.9511, + "step": 3135 + }, + { + "epoch": 10.716723549488055, + "grad_norm": 0.404296875, + "learning_rate": 4.570830881363439e-05, + "loss": 0.9529, + "step": 3140 + }, + { + "epoch": 10.733788395904437, + "grad_norm": 0.5078125, + "learning_rate": 4.537520226261333e-05, + "loss": 0.962, + "step": 3145 + }, + { + "epoch": 10.750853242320819, + "grad_norm": 0.396484375, + "learning_rate": 4.5042957371824057e-05, + "loss": 0.9551, + "step": 3150 + }, + { + "epoch": 10.7679180887372, + "grad_norm": 0.42578125, + "learning_rate": 4.471157938215017e-05, + "loss": 0.9537, + "step": 3155 + }, + { + "epoch": 10.784982935153584, + "grad_norm": 0.3984375, + "learning_rate": 4.438107352080076e-05, + "loss": 0.9573, + "step": 3160 + }, + { + "epoch": 10.802047781569966, + "grad_norm": 0.384765625, + "learning_rate": 4.405144500122772e-05, + "loss": 0.9615, + "step": 3165 + }, + { + "epoch": 10.819112627986348, + "grad_norm": 0.365234375, + "learning_rate": 4.372269902304363e-05, + "loss": 0.9592, + "step": 3170 + }, + { + "epoch": 10.83617747440273, + "grad_norm": 0.38671875, + "learning_rate": 4.339484077193974e-05, + "loss": 0.9518, + "step": 3175 + }, + { + "epoch": 10.853242320819113, + "grad_norm": 0.423828125, + "learning_rate": 4.3067875419604184e-05, + "loss": 0.953, + "step": 3180 + }, + { + "epoch": 10.870307167235495, + "grad_norm": 0.376953125, + "learning_rate": 4.2741808123640335e-05, + "loss": 0.9578, + "step": 3185 + }, + { + "epoch": 10.887372013651877, + "grad_norm": 0.36328125, + "learning_rate": 4.241664402748544e-05, + "loss": 0.9548, + "step": 3190 + }, + { + "epoch": 10.904436860068259, + "grad_norm": 0.361328125, + "learning_rate": 4.209238826032965e-05, + "loss": 0.955, + "step": 3195 + }, + { + "epoch": 10.921501706484642, + "grad_norm": 0.380859375, + "learning_rate": 4.1769045937034876e-05, + "loss": 0.9591, + "step": 3200 + }, + { + "epoch": 10.938566552901024, + "grad_norm": 0.43359375, + "learning_rate": 4.144662215805426e-05, + "loss": 0.9544, + "step": 3205 + }, + { + "epoch": 10.955631399317406, + "grad_norm": 0.58984375, + "learning_rate": 4.1125122009351634e-05, + "loss": 0.9539, + "step": 3210 + }, + { + "epoch": 10.972696245733788, + "grad_norm": 0.416015625, + "learning_rate": 4.080455056232147e-05, + "loss": 0.9497, + "step": 3215 + }, + { + "epoch": 10.98976109215017, + "grad_norm": 0.421875, + "learning_rate": 4.048491287370863e-05, + "loss": 0.952, + "step": 3220 + }, + { + "epoch": 11.0, + "eval_loss": 2.519228935241699, + "eval_runtime": 0.5351, + "eval_samples_per_second": 18.688, + "eval_steps_per_second": 1.869, + "step": 3223 + }, + { + "epoch": 11.006825938566553, + "grad_norm": 0.404296875, + "learning_rate": 4.016621398552877e-05, + "loss": 0.954, + "step": 3225 + }, + { + "epoch": 11.023890784982935, + "grad_norm": 0.390625, + "learning_rate": 3.9848458924988684e-05, + "loss": 0.9494, + "step": 3230 + }, + { + "epoch": 11.040955631399317, + "grad_norm": 0.404296875, + "learning_rate": 3.953165270440721e-05, + "loss": 0.9434, + "step": 3235 + }, + { + "epoch": 11.058020477815699, + "grad_norm": 0.38671875, + "learning_rate": 3.921580032113602e-05, + "loss": 0.9542, + "step": 3240 + }, + { + "epoch": 11.075085324232083, + "grad_norm": 0.388671875, + "learning_rate": 3.8900906757480614e-05, + "loss": 0.9519, + "step": 3245 + }, + { + "epoch": 11.092150170648464, + "grad_norm": 0.388671875, + "learning_rate": 3.858697698062217e-05, + "loss": 0.9597, + "step": 3250 + }, + { + "epoch": 11.109215017064846, + "grad_norm": 0.373046875, + "learning_rate": 3.8274015942538745e-05, + "loss": 0.9437, + "step": 3255 + }, + { + "epoch": 11.126279863481228, + "grad_norm": 0.37890625, + "learning_rate": 3.7962028579927555e-05, + "loss": 0.9545, + "step": 3260 + }, + { + "epoch": 11.143344709897612, + "grad_norm": 0.392578125, + "learning_rate": 3.7651019814126654e-05, + "loss": 0.9524, + "step": 3265 + }, + { + "epoch": 11.160409556313994, + "grad_norm": 0.37890625, + "learning_rate": 3.734099455103779e-05, + "loss": 0.9591, + "step": 3270 + }, + { + "epoch": 11.177474402730375, + "grad_norm": 0.38671875, + "learning_rate": 3.7031957681048604e-05, + "loss": 0.9503, + "step": 3275 + }, + { + "epoch": 11.194539249146757, + "grad_norm": 0.384765625, + "learning_rate": 3.6723914078955825e-05, + "loss": 0.9456, + "step": 3280 + }, + { + "epoch": 11.211604095563139, + "grad_norm": 0.380859375, + "learning_rate": 3.64168686038881e-05, + "loss": 0.9426, + "step": 3285 + }, + { + "epoch": 11.228668941979523, + "grad_norm": 0.390625, + "learning_rate": 3.6110826099229453e-05, + "loss": 0.9496, + "step": 3290 + }, + { + "epoch": 11.245733788395905, + "grad_norm": 0.37109375, + "learning_rate": 3.580579139254303e-05, + "loss": 0.9515, + "step": 3295 + }, + { + "epoch": 11.262798634812286, + "grad_norm": 0.3828125, + "learning_rate": 3.550176929549468e-05, + "loss": 0.9535, + "step": 3300 + }, + { + "epoch": 11.279863481228668, + "grad_norm": 0.3671875, + "learning_rate": 3.5198764603777235e-05, + "loss": 0.9575, + "step": 3305 + }, + { + "epoch": 11.296928327645052, + "grad_norm": 0.376953125, + "learning_rate": 3.489678209703475e-05, + "loss": 0.9468, + "step": 3310 + }, + { + "epoch": 11.313993174061434, + "grad_norm": 0.396484375, + "learning_rate": 3.459582653878731e-05, + "loss": 0.9536, + "step": 3315 + }, + { + "epoch": 11.331058020477816, + "grad_norm": 0.39453125, + "learning_rate": 3.429590267635565e-05, + "loss": 0.9575, + "step": 3320 + }, + { + "epoch": 11.348122866894197, + "grad_norm": 0.38671875, + "learning_rate": 3.399701524078635e-05, + "loss": 0.9533, + "step": 3325 + }, + { + "epoch": 11.365187713310581, + "grad_norm": 0.380859375, + "learning_rate": 3.369916894677733e-05, + "loss": 0.9414, + "step": 3330 + }, + { + "epoch": 11.382252559726963, + "grad_norm": 0.421875, + "learning_rate": 3.340236849260324e-05, + "loss": 0.9494, + "step": 3335 + }, + { + "epoch": 11.399317406143345, + "grad_norm": 0.419921875, + "learning_rate": 3.31066185600417e-05, + "loss": 0.9457, + "step": 3340 + }, + { + "epoch": 11.416382252559726, + "grad_norm": 0.384765625, + "learning_rate": 3.281192381429894e-05, + "loss": 0.9403, + "step": 3345 + }, + { + "epoch": 11.43344709897611, + "grad_norm": 0.375, + "learning_rate": 3.251828890393677e-05, + "loss": 0.9489, + "step": 3350 + }, + { + "epoch": 11.450511945392492, + "grad_norm": 0.412109375, + "learning_rate": 3.222571846079881e-05, + "loss": 0.9525, + "step": 3355 + }, + { + "epoch": 11.467576791808874, + "grad_norm": 0.37109375, + "learning_rate": 3.193421709993779e-05, + "loss": 0.9574, + "step": 3360 + }, + { + "epoch": 11.484641638225256, + "grad_norm": 0.390625, + "learning_rate": 3.1643789419542324e-05, + "loss": 0.9453, + "step": 3365 + }, + { + "epoch": 11.501706484641637, + "grad_norm": 0.3828125, + "learning_rate": 3.135444000086485e-05, + "loss": 0.9462, + "step": 3370 + }, + { + "epoch": 11.518771331058021, + "grad_norm": 0.384765625, + "learning_rate": 3.1066173408148955e-05, + "loss": 0.9551, + "step": 3375 + }, + { + "epoch": 11.535836177474403, + "grad_norm": 0.404296875, + "learning_rate": 3.077899418855772e-05, + "loss": 0.9504, + "step": 3380 + }, + { + "epoch": 11.552901023890785, + "grad_norm": 0.400390625, + "learning_rate": 3.04929068721017e-05, + "loss": 0.9496, + "step": 3385 + }, + { + "epoch": 11.569965870307167, + "grad_norm": 0.380859375, + "learning_rate": 3.0207915971567624e-05, + "loss": 0.9426, + "step": 3390 + }, + { + "epoch": 11.58703071672355, + "grad_norm": 0.384765625, + "learning_rate": 2.992402598244727e-05, + "loss": 0.9458, + "step": 3395 + }, + { + "epoch": 11.604095563139932, + "grad_norm": 0.384765625, + "learning_rate": 2.9641241382866348e-05, + "loss": 0.9525, + "step": 3400 + }, + { + "epoch": 11.621160409556314, + "grad_norm": 0.400390625, + "learning_rate": 2.9359566633514037e-05, + "loss": 0.9449, + "step": 3405 + }, + { + "epoch": 11.638225255972696, + "grad_norm": 0.380859375, + "learning_rate": 2.907900617757252e-05, + "loss": 0.9526, + "step": 3410 + }, + { + "epoch": 11.655290102389078, + "grad_norm": 0.373046875, + "learning_rate": 2.879956444064703e-05, + "loss": 0.9598, + "step": 3415 + }, + { + "epoch": 11.672354948805461, + "grad_norm": 0.388671875, + "learning_rate": 2.8521245830695864e-05, + "loss": 0.9484, + "step": 3420 + }, + { + "epoch": 11.689419795221843, + "grad_norm": 0.3828125, + "learning_rate": 2.8244054737960935e-05, + "loss": 0.9431, + "step": 3425 + }, + { + "epoch": 11.706484641638225, + "grad_norm": 0.365234375, + "learning_rate": 2.7967995534898596e-05, + "loss": 0.9554, + "step": 3430 + }, + { + "epoch": 11.723549488054607, + "grad_norm": 0.390625, + "learning_rate": 2.7693072576110514e-05, + "loss": 0.9519, + "step": 3435 + }, + { + "epoch": 11.74061433447099, + "grad_norm": 0.365234375, + "learning_rate": 2.7419290198275095e-05, + "loss": 0.9509, + "step": 3440 + }, + { + "epoch": 11.757679180887372, + "grad_norm": 0.40234375, + "learning_rate": 2.7146652720079003e-05, + "loss": 0.9578, + "step": 3445 + }, + { + "epoch": 11.774744027303754, + "grad_norm": 0.376953125, + "learning_rate": 2.6875164442149147e-05, + "loss": 0.9449, + "step": 3450 + }, + { + "epoch": 11.791808873720136, + "grad_norm": 0.40625, + "learning_rate": 2.6604829646984686e-05, + "loss": 0.9505, + "step": 3455 + }, + { + "epoch": 11.80887372013652, + "grad_norm": 0.3984375, + "learning_rate": 2.6335652598889683e-05, + "loss": 0.9433, + "step": 3460 + }, + { + "epoch": 11.825938566552901, + "grad_norm": 0.380859375, + "learning_rate": 2.60676375439055e-05, + "loss": 0.9464, + "step": 3465 + }, + { + "epoch": 11.843003412969283, + "grad_norm": 0.384765625, + "learning_rate": 2.5800788709744227e-05, + "loss": 0.955, + "step": 3470 + }, + { + "epoch": 11.860068259385665, + "grad_norm": 0.380859375, + "learning_rate": 2.5535110305721776e-05, + "loss": 0.9458, + "step": 3475 + }, + { + "epoch": 11.877133105802049, + "grad_norm": 0.3828125, + "learning_rate": 2.5270606522691443e-05, + "loss": 0.9544, + "step": 3480 + }, + { + "epoch": 11.89419795221843, + "grad_norm": 0.408203125, + "learning_rate": 2.500728153297788e-05, + "loss": 0.9534, + "step": 3485 + }, + { + "epoch": 11.911262798634812, + "grad_norm": 0.373046875, + "learning_rate": 2.4745139490311254e-05, + "loss": 0.9521, + "step": 3490 + }, + { + "epoch": 11.928327645051194, + "grad_norm": 0.392578125, + "learning_rate": 2.4484184529761834e-05, + "loss": 0.948, + "step": 3495 + }, + { + "epoch": 11.945392491467576, + "grad_norm": 0.39453125, + "learning_rate": 2.4224420767674562e-05, + "loss": 0.9543, + "step": 3500 + }, + { + "epoch": 11.96245733788396, + "grad_norm": 0.375, + "learning_rate": 2.3965852301604254e-05, + "loss": 0.959, + "step": 3505 + }, + { + "epoch": 11.979522184300341, + "grad_norm": 0.375, + "learning_rate": 2.370848321025093e-05, + "loss": 0.9599, + "step": 3510 + }, + { + "epoch": 11.996587030716723, + "grad_norm": 0.37109375, + "learning_rate": 2.345231755339554e-05, + "loss": 0.9505, + "step": 3515 + }, + { + "epoch": 12.0, + "eval_loss": 2.520477771759033, + "eval_runtime": 0.5502, + "eval_samples_per_second": 18.175, + "eval_steps_per_second": 1.818, + "step": 3516 + }, + { + "epoch": 12.013651877133105, + "grad_norm": 0.43359375, + "learning_rate": 2.3197359371835802e-05, + "loss": 0.9615, + "step": 3520 + }, + { + "epoch": 12.030716723549489, + "grad_norm": 0.376953125, + "learning_rate": 2.2943612687322525e-05, + "loss": 0.9485, + "step": 3525 + }, + { + "epoch": 12.04778156996587, + "grad_norm": 0.384765625, + "learning_rate": 2.2691081502496246e-05, + "loss": 0.9475, + "step": 3530 + }, + { + "epoch": 12.064846416382252, + "grad_norm": 0.388671875, + "learning_rate": 2.243976980082394e-05, + "loss": 0.9393, + "step": 3535 + }, + { + "epoch": 12.081911262798634, + "grad_norm": 0.39453125, + "learning_rate": 2.218968154653629e-05, + "loss": 0.9466, + "step": 3540 + }, + { + "epoch": 12.098976109215018, + "grad_norm": 0.376953125, + "learning_rate": 2.194082068456509e-05, + "loss": 0.9537, + "step": 3545 + }, + { + "epoch": 12.1160409556314, + "grad_norm": 0.36328125, + "learning_rate": 2.169319114048114e-05, + "loss": 0.961, + "step": 3550 + }, + { + "epoch": 12.133105802047782, + "grad_norm": 0.38671875, + "learning_rate": 2.1446796820432167e-05, + "loss": 0.9493, + "step": 3555 + }, + { + "epoch": 12.150170648464163, + "grad_norm": 0.384765625, + "learning_rate": 2.1201641611081246e-05, + "loss": 0.948, + "step": 3560 + }, + { + "epoch": 12.167235494880545, + "grad_norm": 0.373046875, + "learning_rate": 2.0957729379545655e-05, + "loss": 0.9584, + "step": 3565 + }, + { + "epoch": 12.184300341296929, + "grad_norm": 0.380859375, + "learning_rate": 2.0715063973335568e-05, + "loss": 0.9503, + "step": 3570 + }, + { + "epoch": 12.20136518771331, + "grad_norm": 0.388671875, + "learning_rate": 2.04736492202937e-05, + "loss": 0.9498, + "step": 3575 + }, + { + "epoch": 12.218430034129693, + "grad_norm": 0.392578125, + "learning_rate": 2.0233488928534673e-05, + "loss": 0.9553, + "step": 3580 + }, + { + "epoch": 12.235494880546074, + "grad_norm": 0.396484375, + "learning_rate": 1.9994586886385046e-05, + "loss": 0.9438, + "step": 3585 + }, + { + "epoch": 12.252559726962458, + "grad_norm": 0.369140625, + "learning_rate": 1.9756946862323535e-05, + "loss": 0.9489, + "step": 3590 + }, + { + "epoch": 12.26962457337884, + "grad_norm": 0.369140625, + "learning_rate": 1.9520572604921672e-05, + "loss": 0.9477, + "step": 3595 + }, + { + "epoch": 12.286689419795222, + "grad_norm": 0.375, + "learning_rate": 1.9285467842784467e-05, + "loss": 0.9457, + "step": 3600 + }, + { + "epoch": 12.303754266211604, + "grad_norm": 0.380859375, + "learning_rate": 1.9051636284491757e-05, + "loss": 0.9541, + "step": 3605 + }, + { + "epoch": 12.320819112627987, + "grad_norm": 0.365234375, + "learning_rate": 1.8819081618539723e-05, + "loss": 0.9393, + "step": 3610 + }, + { + "epoch": 12.337883959044369, + "grad_norm": 0.375, + "learning_rate": 1.858780751328255e-05, + "loss": 0.949, + "step": 3615 + }, + { + "epoch": 12.35494880546075, + "grad_norm": 0.384765625, + "learning_rate": 1.8357817616874694e-05, + "loss": 0.9537, + "step": 3620 + }, + { + "epoch": 12.372013651877133, + "grad_norm": 0.3671875, + "learning_rate": 1.8129115557213262e-05, + "loss": 0.9505, + "step": 3625 + }, + { + "epoch": 12.389078498293514, + "grad_norm": 0.3671875, + "learning_rate": 1.7901704941880914e-05, + "loss": 0.9447, + "step": 3630 + }, + { + "epoch": 12.406143344709898, + "grad_norm": 0.3671875, + "learning_rate": 1.7675589358088763e-05, + "loss": 0.9526, + "step": 3635 + }, + { + "epoch": 12.42320819112628, + "grad_norm": 0.376953125, + "learning_rate": 1.745077237261994e-05, + "loss": 0.9592, + "step": 3640 + }, + { + "epoch": 12.440273037542662, + "grad_norm": 0.40234375, + "learning_rate": 1.7227257531773223e-05, + "loss": 0.9515, + "step": 3645 + }, + { + "epoch": 12.457337883959044, + "grad_norm": 0.408203125, + "learning_rate": 1.7005048361307262e-05, + "loss": 0.9504, + "step": 3650 + }, + { + "epoch": 12.474402730375427, + "grad_norm": 0.388671875, + "learning_rate": 1.6784148366384754e-05, + "loss": 0.9462, + "step": 3655 + }, + { + "epoch": 12.491467576791809, + "grad_norm": 0.384765625, + "learning_rate": 1.656456103151728e-05, + "loss": 0.9456, + "step": 3660 + }, + { + "epoch": 12.508532423208191, + "grad_norm": 0.375, + "learning_rate": 1.6346289820510363e-05, + "loss": 0.9475, + "step": 3665 + }, + { + "epoch": 12.525597269624573, + "grad_norm": 0.384765625, + "learning_rate": 1.612933817640868e-05, + "loss": 0.9478, + "step": 3670 + }, + { + "epoch": 12.542662116040956, + "grad_norm": 0.3671875, + "learning_rate": 1.5913709521441988e-05, + "loss": 0.9415, + "step": 3675 + }, + { + "epoch": 12.559726962457338, + "grad_norm": 0.375, + "learning_rate": 1.5699407256970833e-05, + "loss": 0.9452, + "step": 3680 + }, + { + "epoch": 12.57679180887372, + "grad_norm": 0.375, + "learning_rate": 1.5486434763433222e-05, + "loss": 0.9479, + "step": 3685 + }, + { + "epoch": 12.593856655290102, + "grad_norm": 0.38671875, + "learning_rate": 1.527479540029104e-05, + "loss": 0.9495, + "step": 3690 + }, + { + "epoch": 12.610921501706486, + "grad_norm": 0.3828125, + "learning_rate": 1.5064492505977234e-05, + "loss": 0.936, + "step": 3695 + }, + { + "epoch": 12.627986348122867, + "grad_norm": 0.392578125, + "learning_rate": 1.4855529397843038e-05, + "loss": 0.9476, + "step": 3700 + }, + { + "epoch": 12.64505119453925, + "grad_norm": 0.380859375, + "learning_rate": 1.4647909372105672e-05, + "loss": 0.9525, + "step": 3705 + }, + { + "epoch": 12.662116040955631, + "grad_norm": 0.41796875, + "learning_rate": 1.4441635703796408e-05, + "loss": 0.9477, + "step": 3710 + }, + { + "epoch": 12.679180887372013, + "grad_norm": 0.3984375, + "learning_rate": 1.4236711646708844e-05, + "loss": 0.9505, + "step": 3715 + }, + { + "epoch": 12.696245733788396, + "grad_norm": 0.384765625, + "learning_rate": 1.4033140433347569e-05, + "loss": 0.9464, + "step": 3720 + }, + { + "epoch": 12.713310580204778, + "grad_norm": 0.384765625, + "learning_rate": 1.3830925274877216e-05, + "loss": 0.9392, + "step": 3725 + }, + { + "epoch": 12.73037542662116, + "grad_norm": 0.37890625, + "learning_rate": 1.363006936107183e-05, + "loss": 0.9495, + "step": 3730 + }, + { + "epoch": 12.747440273037542, + "grad_norm": 0.3828125, + "learning_rate": 1.343057586026446e-05, + "loss": 0.9423, + "step": 3735 + }, + { + "epoch": 12.764505119453926, + "grad_norm": 0.416015625, + "learning_rate": 1.3232447919297274e-05, + "loss": 0.9448, + "step": 3740 + }, + { + "epoch": 12.781569965870307, + "grad_norm": 0.404296875, + "learning_rate": 1.3035688663471834e-05, + "loss": 0.9544, + "step": 3745 + }, + { + "epoch": 12.79863481228669, + "grad_norm": 0.37109375, + "learning_rate": 1.2840301196499893e-05, + "loss": 0.9548, + "step": 3750 + }, + { + "epoch": 12.815699658703071, + "grad_norm": 0.376953125, + "learning_rate": 1.2646288600454448e-05, + "loss": 0.9492, + "step": 3755 + }, + { + "epoch": 12.832764505119453, + "grad_norm": 0.373046875, + "learning_rate": 1.2453653935720867e-05, + "loss": 0.9506, + "step": 3760 + }, + { + "epoch": 12.849829351535837, + "grad_norm": 0.388671875, + "learning_rate": 1.2262400240949023e-05, + "loss": 0.9543, + "step": 3765 + }, + { + "epoch": 12.866894197952218, + "grad_norm": 0.369140625, + "learning_rate": 1.2072530533005012e-05, + "loss": 0.9418, + "step": 3770 + }, + { + "epoch": 12.8839590443686, + "grad_norm": 0.369140625, + "learning_rate": 1.1884047806923815e-05, + "loss": 0.9475, + "step": 3775 + }, + { + "epoch": 12.901023890784982, + "grad_norm": 0.39453125, + "learning_rate": 1.169695503586179e-05, + "loss": 0.9428, + "step": 3780 + }, + { + "epoch": 12.918088737201366, + "grad_norm": 0.38671875, + "learning_rate": 1.1511255171050084e-05, + "loss": 0.9529, + "step": 3785 + }, + { + "epoch": 12.935153583617748, + "grad_norm": 0.376953125, + "learning_rate": 1.1326951141747788e-05, + "loss": 0.9455, + "step": 3790 + }, + { + "epoch": 12.95221843003413, + "grad_norm": 0.376953125, + "learning_rate": 1.1144045855195973e-05, + "loss": 0.9537, + "step": 3795 + }, + { + "epoch": 12.969283276450511, + "grad_norm": 0.396484375, + "learning_rate": 1.0962542196571634e-05, + "loss": 0.9426, + "step": 3800 + }, + { + "epoch": 12.986348122866895, + "grad_norm": 0.373046875, + "learning_rate": 1.078244302894229e-05, + "loss": 0.9442, + "step": 3805 + }, + { + "epoch": 13.0, + "eval_loss": 2.522336959838867, + "eval_runtime": 0.5484, + "eval_samples_per_second": 18.236, + "eval_steps_per_second": 1.824, + "step": 3809 + }, + { + "epoch": 13.003412969283277, + "grad_norm": 0.376953125, + "learning_rate": 1.0603751193220846e-05, + "loss": 0.956, + "step": 3810 + }, + { + "epoch": 13.020477815699659, + "grad_norm": 0.392578125, + "learning_rate": 1.0426469508120662e-05, + "loss": 0.9449, + "step": 3815 + }, + { + "epoch": 13.03754266211604, + "grad_norm": 0.390625, + "learning_rate": 1.0250600770111185e-05, + "loss": 0.9479, + "step": 3820 + }, + { + "epoch": 13.054607508532424, + "grad_norm": 0.392578125, + "learning_rate": 1.0076147753373789e-05, + "loss": 0.953, + "step": 3825 + }, + { + "epoch": 13.071672354948806, + "grad_norm": 0.388671875, + "learning_rate": 9.903113209758096e-06, + "loss": 0.9436, + "step": 3830 + }, + { + "epoch": 13.088737201365188, + "grad_norm": 0.380859375, + "learning_rate": 9.731499868738447e-06, + "loss": 0.9454, + "step": 3835 + }, + { + "epoch": 13.10580204778157, + "grad_norm": 0.3828125, + "learning_rate": 9.561310437370907e-06, + "loss": 0.9556, + "step": 3840 + }, + { + "epoch": 13.122866894197951, + "grad_norm": 0.373046875, + "learning_rate": 9.392547600250634e-06, + "loss": 0.949, + "step": 3845 + }, + { + "epoch": 13.139931740614335, + "grad_norm": 0.380859375, + "learning_rate": 9.225214019469385e-06, + "loss": 0.9382, + "step": 3850 + }, + { + "epoch": 13.156996587030717, + "grad_norm": 0.40234375, + "learning_rate": 9.059312334573633e-06, + "loss": 0.943, + "step": 3855 + }, + { + "epoch": 13.174061433447099, + "grad_norm": 0.3828125, + "learning_rate": 8.89484516252287e-06, + "loss": 0.9534, + "step": 3860 + }, + { + "epoch": 13.19112627986348, + "grad_norm": 0.369140625, + "learning_rate": 8.731815097648433e-06, + "loss": 0.9526, + "step": 3865 + }, + { + "epoch": 13.208191126279864, + "grad_norm": 0.392578125, + "learning_rate": 8.570224711612385e-06, + "loss": 0.9419, + "step": 3870 + }, + { + "epoch": 13.225255972696246, + "grad_norm": 0.373046875, + "learning_rate": 8.410076553367208e-06, + "loss": 0.9511, + "step": 3875 + }, + { + "epoch": 13.242320819112628, + "grad_norm": 0.380859375, + "learning_rate": 8.251373149115293e-06, + "loss": 0.9489, + "step": 3880 + }, + { + "epoch": 13.25938566552901, + "grad_norm": 0.36328125, + "learning_rate": 8.094117002269363e-06, + "loss": 0.9428, + "step": 3885 + }, + { + "epoch": 13.276450511945393, + "grad_norm": 0.443359375, + "learning_rate": 7.938310593412879e-06, + "loss": 0.9485, + "step": 3890 + }, + { + "epoch": 13.293515358361775, + "grad_norm": 0.3671875, + "learning_rate": 7.783956380260837e-06, + "loss": 0.955, + "step": 3895 + }, + { + "epoch": 13.310580204778157, + "grad_norm": 0.384765625, + "learning_rate": 7.631056797621106e-06, + "loss": 0.9566, + "step": 3900 + }, + { + "epoch": 13.327645051194539, + "grad_norm": 0.369140625, + "learning_rate": 7.479614257355971e-06, + "loss": 0.9495, + "step": 3905 + }, + { + "epoch": 13.344709897610922, + "grad_norm": 0.376953125, + "learning_rate": 7.329631148344118e-06, + "loss": 0.9535, + "step": 3910 + }, + { + "epoch": 13.361774744027304, + "grad_norm": 0.375, + "learning_rate": 7.181109836442912e-06, + "loss": 0.9473, + "step": 3915 + }, + { + "epoch": 13.378839590443686, + "grad_norm": 0.37890625, + "learning_rate": 7.034052664451118e-06, + "loss": 0.946, + "step": 3920 + }, + { + "epoch": 13.395904436860068, + "grad_norm": 0.380859375, + "learning_rate": 6.88846195207189e-06, + "loss": 0.9526, + "step": 3925 + }, + { + "epoch": 13.41296928327645, + "grad_norm": 0.365234375, + "learning_rate": 6.7443399958762584e-06, + "loss": 0.9416, + "step": 3930 + }, + { + "epoch": 13.430034129692833, + "grad_norm": 0.365234375, + "learning_rate": 6.6016890692668364e-06, + "loss": 0.9529, + "step": 3935 + }, + { + "epoch": 13.447098976109215, + "grad_norm": 0.376953125, + "learning_rate": 6.460511422441984e-06, + "loss": 0.9427, + "step": 3940 + }, + { + "epoch": 13.464163822525597, + "grad_norm": 0.37890625, + "learning_rate": 6.320809282360319e-06, + "loss": 0.9516, + "step": 3945 + }, + { + "epoch": 13.481228668941979, + "grad_norm": 0.380859375, + "learning_rate": 6.1825848527055865e-06, + "loss": 0.9448, + "step": 3950 + }, + { + "epoch": 13.498293515358363, + "grad_norm": 0.384765625, + "learning_rate": 6.04584031385188e-06, + "loss": 0.9542, + "step": 3955 + }, + { + "epoch": 13.515358361774744, + "grad_norm": 0.376953125, + "learning_rate": 5.910577822829233e-06, + "loss": 0.9525, + "step": 3960 + }, + { + "epoch": 13.532423208191126, + "grad_norm": 0.3671875, + "learning_rate": 5.77679951328971e-06, + "loss": 0.9502, + "step": 3965 + }, + { + "epoch": 13.549488054607508, + "grad_norm": 0.373046875, + "learning_rate": 5.644507495473572e-06, + "loss": 0.9464, + "step": 3970 + }, + { + "epoch": 13.56655290102389, + "grad_norm": 0.37890625, + "learning_rate": 5.5137038561761115e-06, + "loss": 0.9531, + "step": 3975 + }, + { + "epoch": 13.583617747440274, + "grad_norm": 0.375, + "learning_rate": 5.3843906587146886e-06, + "loss": 0.9498, + "step": 3980 + }, + { + "epoch": 13.600682593856655, + "grad_norm": 0.37890625, + "learning_rate": 5.256569942896217e-06, + "loss": 0.945, + "step": 3985 + }, + { + "epoch": 13.617747440273037, + "grad_norm": 0.365234375, + "learning_rate": 5.130243724984995e-06, + "loss": 0.9468, + "step": 3990 + }, + { + "epoch": 13.634812286689419, + "grad_norm": 0.369140625, + "learning_rate": 5.005413997670816e-06, + "loss": 0.9517, + "step": 3995 + }, + { + "epoch": 13.651877133105803, + "grad_norm": 0.365234375, + "learning_rate": 4.8820827300376075e-06, + "loss": 0.9502, + "step": 4000 + }, + { + "epoch": 13.668941979522184, + "grad_norm": 0.369140625, + "learning_rate": 4.760251867532362e-06, + "loss": 0.9462, + "step": 4005 + }, + { + "epoch": 13.686006825938566, + "grad_norm": 0.384765625, + "learning_rate": 4.639923331934471e-06, + "loss": 0.9476, + "step": 4010 + }, + { + "epoch": 13.703071672354948, + "grad_norm": 0.369140625, + "learning_rate": 4.521099021325336e-06, + "loss": 0.9556, + "step": 4015 + }, + { + "epoch": 13.720136518771332, + "grad_norm": 0.390625, + "learning_rate": 4.403780810058511e-06, + "loss": 0.9438, + "step": 4020 + }, + { + "epoch": 13.737201365187714, + "grad_norm": 0.470703125, + "learning_rate": 4.287970548730069e-06, + "loss": 0.9495, + "step": 4025 + }, + { + "epoch": 13.754266211604095, + "grad_norm": 0.36328125, + "learning_rate": 4.173670064149482e-06, + "loss": 0.934, + "step": 4030 + }, + { + "epoch": 13.771331058020477, + "grad_norm": 0.384765625, + "learning_rate": 4.060881159310725e-06, + "loss": 0.9502, + "step": 4035 + }, + { + "epoch": 13.788395904436861, + "grad_norm": 0.388671875, + "learning_rate": 3.949605613363882e-06, + "loss": 0.939, + "step": 4040 + }, + { + "epoch": 13.805460750853243, + "grad_norm": 0.37890625, + "learning_rate": 3.839845181587098e-06, + "loss": 0.9559, + "step": 4045 + }, + { + "epoch": 13.822525597269625, + "grad_norm": 0.376953125, + "learning_rate": 3.7316015953588467e-06, + "loss": 0.9547, + "step": 4050 + }, + { + "epoch": 13.839590443686006, + "grad_norm": 0.384765625, + "learning_rate": 3.6248765621306414e-06, + "loss": 0.9463, + "step": 4055 + }, + { + "epoch": 13.856655290102388, + "grad_norm": 0.376953125, + "learning_rate": 3.519671765400079e-06, + "loss": 0.9454, + "step": 4060 + }, + { + "epoch": 13.873720136518772, + "grad_norm": 0.373046875, + "learning_rate": 3.4159888646843495e-06, + "loss": 0.9485, + "step": 4065 + }, + { + "epoch": 13.890784982935154, + "grad_norm": 0.375, + "learning_rate": 3.313829495493992e-06, + "loss": 0.9455, + "step": 4070 + }, + { + "epoch": 13.907849829351536, + "grad_norm": 0.37890625, + "learning_rate": 3.2131952693070898e-06, + "loss": 0.9409, + "step": 4075 + }, + { + "epoch": 13.924914675767917, + "grad_norm": 0.396484375, + "learning_rate": 3.1140877735439387e-06, + "loss": 0.9468, + "step": 4080 + }, + { + "epoch": 13.941979522184301, + "grad_norm": 0.375, + "learning_rate": 3.0165085715418763e-06, + "loss": 0.9434, + "step": 4085 + }, + { + "epoch": 13.959044368600683, + "grad_norm": 0.3671875, + "learning_rate": 2.9204592025307566e-06, + "loss": 0.9455, + "step": 4090 + }, + { + "epoch": 13.976109215017065, + "grad_norm": 0.369140625, + "learning_rate": 2.8259411816085492e-06, + "loss": 0.9437, + "step": 4095 + }, + { + "epoch": 13.993174061433447, + "grad_norm": 0.478515625, + "learning_rate": 2.732955999717546e-06, + "loss": 0.9469, + "step": 4100 + }, + { + "epoch": 14.0, + "eval_loss": 2.5227127075195312, + "eval_runtime": 0.542, + "eval_samples_per_second": 18.45, + "eval_steps_per_second": 1.845, + "step": 4102 + }, + { + "epoch": 14.01023890784983, + "grad_norm": 0.376953125, + "learning_rate": 2.6415051236207355e-06, + "loss": 0.9508, + "step": 4105 + }, + { + "epoch": 14.027303754266212, + "grad_norm": 0.375, + "learning_rate": 2.551589995878789e-06, + "loss": 0.9459, + "step": 4110 + }, + { + "epoch": 14.044368600682594, + "grad_norm": 0.380859375, + "learning_rate": 2.4632120348272003e-06, + "loss": 0.9465, + "step": 4115 + }, + { + "epoch": 14.061433447098976, + "grad_norm": 0.37890625, + "learning_rate": 2.376372634553936e-06, + "loss": 0.9475, + "step": 4120 + }, + { + "epoch": 14.078498293515358, + "grad_norm": 0.376953125, + "learning_rate": 2.291073164877511e-06, + "loss": 0.9435, + "step": 4125 + }, + { + "epoch": 14.095563139931741, + "grad_norm": 0.37890625, + "learning_rate": 2.207314971325292e-06, + "loss": 0.9546, + "step": 4130 + }, + { + "epoch": 14.112627986348123, + "grad_norm": 0.400390625, + "learning_rate": 2.125099375112316e-06, + "loss": 0.9496, + "step": 4135 + }, + { + "epoch": 14.129692832764505, + "grad_norm": 0.3671875, + "learning_rate": 2.0444276731204415e-06, + "loss": 0.9592, + "step": 4140 + }, + { + "epoch": 14.146757679180887, + "grad_norm": 0.37890625, + "learning_rate": 1.9653011378779283e-06, + "loss": 0.9446, + "step": 4145 + }, + { + "epoch": 14.16382252559727, + "grad_norm": 0.5625, + "learning_rate": 1.88772101753929e-06, + "loss": 0.9374, + "step": 4150 + }, + { + "epoch": 14.180887372013652, + "grad_norm": 0.37890625, + "learning_rate": 1.8116885358656744e-06, + "loss": 0.9543, + "step": 4155 + }, + { + "epoch": 14.197952218430034, + "grad_norm": 0.37109375, + "learning_rate": 1.7372048922054906e-06, + "loss": 0.9488, + "step": 4160 + }, + { + "epoch": 14.215017064846416, + "grad_norm": 0.373046875, + "learning_rate": 1.6642712614755695e-06, + "loss": 0.9466, + "step": 4165 + }, + { + "epoch": 14.2320819112628, + "grad_norm": 0.396484375, + "learning_rate": 1.5928887941426107e-06, + "loss": 0.9482, + "step": 4170 + }, + { + "epoch": 14.249146757679181, + "grad_norm": 0.373046875, + "learning_rate": 1.523058616204942e-06, + "loss": 0.9449, + "step": 4175 + }, + { + "epoch": 14.266211604095563, + "grad_norm": 0.3984375, + "learning_rate": 1.4547818291749115e-06, + "loss": 0.9562, + "step": 4180 + }, + { + "epoch": 14.283276450511945, + "grad_norm": 0.388671875, + "learning_rate": 1.3880595100613792e-06, + "loss": 0.9445, + "step": 4185 + }, + { + "epoch": 14.300341296928327, + "grad_norm": 0.376953125, + "learning_rate": 1.3228927113528189e-06, + "loss": 0.9457, + "step": 4190 + }, + { + "epoch": 14.31740614334471, + "grad_norm": 0.388671875, + "learning_rate": 1.2592824610006215e-06, + "loss": 0.9488, + "step": 4195 + }, + { + "epoch": 14.334470989761092, + "grad_norm": 0.38671875, + "learning_rate": 1.1972297624030072e-06, + "loss": 0.9437, + "step": 4200 + }, + { + "epoch": 14.351535836177474, + "grad_norm": 0.3671875, + "learning_rate": 1.1367355943890823e-06, + "loss": 0.9459, + "step": 4205 + }, + { + "epoch": 14.368600682593856, + "grad_norm": 0.396484375, + "learning_rate": 1.0778009112034748e-06, + "loss": 0.9477, + "step": 4210 + }, + { + "epoch": 14.38566552901024, + "grad_norm": 0.375, + "learning_rate": 1.0204266424912123e-06, + "loss": 0.95, + "step": 4215 + }, + { + "epoch": 14.402730375426621, + "grad_norm": 0.40625, + "learning_rate": 9.64613693283123e-07, + "loss": 0.9477, + "step": 4220 + }, + { + "epoch": 14.419795221843003, + "grad_norm": 0.375, + "learning_rate": 9.103629439815354e-07, + "loss": 0.9461, + "step": 4225 + }, + { + "epoch": 14.436860068259385, + "grad_norm": 0.3828125, + "learning_rate": 8.57675250346368e-07, + "loss": 0.9585, + "step": 4230 + }, + { + "epoch": 14.453924914675769, + "grad_norm": 0.443359375, + "learning_rate": 8.065514434816845e-07, + "loss": 0.9434, + "step": 4235 + }, + { + "epoch": 14.47098976109215, + "grad_norm": 0.396484375, + "learning_rate": 7.569923298225146e-07, + "loss": 0.941, + "step": 4240 + }, + { + "epoch": 14.488054607508532, + "grad_norm": 0.375, + "learning_rate": 7.08998691122198e-07, + "loss": 0.9527, + "step": 4245 + }, + { + "epoch": 14.505119453924914, + "grad_norm": 0.380859375, + "learning_rate": 6.625712844400056e-07, + "loss": 0.9484, + "step": 4250 + }, + { + "epoch": 14.522184300341298, + "grad_norm": 0.390625, + "learning_rate": 6.177108421292266e-07, + "loss": 0.9453, + "step": 4255 + }, + { + "epoch": 14.53924914675768, + "grad_norm": 0.404296875, + "learning_rate": 5.744180718255776e-07, + "loss": 0.9464, + "step": 4260 + }, + { + "epoch": 14.556313993174061, + "grad_norm": 0.375, + "learning_rate": 5.326936564361118e-07, + "loss": 0.943, + "step": 4265 + }, + { + "epoch": 14.573378839590443, + "grad_norm": 0.369140625, + "learning_rate": 4.92538254128383e-07, + "loss": 0.9422, + "step": 4270 + }, + { + "epoch": 14.590443686006825, + "grad_norm": 0.390625, + "learning_rate": 4.5395249832007604e-07, + "loss": 0.9591, + "step": 4275 + }, + { + "epoch": 14.607508532423209, + "grad_norm": 0.396484375, + "learning_rate": 4.1693699766902626e-07, + "loss": 0.9475, + "step": 4280 + }, + { + "epoch": 14.62457337883959, + "grad_norm": 0.369140625, + "learning_rate": 3.814923360636158e-07, + "loss": 0.9391, + "step": 4285 + }, + { + "epoch": 14.641638225255972, + "grad_norm": 0.36328125, + "learning_rate": 3.4761907261356976e-07, + "loss": 0.9574, + "step": 4290 + }, + { + "epoch": 14.658703071672354, + "grad_norm": 0.388671875, + "learning_rate": 3.1531774164111903e-07, + "loss": 0.9495, + "step": 4295 + }, + { + "epoch": 14.675767918088738, + "grad_norm": 0.373046875, + "learning_rate": 2.8458885267260705e-07, + "loss": 0.9537, + "step": 4300 + }, + { + "epoch": 14.69283276450512, + "grad_norm": 0.38671875, + "learning_rate": 2.554328904303738e-07, + "loss": 0.9435, + "step": 4305 + }, + { + "epoch": 14.709897610921502, + "grad_norm": 0.39453125, + "learning_rate": 2.2785031482521758e-07, + "loss": 0.9474, + "step": 4310 + }, + { + "epoch": 14.726962457337883, + "grad_norm": 0.376953125, + "learning_rate": 2.0184156094905648e-07, + "loss": 0.947, + "step": 4315 + }, + { + "epoch": 14.744027303754265, + "grad_norm": 0.3671875, + "learning_rate": 1.7740703906810042e-07, + "loss": 0.9431, + "step": 4320 + }, + { + "epoch": 14.761092150170649, + "grad_norm": 0.384765625, + "learning_rate": 1.545471346164007e-07, + "loss": 0.9431, + "step": 4325 + }, + { + "epoch": 14.77815699658703, + "grad_norm": 0.37890625, + "learning_rate": 1.3326220818968838e-07, + "loss": 0.9455, + "step": 4330 + }, + { + "epoch": 14.795221843003413, + "grad_norm": 0.375, + "learning_rate": 1.1355259553978981e-07, + "loss": 0.9512, + "step": 4335 + }, + { + "epoch": 14.812286689419794, + "grad_norm": 0.390625, + "learning_rate": 9.541860756925314e-08, + "loss": 0.9439, + "step": 4340 + }, + { + "epoch": 14.829351535836178, + "grad_norm": 0.37109375, + "learning_rate": 7.886053032649665e-08, + "loss": 0.9548, + "step": 4345 + }, + { + "epoch": 14.84641638225256, + "grad_norm": 0.4921875, + "learning_rate": 6.387862500125685e-08, + "loss": 0.9437, + "step": 4350 + }, + { + "epoch": 14.863481228668942, + "grad_norm": 0.380859375, + "learning_rate": 5.047312792046954e-08, + "loss": 0.9512, + "step": 4355 + }, + { + "epoch": 14.880546075085324, + "grad_norm": 0.39453125, + "learning_rate": 3.8644250544594975e-08, + "loss": 0.9478, + "step": 4360 + }, + { + "epoch": 14.897610921501707, + "grad_norm": 0.380859375, + "learning_rate": 2.839217946422057e-08, + "loss": 0.9362, + "step": 4365 + }, + { + "epoch": 14.914675767918089, + "grad_norm": 0.380859375, + "learning_rate": 1.971707639712994e-08, + "loss": 0.9507, + "step": 4370 + }, + { + "epoch": 14.93174061433447, + "grad_norm": 0.37109375, + "learning_rate": 1.2619078185793776e-08, + "loss": 0.948, + "step": 4375 + }, + { + "epoch": 14.948805460750853, + "grad_norm": 0.400390625, + "learning_rate": 7.098296795138293e-09, + "loss": 0.9524, + "step": 4380 + }, + { + "epoch": 14.965870307167236, + "grad_norm": 0.36328125, + "learning_rate": 3.154819310868806e-09, + "loss": 0.9497, + "step": 4385 + }, + { + "epoch": 14.982935153583618, + "grad_norm": 0.400390625, + "learning_rate": 7.887079380153317e-10, + "loss": 0.9536, + "step": 4390 + }, + { + "epoch": 15.0, + "grad_norm": 0.373046875, + "learning_rate": 0.0, + "loss": 0.9444, + "step": 4395 + }, + { + "epoch": 15.0, + "eval_loss": 2.523277521133423, + "eval_runtime": 0.5592, + "eval_samples_per_second": 17.883, + "eval_steps_per_second": 1.788, + "step": 4395 + }, + { + "epoch": 15.0, + "step": 4395, + "total_flos": 2.581505823377195e+18, + "train_loss": 1.0488379673203783, + "train_runtime": 23446.7186, + "train_samples_per_second": 8.983, + "train_steps_per_second": 0.187 + } + ], + "logging_steps": 5, + "max_steps": 4395, + "num_input_tokens_seen": 0, + "num_train_epochs": 15, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.581505823377195e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}