lightnovel_cpt / trainer_state.json
zhangxiaolong
First model init
99e39c9
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9133519673188562,
"eval_steps": 500,
"global_step": 531,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0017200602021070737,
"grad_norm": 9.863033294677734,
"learning_rate": 1.0000000000000001e-07,
"loss": 46.5684,
"step": 1
},
{
"epoch": 0.0034401204042141475,
"grad_norm": 9.501362800598145,
"learning_rate": 2.0000000000000002e-07,
"loss": 46.7594,
"step": 2
},
{
"epoch": 0.005160180606321221,
"grad_norm": 10.306024551391602,
"learning_rate": 3.0000000000000004e-07,
"loss": 47.7285,
"step": 3
},
{
"epoch": 0.006880240808428295,
"grad_norm": 10.35928726196289,
"learning_rate": 4.0000000000000003e-07,
"loss": 46.2416,
"step": 4
},
{
"epoch": 0.008600301010535369,
"grad_norm": 10.284590721130371,
"learning_rate": 5.000000000000001e-07,
"loss": 46.4638,
"step": 5
},
{
"epoch": 0.010320361212642442,
"grad_norm": 9.508557319641113,
"learning_rate": 6.000000000000001e-07,
"loss": 47.0967,
"step": 6
},
{
"epoch": 0.012040421414749517,
"grad_norm": 9.53107738494873,
"learning_rate": 7.000000000000001e-07,
"loss": 46.8945,
"step": 7
},
{
"epoch": 0.01376048161685659,
"grad_norm": 8.917387008666992,
"learning_rate": 8.000000000000001e-07,
"loss": 46.0318,
"step": 8
},
{
"epoch": 0.015480541818963663,
"grad_norm": 9.046576499938965,
"learning_rate": 9.000000000000001e-07,
"loss": 46.7626,
"step": 9
},
{
"epoch": 0.017200602021070738,
"grad_norm": 8.346121788024902,
"learning_rate": 1.0000000000000002e-06,
"loss": 46.4023,
"step": 10
},
{
"epoch": 0.01892066222317781,
"grad_norm": 7.368544578552246,
"learning_rate": 1.1e-06,
"loss": 46.9599,
"step": 11
},
{
"epoch": 0.020640722425284884,
"grad_norm": 7.151632308959961,
"learning_rate": 1.2000000000000002e-06,
"loss": 46.8859,
"step": 12
},
{
"epoch": 0.022360782627391957,
"grad_norm": 6.858343124389648,
"learning_rate": 1.3e-06,
"loss": 47.0505,
"step": 13
},
{
"epoch": 0.024080842829499034,
"grad_norm": 6.4683918952941895,
"learning_rate": 1.4000000000000001e-06,
"loss": 46.6439,
"step": 14
},
{
"epoch": 0.025800903031606107,
"grad_norm": 8.405121803283691,
"learning_rate": 1.5e-06,
"loss": 46.8736,
"step": 15
},
{
"epoch": 0.02752096323371318,
"grad_norm": 10.191279411315918,
"learning_rate": 1.6000000000000001e-06,
"loss": 46.2945,
"step": 16
},
{
"epoch": 0.029241023435820253,
"grad_norm": 10.24738597869873,
"learning_rate": 1.7000000000000002e-06,
"loss": 46.0618,
"step": 17
},
{
"epoch": 0.030961083637927326,
"grad_norm": 9.236139297485352,
"learning_rate": 1.8000000000000001e-06,
"loss": 47.0773,
"step": 18
},
{
"epoch": 0.0326811438400344,
"grad_norm": 8.437068939208984,
"learning_rate": 1.9000000000000002e-06,
"loss": 46.6797,
"step": 19
},
{
"epoch": 0.034401204042141476,
"grad_norm": 8.35542106628418,
"learning_rate": 2.0000000000000003e-06,
"loss": 46.8348,
"step": 20
},
{
"epoch": 0.03612126424424855,
"grad_norm": 9.210644721984863,
"learning_rate": 2.1000000000000002e-06,
"loss": 46.7521,
"step": 21
},
{
"epoch": 0.03784132444635562,
"grad_norm": 9.216105461120605,
"learning_rate": 2.2e-06,
"loss": 46.4541,
"step": 22
},
{
"epoch": 0.0395613846484627,
"grad_norm": 8.33311939239502,
"learning_rate": 2.3000000000000004e-06,
"loss": 46.7192,
"step": 23
},
{
"epoch": 0.04128144485056977,
"grad_norm": 7.9267449378967285,
"learning_rate": 2.4000000000000003e-06,
"loss": 46.7559,
"step": 24
},
{
"epoch": 0.043001505052676844,
"grad_norm": 7.363073348999023,
"learning_rate": 2.5e-06,
"loss": 45.9578,
"step": 25
},
{
"epoch": 0.044721565254783914,
"grad_norm": 7.073836326599121,
"learning_rate": 2.6e-06,
"loss": 45.9155,
"step": 26
},
{
"epoch": 0.04644162545689099,
"grad_norm": 7.0409088134765625,
"learning_rate": 2.7000000000000004e-06,
"loss": 46.6551,
"step": 27
},
{
"epoch": 0.04816168565899807,
"grad_norm": 6.685385227203369,
"learning_rate": 2.8000000000000003e-06,
"loss": 46.3889,
"step": 28
},
{
"epoch": 0.04988174586110514,
"grad_norm": 6.540524959564209,
"learning_rate": 2.9e-06,
"loss": 47.0662,
"step": 29
},
{
"epoch": 0.05160180606321221,
"grad_norm": 6.627730846405029,
"learning_rate": 3e-06,
"loss": 46.101,
"step": 30
},
{
"epoch": 0.05332186626531928,
"grad_norm": 7.054595470428467,
"learning_rate": 3.1000000000000004e-06,
"loss": 46.7379,
"step": 31
},
{
"epoch": 0.05504192646742636,
"grad_norm": 6.73618221282959,
"learning_rate": 3.2000000000000003e-06,
"loss": 46.3172,
"step": 32
},
{
"epoch": 0.056761986669533436,
"grad_norm": 5.943539619445801,
"learning_rate": 3.3000000000000006e-06,
"loss": 47.0245,
"step": 33
},
{
"epoch": 0.058482046871640506,
"grad_norm": 6.056912899017334,
"learning_rate": 3.4000000000000005e-06,
"loss": 46.0208,
"step": 34
},
{
"epoch": 0.06020210707374758,
"grad_norm": 5.7798309326171875,
"learning_rate": 3.5e-06,
"loss": 46.0746,
"step": 35
},
{
"epoch": 0.06192216727585465,
"grad_norm": 5.896692276000977,
"learning_rate": 3.6000000000000003e-06,
"loss": 46.4911,
"step": 36
},
{
"epoch": 0.06364222747796174,
"grad_norm": 5.496098518371582,
"learning_rate": 3.7e-06,
"loss": 45.9957,
"step": 37
},
{
"epoch": 0.0653622876800688,
"grad_norm": 5.253308296203613,
"learning_rate": 3.8000000000000005e-06,
"loss": 45.7145,
"step": 38
},
{
"epoch": 0.06708234788217587,
"grad_norm": 5.340756416320801,
"learning_rate": 3.900000000000001e-06,
"loss": 46.7068,
"step": 39
},
{
"epoch": 0.06880240808428295,
"grad_norm": 5.312371730804443,
"learning_rate": 4.000000000000001e-06,
"loss": 46.4172,
"step": 40
},
{
"epoch": 0.07052246828639003,
"grad_norm": 5.484511375427246,
"learning_rate": 4.1e-06,
"loss": 45.6433,
"step": 41
},
{
"epoch": 0.0722425284884971,
"grad_norm": 5.260024547576904,
"learning_rate": 4.2000000000000004e-06,
"loss": 46.1259,
"step": 42
},
{
"epoch": 0.07396258869060417,
"grad_norm": 5.4440999031066895,
"learning_rate": 4.3e-06,
"loss": 46.2947,
"step": 43
},
{
"epoch": 0.07568264889271124,
"grad_norm": 5.153471946716309,
"learning_rate": 4.4e-06,
"loss": 46.2882,
"step": 44
},
{
"epoch": 0.07740270909481832,
"grad_norm": 5.6796488761901855,
"learning_rate": 4.5e-06,
"loss": 46.3314,
"step": 45
},
{
"epoch": 0.0791227692969254,
"grad_norm": 5.253461837768555,
"learning_rate": 4.600000000000001e-06,
"loss": 46.4205,
"step": 46
},
{
"epoch": 0.08084282949903247,
"grad_norm": 5.475714683532715,
"learning_rate": 4.7e-06,
"loss": 45.9359,
"step": 47
},
{
"epoch": 0.08256288970113954,
"grad_norm": 5.880488872528076,
"learning_rate": 4.800000000000001e-06,
"loss": 46.0665,
"step": 48
},
{
"epoch": 0.08428294990324661,
"grad_norm": 5.69149923324585,
"learning_rate": 4.9000000000000005e-06,
"loss": 45.8811,
"step": 49
},
{
"epoch": 0.08600301010535369,
"grad_norm": 5.086974143981934,
"learning_rate": 5e-06,
"loss": 46.1759,
"step": 50
},
{
"epoch": 0.08772307030746077,
"grad_norm": 5.902041435241699,
"learning_rate": 5.1e-06,
"loss": 46.4326,
"step": 51
},
{
"epoch": 0.08944313050956783,
"grad_norm": 5.740163803100586,
"learning_rate": 5.2e-06,
"loss": 45.607,
"step": 52
},
{
"epoch": 0.0911631907116749,
"grad_norm": 5.547687530517578,
"learning_rate": 5.300000000000001e-06,
"loss": 45.8864,
"step": 53
},
{
"epoch": 0.09288325091378198,
"grad_norm": 6.20143461227417,
"learning_rate": 5.400000000000001e-06,
"loss": 45.7776,
"step": 54
},
{
"epoch": 0.09460331111588906,
"grad_norm": 5.544395446777344,
"learning_rate": 5.500000000000001e-06,
"loss": 46.3334,
"step": 55
},
{
"epoch": 0.09632337131799613,
"grad_norm": 5.583694934844971,
"learning_rate": 5.600000000000001e-06,
"loss": 45.5733,
"step": 56
},
{
"epoch": 0.0980434315201032,
"grad_norm": 5.751343250274658,
"learning_rate": 5.7e-06,
"loss": 45.2805,
"step": 57
},
{
"epoch": 0.09976349172221027,
"grad_norm": 6.024663925170898,
"learning_rate": 5.8e-06,
"loss": 45.6569,
"step": 58
},
{
"epoch": 0.10148355192431735,
"grad_norm": 5.834673881530762,
"learning_rate": 5.9e-06,
"loss": 45.7854,
"step": 59
},
{
"epoch": 0.10320361212642443,
"grad_norm": 6.796127796173096,
"learning_rate": 6e-06,
"loss": 46.4184,
"step": 60
},
{
"epoch": 0.1049236723285315,
"grad_norm": 6.303890705108643,
"learning_rate": 6.1e-06,
"loss": 46.0877,
"step": 61
},
{
"epoch": 0.10664373253063857,
"grad_norm": 6.320569038391113,
"learning_rate": 6.200000000000001e-06,
"loss": 45.8802,
"step": 62
},
{
"epoch": 0.10836379273274564,
"grad_norm": 6.015869617462158,
"learning_rate": 6.300000000000001e-06,
"loss": 45.6733,
"step": 63
},
{
"epoch": 0.11008385293485272,
"grad_norm": 6.314846515655518,
"learning_rate": 6.4000000000000006e-06,
"loss": 45.8531,
"step": 64
},
{
"epoch": 0.1118039131369598,
"grad_norm": 6.106888294219971,
"learning_rate": 6.5000000000000004e-06,
"loss": 45.2966,
"step": 65
},
{
"epoch": 0.11352397333906687,
"grad_norm": 6.093003273010254,
"learning_rate": 6.600000000000001e-06,
"loss": 46.0552,
"step": 66
},
{
"epoch": 0.11524403354117393,
"grad_norm": 5.634897708892822,
"learning_rate": 6.700000000000001e-06,
"loss": 45.3367,
"step": 67
},
{
"epoch": 0.11696409374328101,
"grad_norm": 6.261721134185791,
"learning_rate": 6.800000000000001e-06,
"loss": 45.7655,
"step": 68
},
{
"epoch": 0.11868415394538809,
"grad_norm": 5.926329135894775,
"learning_rate": 6.9e-06,
"loss": 45.1391,
"step": 69
},
{
"epoch": 0.12040421414749516,
"grad_norm": 5.887923717498779,
"learning_rate": 7e-06,
"loss": 45.43,
"step": 70
},
{
"epoch": 0.12212427434960224,
"grad_norm": 5.79194450378418,
"learning_rate": 7.100000000000001e-06,
"loss": 45.763,
"step": 71
},
{
"epoch": 0.1238443345517093,
"grad_norm": 6.2670722007751465,
"learning_rate": 7.2000000000000005e-06,
"loss": 45.137,
"step": 72
},
{
"epoch": 0.1255643947538164,
"grad_norm": 5.60247278213501,
"learning_rate": 7.3e-06,
"loss": 45.7123,
"step": 73
},
{
"epoch": 0.12728445495592347,
"grad_norm": 7.346188545227051,
"learning_rate": 7.4e-06,
"loss": 45.5004,
"step": 74
},
{
"epoch": 0.12900451515803052,
"grad_norm": 6.078243255615234,
"learning_rate": 7.500000000000001e-06,
"loss": 45.5731,
"step": 75
},
{
"epoch": 0.1307245753601376,
"grad_norm": 7.248181343078613,
"learning_rate": 7.600000000000001e-06,
"loss": 45.7506,
"step": 76
},
{
"epoch": 0.13244463556224467,
"grad_norm": 5.837612628936768,
"learning_rate": 7.7e-06,
"loss": 45.7486,
"step": 77
},
{
"epoch": 0.13416469576435175,
"grad_norm": 7.661535739898682,
"learning_rate": 7.800000000000002e-06,
"loss": 45.5743,
"step": 78
},
{
"epoch": 0.13588475596645883,
"grad_norm": 5.925168991088867,
"learning_rate": 7.9e-06,
"loss": 45.0878,
"step": 79
},
{
"epoch": 0.1376048161685659,
"grad_norm": 6.47566032409668,
"learning_rate": 8.000000000000001e-06,
"loss": 45.4567,
"step": 80
},
{
"epoch": 0.13932487637067298,
"grad_norm": 5.9641852378845215,
"learning_rate": 8.1e-06,
"loss": 45.0619,
"step": 81
},
{
"epoch": 0.14104493657278006,
"grad_norm": 6.4052324295043945,
"learning_rate": 8.2e-06,
"loss": 45.9375,
"step": 82
},
{
"epoch": 0.14276499677488713,
"grad_norm": 6.149839401245117,
"learning_rate": 8.3e-06,
"loss": 45.5695,
"step": 83
},
{
"epoch": 0.1444850569769942,
"grad_norm": 6.546520233154297,
"learning_rate": 8.400000000000001e-06,
"loss": 45.8453,
"step": 84
},
{
"epoch": 0.14620511717910126,
"grad_norm": 6.221002578735352,
"learning_rate": 8.5e-06,
"loss": 44.9717,
"step": 85
},
{
"epoch": 0.14792517738120833,
"grad_norm": 6.772427558898926,
"learning_rate": 8.6e-06,
"loss": 45.8218,
"step": 86
},
{
"epoch": 0.1496452375833154,
"grad_norm": 7.193835258483887,
"learning_rate": 8.700000000000001e-06,
"loss": 45.4445,
"step": 87
},
{
"epoch": 0.1513652977854225,
"grad_norm": 6.1274895668029785,
"learning_rate": 8.8e-06,
"loss": 45.127,
"step": 88
},
{
"epoch": 0.15308535798752956,
"grad_norm": 7.984500885009766,
"learning_rate": 8.900000000000001e-06,
"loss": 44.9928,
"step": 89
},
{
"epoch": 0.15480541818963664,
"grad_norm": 6.346322059631348,
"learning_rate": 9e-06,
"loss": 45.4616,
"step": 90
},
{
"epoch": 0.15652547839174372,
"grad_norm": 7.460362911224365,
"learning_rate": 9.100000000000001e-06,
"loss": 45.2488,
"step": 91
},
{
"epoch": 0.1582455385938508,
"grad_norm": 7.510838508605957,
"learning_rate": 9.200000000000002e-06,
"loss": 45.0087,
"step": 92
},
{
"epoch": 0.15996559879595787,
"grad_norm": 6.567174911499023,
"learning_rate": 9.3e-06,
"loss": 44.3216,
"step": 93
},
{
"epoch": 0.16168565899806495,
"grad_norm": 8.357160568237305,
"learning_rate": 9.4e-06,
"loss": 45.1102,
"step": 94
},
{
"epoch": 0.163405719200172,
"grad_norm": 6.190718650817871,
"learning_rate": 9.5e-06,
"loss": 45.0187,
"step": 95
},
{
"epoch": 0.16512577940227907,
"grad_norm": 6.47848653793335,
"learning_rate": 9.600000000000001e-06,
"loss": 45.9968,
"step": 96
},
{
"epoch": 0.16684583960438615,
"grad_norm": 6.271050453186035,
"learning_rate": 9.7e-06,
"loss": 45.6172,
"step": 97
},
{
"epoch": 0.16856589980649322,
"grad_norm": 7.013180255889893,
"learning_rate": 9.800000000000001e-06,
"loss": 45.1628,
"step": 98
},
{
"epoch": 0.1702859600086003,
"grad_norm": 6.151601314544678,
"learning_rate": 9.9e-06,
"loss": 45.3331,
"step": 99
},
{
"epoch": 0.17200602021070738,
"grad_norm": 6.522064685821533,
"learning_rate": 1e-05,
"loss": 45.01,
"step": 100
},
{
"epoch": 0.17372608041281445,
"grad_norm": 6.936015605926514,
"learning_rate": 9.999990859614876e-06,
"loss": 44.7838,
"step": 101
},
{
"epoch": 0.17544614061492153,
"grad_norm": 6.696622371673584,
"learning_rate": 9.99996343849292e-06,
"loss": 44.4096,
"step": 102
},
{
"epoch": 0.1771662008170286,
"grad_norm": 6.770718574523926,
"learning_rate": 9.999917736734387e-06,
"loss": 44.6306,
"step": 103
},
{
"epoch": 0.17888626101913566,
"grad_norm": 8.121861457824707,
"learning_rate": 9.999853754506375e-06,
"loss": 45.3421,
"step": 104
},
{
"epoch": 0.18060632122124273,
"grad_norm": 7.101470947265625,
"learning_rate": 9.999771492042807e-06,
"loss": 45.415,
"step": 105
},
{
"epoch": 0.1823263814233498,
"grad_norm": 8.22966194152832,
"learning_rate": 9.99967094964445e-06,
"loss": 45.3149,
"step": 106
},
{
"epoch": 0.18404644162545689,
"grad_norm": 10.975841522216797,
"learning_rate": 9.9995521276789e-06,
"loss": 45.2939,
"step": 107
},
{
"epoch": 0.18576650182756396,
"grad_norm": 7.771969795227051,
"learning_rate": 9.999415026580592e-06,
"loss": 45.2433,
"step": 108
},
{
"epoch": 0.18748656202967104,
"grad_norm": 10.387533187866211,
"learning_rate": 9.999259646850787e-06,
"loss": 45.2594,
"step": 109
},
{
"epoch": 0.18920662223177812,
"grad_norm": 8.546263694763184,
"learning_rate": 9.999085989057578e-06,
"loss": 45.0587,
"step": 110
},
{
"epoch": 0.1909266824338852,
"grad_norm": 9.245797157287598,
"learning_rate": 9.998894053835883e-06,
"loss": 45.2273,
"step": 111
},
{
"epoch": 0.19264674263599227,
"grad_norm": 9.295574188232422,
"learning_rate": 9.998683841887449e-06,
"loss": 45.5471,
"step": 112
},
{
"epoch": 0.19436680283809935,
"grad_norm": 8.650074005126953,
"learning_rate": 9.99845535398084e-06,
"loss": 45.2802,
"step": 113
},
{
"epoch": 0.1960868630402064,
"grad_norm": 11.102032661437988,
"learning_rate": 9.998208590951449e-06,
"loss": 45.3066,
"step": 114
},
{
"epoch": 0.19780692324231347,
"grad_norm": 6.938677787780762,
"learning_rate": 9.99794355370147e-06,
"loss": 45.1454,
"step": 115
},
{
"epoch": 0.19952698344442055,
"grad_norm": 16.639892578125,
"learning_rate": 9.997660243199928e-06,
"loss": 45.6125,
"step": 116
},
{
"epoch": 0.20124704364652762,
"grad_norm": 11.0328369140625,
"learning_rate": 9.99735866048265e-06,
"loss": 45.4087,
"step": 117
},
{
"epoch": 0.2029671038486347,
"grad_norm": 13.419515609741211,
"learning_rate": 9.997038806652264e-06,
"loss": 45.8139,
"step": 118
},
{
"epoch": 0.20468716405074178,
"grad_norm": 13.69460678100586,
"learning_rate": 9.996700682878206e-06,
"loss": 45.304,
"step": 119
},
{
"epoch": 0.20640722425284885,
"grad_norm": 7.643240928649902,
"learning_rate": 9.996344290396713e-06,
"loss": 45.1699,
"step": 120
},
{
"epoch": 0.20812728445495593,
"grad_norm": 9.692140579223633,
"learning_rate": 9.995969630510805e-06,
"loss": 44.7309,
"step": 121
},
{
"epoch": 0.209847344657063,
"grad_norm": 7.550359725952148,
"learning_rate": 9.995576704590299e-06,
"loss": 44.6299,
"step": 122
},
{
"epoch": 0.21156740485917008,
"grad_norm": 6.425361156463623,
"learning_rate": 9.995165514071793e-06,
"loss": 44.8296,
"step": 123
},
{
"epoch": 0.21328746506127713,
"grad_norm": 8.089837074279785,
"learning_rate": 9.994736060458665e-06,
"loss": 45.2622,
"step": 124
},
{
"epoch": 0.2150075252633842,
"grad_norm": 6.491065979003906,
"learning_rate": 9.994288345321059e-06,
"loss": 44.3288,
"step": 125
},
{
"epoch": 0.21672758546549128,
"grad_norm": 5.423704147338867,
"learning_rate": 9.993822370295892e-06,
"loss": 44.7233,
"step": 126
},
{
"epoch": 0.21844764566759836,
"grad_norm": 7.901766777038574,
"learning_rate": 9.993338137086841e-06,
"loss": 45.0159,
"step": 127
},
{
"epoch": 0.22016770586970544,
"grad_norm": 7.50944709777832,
"learning_rate": 9.992835647464339e-06,
"loss": 44.8785,
"step": 128
},
{
"epoch": 0.22188776607181251,
"grad_norm": 7.400681972503662,
"learning_rate": 9.992314903265561e-06,
"loss": 45.3053,
"step": 129
},
{
"epoch": 0.2236078262739196,
"grad_norm": 8.214972496032715,
"learning_rate": 9.991775906394434e-06,
"loss": 44.8066,
"step": 130
},
{
"epoch": 0.22532788647602667,
"grad_norm": 7.00181770324707,
"learning_rate": 9.991218658821609e-06,
"loss": 44.5013,
"step": 131
},
{
"epoch": 0.22704794667813374,
"grad_norm": 6.811863422393799,
"learning_rate": 9.990643162584467e-06,
"loss": 44.7327,
"step": 132
},
{
"epoch": 0.22876800688024082,
"grad_norm": 8.760518074035645,
"learning_rate": 9.99004941978712e-06,
"loss": 44.6706,
"step": 133
},
{
"epoch": 0.23048806708234787,
"grad_norm": 7.204503536224365,
"learning_rate": 9.989437432600373e-06,
"loss": 44.8113,
"step": 134
},
{
"epoch": 0.23220812728445495,
"grad_norm": 8.635207176208496,
"learning_rate": 9.988807203261752e-06,
"loss": 45.3002,
"step": 135
},
{
"epoch": 0.23392818748656202,
"grad_norm": 7.5076823234558105,
"learning_rate": 9.988158734075468e-06,
"loss": 44.919,
"step": 136
},
{
"epoch": 0.2356482476886691,
"grad_norm": 8.329002380371094,
"learning_rate": 9.98749202741243e-06,
"loss": 44.9676,
"step": 137
},
{
"epoch": 0.23736830789077618,
"grad_norm": 7.5155158042907715,
"learning_rate": 9.986807085710213e-06,
"loss": 45.167,
"step": 138
},
{
"epoch": 0.23908836809288325,
"grad_norm": 8.341093063354492,
"learning_rate": 9.986103911473075e-06,
"loss": 45.1884,
"step": 139
},
{
"epoch": 0.24080842829499033,
"grad_norm": 7.4146342277526855,
"learning_rate": 9.985382507271928e-06,
"loss": 44.8973,
"step": 140
},
{
"epoch": 0.2425284884970974,
"grad_norm": 8.28603458404541,
"learning_rate": 9.984642875744338e-06,
"loss": 44.6793,
"step": 141
},
{
"epoch": 0.24424854869920448,
"grad_norm": 7.7817583084106445,
"learning_rate": 9.983885019594506e-06,
"loss": 44.7361,
"step": 142
},
{
"epoch": 0.24596860890131156,
"grad_norm": 8.71711254119873,
"learning_rate": 9.983108941593277e-06,
"loss": 44.514,
"step": 143
},
{
"epoch": 0.2476886691034186,
"grad_norm": 8.746935844421387,
"learning_rate": 9.982314644578111e-06,
"loss": 45.4915,
"step": 144
},
{
"epoch": 0.24940872930552568,
"grad_norm": 8.616266250610352,
"learning_rate": 9.981502131453077e-06,
"loss": 44.4955,
"step": 145
},
{
"epoch": 0.2511287895076328,
"grad_norm": 8.73255443572998,
"learning_rate": 9.980671405188852e-06,
"loss": 44.7087,
"step": 146
},
{
"epoch": 0.25284884970973986,
"grad_norm": 7.063310623168945,
"learning_rate": 9.979822468822696e-06,
"loss": 44.7221,
"step": 147
},
{
"epoch": 0.25456890991184694,
"grad_norm": 8.635971069335938,
"learning_rate": 9.978955325458453e-06,
"loss": 44.5993,
"step": 148
},
{
"epoch": 0.25628897011395396,
"grad_norm": 7.771418571472168,
"learning_rate": 9.978069978266534e-06,
"loss": 45.1399,
"step": 149
},
{
"epoch": 0.25800903031606104,
"grad_norm": 9.317761421203613,
"learning_rate": 9.9771664304839e-06,
"loss": 44.8672,
"step": 150
},
{
"epoch": 0.2597290905181681,
"grad_norm": 8.102662086486816,
"learning_rate": 9.976244685414065e-06,
"loss": 45.4121,
"step": 151
},
{
"epoch": 0.2614491507202752,
"grad_norm": 12.453920364379883,
"learning_rate": 9.97530474642707e-06,
"loss": 44.6091,
"step": 152
},
{
"epoch": 0.26316921092238227,
"grad_norm": 9.156878471374512,
"learning_rate": 9.974346616959476e-06,
"loss": 45.1213,
"step": 153
},
{
"epoch": 0.26488927112448934,
"grad_norm": 11.89529037475586,
"learning_rate": 9.973370300514353e-06,
"loss": 44.5162,
"step": 154
},
{
"epoch": 0.2666093313265964,
"grad_norm": 9.946409225463867,
"learning_rate": 9.972375800661264e-06,
"loss": 44.106,
"step": 155
},
{
"epoch": 0.2683293915287035,
"grad_norm": 10.854037284851074,
"learning_rate": 9.971363121036252e-06,
"loss": 44.9188,
"step": 156
},
{
"epoch": 0.2700494517308106,
"grad_norm": 9.196579933166504,
"learning_rate": 9.970332265341833e-06,
"loss": 44.6235,
"step": 157
},
{
"epoch": 0.27176951193291765,
"grad_norm": 11.79194164276123,
"learning_rate": 9.969283237346973e-06,
"loss": 45.7572,
"step": 158
},
{
"epoch": 0.2734895721350247,
"grad_norm": 7.789967060089111,
"learning_rate": 9.968216040887078e-06,
"loss": 44.8069,
"step": 159
},
{
"epoch": 0.2752096323371318,
"grad_norm": 10.540818214416504,
"learning_rate": 9.967130679863984e-06,
"loss": 44.7341,
"step": 160
},
{
"epoch": 0.2769296925392389,
"grad_norm": 7.4835405349731445,
"learning_rate": 9.966027158245939e-06,
"loss": 45.4445,
"step": 161
},
{
"epoch": 0.27864975274134596,
"grad_norm": 9.472413063049316,
"learning_rate": 9.964905480067585e-06,
"loss": 44.5498,
"step": 162
},
{
"epoch": 0.28036981294345303,
"grad_norm": 9.155404090881348,
"learning_rate": 9.963765649429954e-06,
"loss": 44.3932,
"step": 163
},
{
"epoch": 0.2820898731455601,
"grad_norm": 8.14942455291748,
"learning_rate": 9.962607670500442e-06,
"loss": 44.5368,
"step": 164
},
{
"epoch": 0.2838099333476672,
"grad_norm": 10.103939056396484,
"learning_rate": 9.961431547512794e-06,
"loss": 44.1877,
"step": 165
},
{
"epoch": 0.28552999354977426,
"grad_norm": 7.972369194030762,
"learning_rate": 9.960237284767103e-06,
"loss": 44.775,
"step": 166
},
{
"epoch": 0.28725005375188134,
"grad_norm": 13.627598762512207,
"learning_rate": 9.959024886629772e-06,
"loss": 44.3529,
"step": 167
},
{
"epoch": 0.2889701139539884,
"grad_norm": 9.467161178588867,
"learning_rate": 9.957794357533518e-06,
"loss": 44.7574,
"step": 168
},
{
"epoch": 0.29069017415609544,
"grad_norm": 12.242958068847656,
"learning_rate": 9.956545701977347e-06,
"loss": 44.5439,
"step": 169
},
{
"epoch": 0.2924102343582025,
"grad_norm": 10.142478942871094,
"learning_rate": 9.955278924526532e-06,
"loss": 44.4951,
"step": 170
},
{
"epoch": 0.2941302945603096,
"grad_norm": 12.610379219055176,
"learning_rate": 9.95399402981261e-06,
"loss": 44.7777,
"step": 171
},
{
"epoch": 0.29585035476241667,
"grad_norm": 9.542120933532715,
"learning_rate": 9.952691022533352e-06,
"loss": 44.5877,
"step": 172
},
{
"epoch": 0.29757041496452374,
"grad_norm": 13.538004875183105,
"learning_rate": 9.951369907452752e-06,
"loss": 44.4161,
"step": 173
},
{
"epoch": 0.2992904751666308,
"grad_norm": 11.80547046661377,
"learning_rate": 9.950030689401014e-06,
"loss": 44.8214,
"step": 174
},
{
"epoch": 0.3010105353687379,
"grad_norm": 11.983033180236816,
"learning_rate": 9.948673373274523e-06,
"loss": 44.7073,
"step": 175
},
{
"epoch": 0.302730595570845,
"grad_norm": 10.844096183776855,
"learning_rate": 9.947297964035837e-06,
"loss": 44.4097,
"step": 176
},
{
"epoch": 0.30445065577295205,
"grad_norm": 11.820161819458008,
"learning_rate": 9.94590446671366e-06,
"loss": 44.7555,
"step": 177
},
{
"epoch": 0.3061707159750591,
"grad_norm": 11.587363243103027,
"learning_rate": 9.94449288640284e-06,
"loss": 44.2532,
"step": 178
},
{
"epoch": 0.3078907761771662,
"grad_norm": 10.775352478027344,
"learning_rate": 9.943063228264327e-06,
"loss": 44.8222,
"step": 179
},
{
"epoch": 0.3096108363792733,
"grad_norm": 12.077004432678223,
"learning_rate": 9.941615497525172e-06,
"loss": 45.0694,
"step": 180
},
{
"epoch": 0.31133089658138036,
"grad_norm": 9.174921035766602,
"learning_rate": 9.940149699478502e-06,
"loss": 44.4424,
"step": 181
},
{
"epoch": 0.31305095678348743,
"grad_norm": 9.949912071228027,
"learning_rate": 9.938665839483503e-06,
"loss": 44.82,
"step": 182
},
{
"epoch": 0.3147710169855945,
"grad_norm": 10.341442108154297,
"learning_rate": 9.937163922965394e-06,
"loss": 43.993,
"step": 183
},
{
"epoch": 0.3164910771877016,
"grad_norm": 7.373370170593262,
"learning_rate": 9.93564395541541e-06,
"loss": 44.6727,
"step": 184
},
{
"epoch": 0.31821113738980866,
"grad_norm": 11.212616920471191,
"learning_rate": 9.93410594239079e-06,
"loss": 44.3283,
"step": 185
},
{
"epoch": 0.31993119759191574,
"grad_norm": 8.05331802368164,
"learning_rate": 9.932549889514747e-06,
"loss": 44.6646,
"step": 186
},
{
"epoch": 0.3216512577940228,
"grad_norm": 11.395249366760254,
"learning_rate": 9.930975802476448e-06,
"loss": 43.9711,
"step": 187
},
{
"epoch": 0.3233713179961299,
"grad_norm": 7.730330944061279,
"learning_rate": 9.929383687030999e-06,
"loss": 45.2283,
"step": 188
},
{
"epoch": 0.3250913781982369,
"grad_norm": 11.927477836608887,
"learning_rate": 9.927773548999419e-06,
"loss": 44.723,
"step": 189
},
{
"epoch": 0.326811438400344,
"grad_norm": 8.933055877685547,
"learning_rate": 9.92614539426862e-06,
"loss": 44.5058,
"step": 190
},
{
"epoch": 0.32853149860245107,
"grad_norm": 13.985485076904297,
"learning_rate": 9.924499228791387e-06,
"loss": 44.8917,
"step": 191
},
{
"epoch": 0.33025155880455814,
"grad_norm": 13.561887741088867,
"learning_rate": 9.922835058586353e-06,
"loss": 44.6659,
"step": 192
},
{
"epoch": 0.3319716190066652,
"grad_norm": 8.802722930908203,
"learning_rate": 9.921152889737985e-06,
"loss": 43.9996,
"step": 193
},
{
"epoch": 0.3336916792087723,
"grad_norm": 10.387024879455566,
"learning_rate": 9.919452728396548e-06,
"loss": 44.7691,
"step": 194
},
{
"epoch": 0.3354117394108794,
"grad_norm": 8.743358612060547,
"learning_rate": 9.917734580778094e-06,
"loss": 45.2034,
"step": 195
},
{
"epoch": 0.33713179961298645,
"grad_norm": 7.972978591918945,
"learning_rate": 9.915998453164435e-06,
"loss": 45.5303,
"step": 196
},
{
"epoch": 0.3388518598150935,
"grad_norm": 8.630827903747559,
"learning_rate": 9.914244351903122e-06,
"loss": 44.7464,
"step": 197
},
{
"epoch": 0.3405719200172006,
"grad_norm": 6.4518609046936035,
"learning_rate": 9.912472283407421e-06,
"loss": 44.2976,
"step": 198
},
{
"epoch": 0.3422919802193077,
"grad_norm": 8.886955261230469,
"learning_rate": 9.910682254156284e-06,
"loss": 44.8556,
"step": 199
},
{
"epoch": 0.34401204042141476,
"grad_norm": 7.261415004730225,
"learning_rate": 9.908874270694337e-06,
"loss": 44.466,
"step": 200
},
{
"epoch": 0.34573210062352183,
"grad_norm": 8.031500816345215,
"learning_rate": 9.907048339631843e-06,
"loss": 44.4753,
"step": 201
},
{
"epoch": 0.3474521608256289,
"grad_norm": 7.2853569984436035,
"learning_rate": 9.905204467644688e-06,
"loss": 45.0669,
"step": 202
},
{
"epoch": 0.349172221027736,
"grad_norm": 8.290436744689941,
"learning_rate": 9.903342661474355e-06,
"loss": 44.7901,
"step": 203
},
{
"epoch": 0.35089228122984306,
"grad_norm": 9.512991905212402,
"learning_rate": 9.901462927927891e-06,
"loss": 44.7863,
"step": 204
},
{
"epoch": 0.35261234143195014,
"grad_norm": 8.04033374786377,
"learning_rate": 9.899565273877892e-06,
"loss": 44.9776,
"step": 205
},
{
"epoch": 0.3543324016340572,
"grad_norm": 11.140141487121582,
"learning_rate": 9.897649706262474e-06,
"loss": 45.1048,
"step": 206
},
{
"epoch": 0.3560524618361643,
"grad_norm": 8.036794662475586,
"learning_rate": 9.895716232085247e-06,
"loss": 44.763,
"step": 207
},
{
"epoch": 0.3577725220382713,
"grad_norm": 11.508170127868652,
"learning_rate": 9.89376485841529e-06,
"loss": 44.7094,
"step": 208
},
{
"epoch": 0.3594925822403784,
"grad_norm": 8.577386856079102,
"learning_rate": 9.891795592387127e-06,
"loss": 44.5892,
"step": 209
},
{
"epoch": 0.36121264244248547,
"grad_norm": 11.701868057250977,
"learning_rate": 9.889808441200697e-06,
"loss": 44.4664,
"step": 210
},
{
"epoch": 0.36293270264459254,
"grad_norm": 7.955048561096191,
"learning_rate": 9.887803412121331e-06,
"loss": 44.5424,
"step": 211
},
{
"epoch": 0.3646527628466996,
"grad_norm": 11.340240478515625,
"learning_rate": 9.885780512479725e-06,
"loss": 44.3322,
"step": 212
},
{
"epoch": 0.3663728230488067,
"grad_norm": 8.020219802856445,
"learning_rate": 9.88373974967191e-06,
"loss": 44.2046,
"step": 213
},
{
"epoch": 0.36809288325091377,
"grad_norm": 10.230839729309082,
"learning_rate": 9.881681131159232e-06,
"loss": 44.2038,
"step": 214
},
{
"epoch": 0.36981294345302085,
"grad_norm": 8.711820602416992,
"learning_rate": 9.879604664468315e-06,
"loss": 44.0065,
"step": 215
},
{
"epoch": 0.3715330036551279,
"grad_norm": 10.792113304138184,
"learning_rate": 9.877510357191042e-06,
"loss": 44.3272,
"step": 216
},
{
"epoch": 0.373253063857235,
"grad_norm": 8.520330429077148,
"learning_rate": 9.875398216984521e-06,
"loss": 44.403,
"step": 217
},
{
"epoch": 0.3749731240593421,
"grad_norm": 12.095443725585938,
"learning_rate": 9.873268251571065e-06,
"loss": 44.8801,
"step": 218
},
{
"epoch": 0.37669318426144915,
"grad_norm": 10.553284645080566,
"learning_rate": 9.871120468738156e-06,
"loss": 44.1855,
"step": 219
},
{
"epoch": 0.37841324446355623,
"grad_norm": 10.677011489868164,
"learning_rate": 9.868954876338414e-06,
"loss": 44.9765,
"step": 220
},
{
"epoch": 0.3801333046656633,
"grad_norm": 10.681044578552246,
"learning_rate": 9.866771482289585e-06,
"loss": 44.3767,
"step": 221
},
{
"epoch": 0.3818533648677704,
"grad_norm": 8.905034065246582,
"learning_rate": 9.86457029457449e-06,
"loss": 44.5735,
"step": 222
},
{
"epoch": 0.38357342506987746,
"grad_norm": 11.448368072509766,
"learning_rate": 9.86235132124101e-06,
"loss": 44.3242,
"step": 223
},
{
"epoch": 0.38529348527198454,
"grad_norm": 8.515959739685059,
"learning_rate": 9.860114570402055e-06,
"loss": 45.0004,
"step": 224
},
{
"epoch": 0.3870135454740916,
"grad_norm": 11.419853210449219,
"learning_rate": 9.85786005023553e-06,
"loss": 44.7329,
"step": 225
},
{
"epoch": 0.3887336056761987,
"grad_norm": 10.274301528930664,
"learning_rate": 9.855587768984308e-06,
"loss": 44.5693,
"step": 226
},
{
"epoch": 0.39045366587830577,
"grad_norm": 9.408884048461914,
"learning_rate": 9.8532977349562e-06,
"loss": 44.2397,
"step": 227
},
{
"epoch": 0.3921737260804128,
"grad_norm": 8.234622955322266,
"learning_rate": 9.850989956523922e-06,
"loss": 44.3023,
"step": 228
},
{
"epoch": 0.39389378628251986,
"grad_norm": 10.358318328857422,
"learning_rate": 9.848664442125068e-06,
"loss": 44.6614,
"step": 229
},
{
"epoch": 0.39561384648462694,
"grad_norm": 9.553954124450684,
"learning_rate": 9.846321200262079e-06,
"loss": 44.8496,
"step": 230
},
{
"epoch": 0.397333906686734,
"grad_norm": 9.689641952514648,
"learning_rate": 9.843960239502205e-06,
"loss": 44.1763,
"step": 231
},
{
"epoch": 0.3990539668888411,
"grad_norm": 8.881220817565918,
"learning_rate": 9.841581568477483e-06,
"loss": 44.5679,
"step": 232
},
{
"epoch": 0.40077402709094817,
"grad_norm": 8.874262809753418,
"learning_rate": 9.839185195884702e-06,
"loss": 45.0497,
"step": 233
},
{
"epoch": 0.40249408729305525,
"grad_norm": 8.762188911437988,
"learning_rate": 9.836771130485367e-06,
"loss": 44.415,
"step": 234
},
{
"epoch": 0.4042141474951623,
"grad_norm": 7.233129501342773,
"learning_rate": 9.834339381105676e-06,
"loss": 44.4637,
"step": 235
},
{
"epoch": 0.4059342076972694,
"grad_norm": 10.540146827697754,
"learning_rate": 9.831889956636478e-06,
"loss": 44.525,
"step": 236
},
{
"epoch": 0.4076542678993765,
"grad_norm": 5.345295429229736,
"learning_rate": 9.829422866033246e-06,
"loss": 43.6553,
"step": 237
},
{
"epoch": 0.40937432810148355,
"grad_norm": 9.131731033325195,
"learning_rate": 9.826938118316044e-06,
"loss": 44.2395,
"step": 238
},
{
"epoch": 0.41109438830359063,
"grad_norm": 6.4219560623168945,
"learning_rate": 9.82443572256949e-06,
"loss": 44.541,
"step": 239
},
{
"epoch": 0.4128144485056977,
"grad_norm": 7.843993663787842,
"learning_rate": 9.821915687942729e-06,
"loss": 44.6975,
"step": 240
},
{
"epoch": 0.4145345087078048,
"grad_norm": 7.926616668701172,
"learning_rate": 9.8193780236494e-06,
"loss": 43.983,
"step": 241
},
{
"epoch": 0.41625456890991186,
"grad_norm": 7.5367045402526855,
"learning_rate": 9.81682273896759e-06,
"loss": 43.9646,
"step": 242
},
{
"epoch": 0.41797462911201894,
"grad_norm": 10.298775672912598,
"learning_rate": 9.814249843239816e-06,
"loss": 44.0679,
"step": 243
},
{
"epoch": 0.419694689314126,
"grad_norm": 8.142918586730957,
"learning_rate": 9.811659345872979e-06,
"loss": 44.9597,
"step": 244
},
{
"epoch": 0.4214147495162331,
"grad_norm": 11.297587394714355,
"learning_rate": 9.809051256338338e-06,
"loss": 44.3569,
"step": 245
},
{
"epoch": 0.42313480971834017,
"grad_norm": 8.743409156799316,
"learning_rate": 9.806425584171468e-06,
"loss": 43.6217,
"step": 246
},
{
"epoch": 0.42485486992044724,
"grad_norm": 9.554738998413086,
"learning_rate": 9.803782338972235e-06,
"loss": 44.3762,
"step": 247
},
{
"epoch": 0.42657493012255426,
"grad_norm": 8.766114234924316,
"learning_rate": 9.801121530404746e-06,
"loss": 44.1824,
"step": 248
},
{
"epoch": 0.42829499032466134,
"grad_norm": 8.465466499328613,
"learning_rate": 9.798443168197332e-06,
"loss": 44.0283,
"step": 249
},
{
"epoch": 0.4300150505267684,
"grad_norm": 8.999267578125,
"learning_rate": 9.795747262142494e-06,
"loss": 44.1171,
"step": 250
},
{
"epoch": 0.4317351107288755,
"grad_norm": 6.678277492523193,
"learning_rate": 9.79303382209688e-06,
"loss": 44.7172,
"step": 251
},
{
"epoch": 0.43345517093098257,
"grad_norm": 9.430837631225586,
"learning_rate": 9.790302857981247e-06,
"loss": 44.3632,
"step": 252
},
{
"epoch": 0.43517523113308965,
"grad_norm": 6.532567501068115,
"learning_rate": 9.787554379780417e-06,
"loss": 44.2348,
"step": 253
},
{
"epoch": 0.4368952913351967,
"grad_norm": 9.008966445922852,
"learning_rate": 9.784788397543254e-06,
"loss": 43.9189,
"step": 254
},
{
"epoch": 0.4386153515373038,
"grad_norm": 7.171030521392822,
"learning_rate": 9.782004921382612e-06,
"loss": 44.719,
"step": 255
},
{
"epoch": 0.4403354117394109,
"grad_norm": 8.457947731018066,
"learning_rate": 9.77920396147531e-06,
"loss": 44.3203,
"step": 256
},
{
"epoch": 0.44205547194151795,
"grad_norm": 8.303704261779785,
"learning_rate": 9.77638552806209e-06,
"loss": 44.6251,
"step": 257
},
{
"epoch": 0.44377553214362503,
"grad_norm": 7.793144702911377,
"learning_rate": 9.773549631447576e-06,
"loss": 44.4527,
"step": 258
},
{
"epoch": 0.4454955923457321,
"grad_norm": 9.074666976928711,
"learning_rate": 9.770696282000245e-06,
"loss": 44.4602,
"step": 259
},
{
"epoch": 0.4472156525478392,
"grad_norm": 7.790366172790527,
"learning_rate": 9.767825490152381e-06,
"loss": 44.0525,
"step": 260
},
{
"epoch": 0.44893571274994626,
"grad_norm": 6.790820598602295,
"learning_rate": 9.764937266400042e-06,
"loss": 44.3677,
"step": 261
},
{
"epoch": 0.45065577295205334,
"grad_norm": 7.48856782913208,
"learning_rate": 9.76203162130302e-06,
"loss": 44.281,
"step": 262
},
{
"epoch": 0.4523758331541604,
"grad_norm": 6.4276814460754395,
"learning_rate": 9.759108565484796e-06,
"loss": 44.5151,
"step": 263
},
{
"epoch": 0.4540958933562675,
"grad_norm": 7.4707350730896,
"learning_rate": 9.756168109632519e-06,
"loss": 44.4653,
"step": 264
},
{
"epoch": 0.45581595355837456,
"grad_norm": 6.78653621673584,
"learning_rate": 9.753210264496943e-06,
"loss": 45.1088,
"step": 265
},
{
"epoch": 0.45753601376048164,
"grad_norm": 6.732356548309326,
"learning_rate": 9.75023504089241e-06,
"loss": 44.3737,
"step": 266
},
{
"epoch": 0.4592560739625887,
"grad_norm": 6.069101810455322,
"learning_rate": 9.747242449696794e-06,
"loss": 44.5619,
"step": 267
},
{
"epoch": 0.46097613416469574,
"grad_norm": 6.123370170593262,
"learning_rate": 9.74423250185147e-06,
"loss": 44.3021,
"step": 268
},
{
"epoch": 0.4626961943668028,
"grad_norm": 6.104144096374512,
"learning_rate": 9.74120520836127e-06,
"loss": 44.2148,
"step": 269
},
{
"epoch": 0.4644162545689099,
"grad_norm": 8.752744674682617,
"learning_rate": 9.738160580294444e-06,
"loss": 44.6205,
"step": 270
},
{
"epoch": 0.46613631477101697,
"grad_norm": 7.354405879974365,
"learning_rate": 9.735098628782624e-06,
"loss": 44.459,
"step": 271
},
{
"epoch": 0.46785637497312405,
"grad_norm": 8.801732063293457,
"learning_rate": 9.732019365020778e-06,
"loss": 44.635,
"step": 272
},
{
"epoch": 0.4695764351752311,
"grad_norm": 7.110331058502197,
"learning_rate": 9.728922800267162e-06,
"loss": 44.3402,
"step": 273
},
{
"epoch": 0.4712964953773382,
"grad_norm": 8.147953033447266,
"learning_rate": 9.7258089458433e-06,
"loss": 44.6291,
"step": 274
},
{
"epoch": 0.4730165555794453,
"grad_norm": 7.08845853805542,
"learning_rate": 9.722677813133921e-06,
"loss": 45.0577,
"step": 275
},
{
"epoch": 0.47473661578155235,
"grad_norm": 6.4095025062561035,
"learning_rate": 9.719529413586928e-06,
"loss": 43.9258,
"step": 276
},
{
"epoch": 0.47645667598365943,
"grad_norm": 7.270499229431152,
"learning_rate": 9.716363758713357e-06,
"loss": 44.8198,
"step": 277
},
{
"epoch": 0.4781767361857665,
"grad_norm": 7.040411949157715,
"learning_rate": 9.713180860087328e-06,
"loss": 44.1966,
"step": 278
},
{
"epoch": 0.4798967963878736,
"grad_norm": 6.831189155578613,
"learning_rate": 9.709980729346009e-06,
"loss": 44.88,
"step": 279
},
{
"epoch": 0.48161685658998066,
"grad_norm": 7.017982482910156,
"learning_rate": 9.706763378189571e-06,
"loss": 44.3914,
"step": 280
},
{
"epoch": 0.48333691679208773,
"grad_norm": 7.631150245666504,
"learning_rate": 9.703528818381144e-06,
"loss": 44.3798,
"step": 281
},
{
"epoch": 0.4850569769941948,
"grad_norm": 6.717972278594971,
"learning_rate": 9.70027706174678e-06,
"loss": 43.8038,
"step": 282
},
{
"epoch": 0.4867770371963019,
"grad_norm": 8.252500534057617,
"learning_rate": 9.697008120175402e-06,
"loss": 44.4889,
"step": 283
},
{
"epoch": 0.48849709739840896,
"grad_norm": 7.6612420082092285,
"learning_rate": 9.693722005618763e-06,
"loss": 44.2001,
"step": 284
},
{
"epoch": 0.49021715760051604,
"grad_norm": 6.846263408660889,
"learning_rate": 9.690418730091403e-06,
"loss": 44.6985,
"step": 285
},
{
"epoch": 0.4919372178026231,
"grad_norm": 6.461937427520752,
"learning_rate": 9.687098305670606e-06,
"loss": 44.2506,
"step": 286
},
{
"epoch": 0.49365727800473014,
"grad_norm": 6.650141716003418,
"learning_rate": 9.683760744496356e-06,
"loss": 44.3858,
"step": 287
},
{
"epoch": 0.4953773382068372,
"grad_norm": 6.5828986167907715,
"learning_rate": 9.68040605877129e-06,
"loss": 43.6322,
"step": 288
},
{
"epoch": 0.4970973984089443,
"grad_norm": 6.385183334350586,
"learning_rate": 9.677034260760658e-06,
"loss": 44.7745,
"step": 289
},
{
"epoch": 0.49881745861105137,
"grad_norm": 7.130415916442871,
"learning_rate": 9.673645362792273e-06,
"loss": 44.1543,
"step": 290
},
{
"epoch": 0.5005375188131584,
"grad_norm": 6.580416202545166,
"learning_rate": 9.670239377256467e-06,
"loss": 43.8422,
"step": 291
},
{
"epoch": 0.5022575790152656,
"grad_norm": 7.959731101989746,
"learning_rate": 9.666816316606044e-06,
"loss": 44.4367,
"step": 292
},
{
"epoch": 0.5039776392173726,
"grad_norm": 6.089702606201172,
"learning_rate": 9.663376193356249e-06,
"loss": 43.9484,
"step": 293
},
{
"epoch": 0.5056976994194797,
"grad_norm": 8.458806037902832,
"learning_rate": 9.659919020084695e-06,
"loss": 44.1408,
"step": 294
},
{
"epoch": 0.5074177596215868,
"grad_norm": 6.733780860900879,
"learning_rate": 9.656444809431344e-06,
"loss": 43.9267,
"step": 295
},
{
"epoch": 0.5091378198236939,
"grad_norm": 8.281241416931152,
"learning_rate": 9.652953574098444e-06,
"loss": 44.7447,
"step": 296
},
{
"epoch": 0.5108578800258009,
"grad_norm": 7.2893195152282715,
"learning_rate": 9.649445326850491e-06,
"loss": 44.1749,
"step": 297
},
{
"epoch": 0.5125779402279079,
"grad_norm": 8.188138961791992,
"learning_rate": 9.645920080514176e-06,
"loss": 44.5725,
"step": 298
},
{
"epoch": 0.5142980004300151,
"grad_norm": 8.285508155822754,
"learning_rate": 9.642377847978343e-06,
"loss": 44.4519,
"step": 299
},
{
"epoch": 0.5160180606321221,
"grad_norm": 12.107803344726562,
"learning_rate": 9.638818642193939e-06,
"loss": 43.6642,
"step": 300
},
{
"epoch": 0.5177381208342292,
"grad_norm": 10.988150596618652,
"learning_rate": 9.63524247617397e-06,
"loss": 43.9385,
"step": 301
},
{
"epoch": 0.5194581810363362,
"grad_norm": 12.65985107421875,
"learning_rate": 9.631649362993447e-06,
"loss": 44.304,
"step": 302
},
{
"epoch": 0.5211782412384434,
"grad_norm": 12.63979721069336,
"learning_rate": 9.62803931578935e-06,
"loss": 44.2028,
"step": 303
},
{
"epoch": 0.5228983014405504,
"grad_norm": 7.90657377243042,
"learning_rate": 9.624412347760564e-06,
"loss": 44.1649,
"step": 304
},
{
"epoch": 0.5246183616426575,
"grad_norm": 9.31624698638916,
"learning_rate": 9.620768472167844e-06,
"loss": 43.996,
"step": 305
},
{
"epoch": 0.5263384218447645,
"grad_norm": 8.557055473327637,
"learning_rate": 9.61710770233376e-06,
"loss": 44.3358,
"step": 306
},
{
"epoch": 0.5280584820468717,
"grad_norm": 7.057743549346924,
"learning_rate": 9.613430051642652e-06,
"loss": 44.583,
"step": 307
},
{
"epoch": 0.5297785422489787,
"grad_norm": 7.244456768035889,
"learning_rate": 9.609735533540576e-06,
"loss": 43.7423,
"step": 308
},
{
"epoch": 0.5314986024510858,
"grad_norm": 6.6239333152771,
"learning_rate": 9.606024161535261e-06,
"loss": 43.719,
"step": 309
},
{
"epoch": 0.5332186626531928,
"grad_norm": 7.109512805938721,
"learning_rate": 9.602295949196052e-06,
"loss": 43.8263,
"step": 310
},
{
"epoch": 0.5349387228553,
"grad_norm": 7.938423156738281,
"learning_rate": 9.59855091015387e-06,
"loss": 43.513,
"step": 311
},
{
"epoch": 0.536658783057407,
"grad_norm": 7.159519195556641,
"learning_rate": 9.594789058101154e-06,
"loss": 43.7476,
"step": 312
},
{
"epoch": 0.5383788432595141,
"grad_norm": 9.051861763000488,
"learning_rate": 9.591010406791814e-06,
"loss": 44.6808,
"step": 313
},
{
"epoch": 0.5400989034616211,
"grad_norm": 7.437441825866699,
"learning_rate": 9.587214970041181e-06,
"loss": 44.2018,
"step": 314
},
{
"epoch": 0.5418189636637283,
"grad_norm": 8.533609390258789,
"learning_rate": 9.58340276172596e-06,
"loss": 44.35,
"step": 315
},
{
"epoch": 0.5435390238658353,
"grad_norm": 7.41975736618042,
"learning_rate": 9.579573795784167e-06,
"loss": 44.1627,
"step": 316
},
{
"epoch": 0.5452590840679423,
"grad_norm": 6.5013580322265625,
"learning_rate": 9.575728086215093e-06,
"loss": 44.0411,
"step": 317
},
{
"epoch": 0.5469791442700495,
"grad_norm": 8.311059951782227,
"learning_rate": 9.571865647079246e-06,
"loss": 44.6953,
"step": 318
},
{
"epoch": 0.5486992044721565,
"grad_norm": 5.960739612579346,
"learning_rate": 9.567986492498299e-06,
"loss": 44.1261,
"step": 319
},
{
"epoch": 0.5504192646742636,
"grad_norm": 7.7150959968566895,
"learning_rate": 9.564090636655033e-06,
"loss": 44.0052,
"step": 320
},
{
"epoch": 0.5521393248763706,
"grad_norm": 7.0516815185546875,
"learning_rate": 9.560178093793304e-06,
"loss": 44.4024,
"step": 321
},
{
"epoch": 0.5538593850784778,
"grad_norm": 6.510403633117676,
"learning_rate": 9.55624887821797e-06,
"loss": 44.3171,
"step": 322
},
{
"epoch": 0.5555794452805848,
"grad_norm": 6.586174488067627,
"learning_rate": 9.552303004294845e-06,
"loss": 44.1694,
"step": 323
},
{
"epoch": 0.5572995054826919,
"grad_norm": 7.093349456787109,
"learning_rate": 9.548340486450656e-06,
"loss": 43.9714,
"step": 324
},
{
"epoch": 0.5590195656847989,
"grad_norm": 5.638337135314941,
"learning_rate": 9.544361339172976e-06,
"loss": 43.9597,
"step": 325
},
{
"epoch": 0.5607396258869061,
"grad_norm": 6.339056491851807,
"learning_rate": 9.54036557701018e-06,
"loss": 43.9718,
"step": 326
},
{
"epoch": 0.5624596860890131,
"grad_norm": 6.259551525115967,
"learning_rate": 9.536353214571393e-06,
"loss": 44.1765,
"step": 327
},
{
"epoch": 0.5641797462911202,
"grad_norm": 6.071033477783203,
"learning_rate": 9.53232426652643e-06,
"loss": 44.1298,
"step": 328
},
{
"epoch": 0.5658998064932272,
"grad_norm": 5.966522216796875,
"learning_rate": 9.528278747605741e-06,
"loss": 43.5899,
"step": 329
},
{
"epoch": 0.5676198666953344,
"grad_norm": 6.373861312866211,
"learning_rate": 9.52421667260037e-06,
"loss": 43.9747,
"step": 330
},
{
"epoch": 0.5693399268974414,
"grad_norm": 7.79583215713501,
"learning_rate": 9.52013805636189e-06,
"loss": 44.1957,
"step": 331
},
{
"epoch": 0.5710599870995485,
"grad_norm": 6.730131149291992,
"learning_rate": 9.516042913802349e-06,
"loss": 44.5105,
"step": 332
},
{
"epoch": 0.5727800473016555,
"grad_norm": 6.649820804595947,
"learning_rate": 9.511931259894219e-06,
"loss": 43.4763,
"step": 333
},
{
"epoch": 0.5745001075037627,
"grad_norm": 8.12640380859375,
"learning_rate": 9.507803109670337e-06,
"loss": 43.5421,
"step": 334
},
{
"epoch": 0.5762201677058697,
"grad_norm": 7.1897783279418945,
"learning_rate": 9.503658478223862e-06,
"loss": 43.661,
"step": 335
},
{
"epoch": 0.5779402279079768,
"grad_norm": 6.837246417999268,
"learning_rate": 9.499497380708202e-06,
"loss": 43.7997,
"step": 336
},
{
"epoch": 0.5796602881100839,
"grad_norm": 8.631741523742676,
"learning_rate": 9.495319832336969e-06,
"loss": 43.8287,
"step": 337
},
{
"epoch": 0.5813803483121909,
"grad_norm": 8.00960636138916,
"learning_rate": 9.491125848383926e-06,
"loss": 43.6861,
"step": 338
},
{
"epoch": 0.583100408514298,
"grad_norm": 6.487185001373291,
"learning_rate": 9.486915444182926e-06,
"loss": 43.8275,
"step": 339
},
{
"epoch": 0.584820468716405,
"grad_norm": 7.411306381225586,
"learning_rate": 9.482688635127849e-06,
"loss": 43.4639,
"step": 340
},
{
"epoch": 0.5865405289185122,
"grad_norm": 6.86546516418457,
"learning_rate": 9.478445436672566e-06,
"loss": 44.094,
"step": 341
},
{
"epoch": 0.5882605891206192,
"grad_norm": 8.094916343688965,
"learning_rate": 9.474185864330861e-06,
"loss": 43.8569,
"step": 342
},
{
"epoch": 0.5899806493227263,
"grad_norm": 7.7384138107299805,
"learning_rate": 9.469909933676388e-06,
"loss": 43.8937,
"step": 343
},
{
"epoch": 0.5917007095248333,
"grad_norm": 8.301685333251953,
"learning_rate": 9.46561766034261e-06,
"loss": 44.288,
"step": 344
},
{
"epoch": 0.5934207697269405,
"grad_norm": 10.100611686706543,
"learning_rate": 9.461309060022737e-06,
"loss": 43.9889,
"step": 345
},
{
"epoch": 0.5951408299290475,
"grad_norm": 6.778660774230957,
"learning_rate": 9.456984148469674e-06,
"loss": 43.697,
"step": 346
},
{
"epoch": 0.5968608901311546,
"grad_norm": 11.071110725402832,
"learning_rate": 9.452642941495967e-06,
"loss": 44.3392,
"step": 347
},
{
"epoch": 0.5985809503332616,
"grad_norm": 7.321798324584961,
"learning_rate": 9.448285454973739e-06,
"loss": 43.5908,
"step": 348
},
{
"epoch": 0.6003010105353688,
"grad_norm": 10.584439277648926,
"learning_rate": 9.443911704834624e-06,
"loss": 44.094,
"step": 349
},
{
"epoch": 0.6020210707374758,
"grad_norm": 7.291213035583496,
"learning_rate": 9.439521707069737e-06,
"loss": 43.7771,
"step": 350
},
{
"epoch": 0.6037411309395829,
"grad_norm": 10.645991325378418,
"learning_rate": 9.435115477729577e-06,
"loss": 44.2799,
"step": 351
},
{
"epoch": 0.60546119114169,
"grad_norm": 8.463363647460938,
"learning_rate": 9.430693032924003e-06,
"loss": 43.9206,
"step": 352
},
{
"epoch": 0.6071812513437971,
"grad_norm": 11.510597229003906,
"learning_rate": 9.426254388822152e-06,
"loss": 43.8839,
"step": 353
},
{
"epoch": 0.6089013115459041,
"grad_norm": 9.736458778381348,
"learning_rate": 9.421799561652391e-06,
"loss": 44.6627,
"step": 354
},
{
"epoch": 0.6106213717480112,
"grad_norm": 11.647321701049805,
"learning_rate": 9.417328567702256e-06,
"loss": 43.6068,
"step": 355
},
{
"epoch": 0.6123414319501183,
"grad_norm": 12.150289535522461,
"learning_rate": 9.412841423318386e-06,
"loss": 43.6395,
"step": 356
},
{
"epoch": 0.6140614921522253,
"grad_norm": 9.2246675491333,
"learning_rate": 9.408338144906475e-06,
"loss": 44.1518,
"step": 357
},
{
"epoch": 0.6157815523543324,
"grad_norm": 10.310601234436035,
"learning_rate": 9.403818748931201e-06,
"loss": 43.9735,
"step": 358
},
{
"epoch": 0.6175016125564394,
"grad_norm": 8.350564002990723,
"learning_rate": 9.399283251916174e-06,
"loss": 43.7255,
"step": 359
},
{
"epoch": 0.6192216727585466,
"grad_norm": 9.116833686828613,
"learning_rate": 9.394731670443869e-06,
"loss": 43.6732,
"step": 360
},
{
"epoch": 0.6209417329606536,
"grad_norm": 8.664816856384277,
"learning_rate": 9.390164021155568e-06,
"loss": 43.6788,
"step": 361
},
{
"epoch": 0.6226617931627607,
"grad_norm": 8.421276092529297,
"learning_rate": 9.385580320751301e-06,
"loss": 43.7357,
"step": 362
},
{
"epoch": 0.6243818533648677,
"grad_norm": 7.4578680992126465,
"learning_rate": 9.380980585989782e-06,
"loss": 43.7452,
"step": 363
},
{
"epoch": 0.6261019135669749,
"grad_norm": 8.058576583862305,
"learning_rate": 9.376364833688352e-06,
"loss": 43.839,
"step": 364
},
{
"epoch": 0.6278219737690819,
"grad_norm": 7.826845169067383,
"learning_rate": 9.371733080722911e-06,
"loss": 43.8903,
"step": 365
},
{
"epoch": 0.629542033971189,
"grad_norm": 7.76792573928833,
"learning_rate": 9.367085344027862e-06,
"loss": 44.3217,
"step": 366
},
{
"epoch": 0.631262094173296,
"grad_norm": 8.252405166625977,
"learning_rate": 9.362421640596044e-06,
"loss": 44.0905,
"step": 367
},
{
"epoch": 0.6329821543754032,
"grad_norm": 7.431006908416748,
"learning_rate": 9.35774198747868e-06,
"loss": 43.864,
"step": 368
},
{
"epoch": 0.6347022145775102,
"grad_norm": 9.33600902557373,
"learning_rate": 9.353046401785297e-06,
"loss": 43.0342,
"step": 369
},
{
"epoch": 0.6364222747796173,
"grad_norm": 7.454495906829834,
"learning_rate": 9.348334900683685e-06,
"loss": 43.7442,
"step": 370
},
{
"epoch": 0.6381423349817243,
"grad_norm": 8.429414749145508,
"learning_rate": 9.343607501399812e-06,
"loss": 43.7836,
"step": 371
},
{
"epoch": 0.6398623951838315,
"grad_norm": 7.481090545654297,
"learning_rate": 9.338864221217783e-06,
"loss": 43.9994,
"step": 372
},
{
"epoch": 0.6415824553859385,
"grad_norm": 7.106781482696533,
"learning_rate": 9.33410507747976e-06,
"loss": 43.6922,
"step": 373
},
{
"epoch": 0.6433025155880456,
"grad_norm": 7.0425615310668945,
"learning_rate": 9.329330087585905e-06,
"loss": 44.5278,
"step": 374
},
{
"epoch": 0.6450225757901527,
"grad_norm": 7.197376728057861,
"learning_rate": 9.324539268994317e-06,
"loss": 43.8955,
"step": 375
},
{
"epoch": 0.6467426359922598,
"grad_norm": 6.9038286209106445,
"learning_rate": 9.319732639220965e-06,
"loss": 43.9268,
"step": 376
},
{
"epoch": 0.6484626961943668,
"grad_norm": 7.032724380493164,
"learning_rate": 9.31491021583963e-06,
"loss": 43.5221,
"step": 377
},
{
"epoch": 0.6501827563964738,
"grad_norm": 7.234856605529785,
"learning_rate": 9.310072016481832e-06,
"loss": 43.4951,
"step": 378
},
{
"epoch": 0.651902816598581,
"grad_norm": 6.546868801116943,
"learning_rate": 9.305218058836778e-06,
"loss": 44.0876,
"step": 379
},
{
"epoch": 0.653622876800688,
"grad_norm": 6.904932498931885,
"learning_rate": 9.300348360651282e-06,
"loss": 44.5423,
"step": 380
},
{
"epoch": 0.6553429370027951,
"grad_norm": 7.505612850189209,
"learning_rate": 9.295462939729711e-06,
"loss": 43.9986,
"step": 381
},
{
"epoch": 0.6570629972049021,
"grad_norm": 5.859259605407715,
"learning_rate": 9.290561813933916e-06,
"loss": 43.8683,
"step": 382
},
{
"epoch": 0.6587830574070093,
"grad_norm": 8.047765731811523,
"learning_rate": 9.285645001183167e-06,
"loss": 44.4658,
"step": 383
},
{
"epoch": 0.6605031176091163,
"grad_norm": 6.570570945739746,
"learning_rate": 9.280712519454092e-06,
"loss": 43.6115,
"step": 384
},
{
"epoch": 0.6622231778112234,
"grad_norm": 6.266587734222412,
"learning_rate": 9.2757643867806e-06,
"loss": 43.6822,
"step": 385
},
{
"epoch": 0.6639432380133304,
"grad_norm": 7.281513214111328,
"learning_rate": 9.270800621253833e-06,
"loss": 43.8285,
"step": 386
},
{
"epoch": 0.6656632982154376,
"grad_norm": 6.563234806060791,
"learning_rate": 9.265821241022074e-06,
"loss": 43.6976,
"step": 387
},
{
"epoch": 0.6673833584175446,
"grad_norm": 6.870432376861572,
"learning_rate": 9.26082626429071e-06,
"loss": 43.9566,
"step": 388
},
{
"epoch": 0.6691034186196517,
"grad_norm": 8.11976432800293,
"learning_rate": 9.255815709322142e-06,
"loss": 43.8613,
"step": 389
},
{
"epoch": 0.6708234788217587,
"grad_norm": 7.040714740753174,
"learning_rate": 9.250789594435735e-06,
"loss": 43.3387,
"step": 390
},
{
"epoch": 0.6725435390238659,
"grad_norm": 6.891185283660889,
"learning_rate": 9.245747938007734e-06,
"loss": 43.596,
"step": 391
},
{
"epoch": 0.6742635992259729,
"grad_norm": 7.045391082763672,
"learning_rate": 9.240690758471216e-06,
"loss": 43.1001,
"step": 392
},
{
"epoch": 0.67598365942808,
"grad_norm": 6.838486194610596,
"learning_rate": 9.235618074316005e-06,
"loss": 44.0918,
"step": 393
},
{
"epoch": 0.677703719630187,
"grad_norm": 8.006799697875977,
"learning_rate": 9.230529904088621e-06,
"loss": 43.4563,
"step": 394
},
{
"epoch": 0.6794237798322942,
"grad_norm": 7.786087989807129,
"learning_rate": 9.225426266392191e-06,
"loss": 44.1002,
"step": 395
},
{
"epoch": 0.6811438400344012,
"grad_norm": 7.782168388366699,
"learning_rate": 9.220307179886408e-06,
"loss": 44.15,
"step": 396
},
{
"epoch": 0.6828639002365082,
"grad_norm": 7.179986953735352,
"learning_rate": 9.215172663287435e-06,
"loss": 43.8326,
"step": 397
},
{
"epoch": 0.6845839604386154,
"grad_norm": 8.045145988464355,
"learning_rate": 9.210022735367857e-06,
"loss": 43.3196,
"step": 398
},
{
"epoch": 0.6863040206407224,
"grad_norm": 7.906603813171387,
"learning_rate": 9.204857414956606e-06,
"loss": 44.152,
"step": 399
},
{
"epoch": 0.6880240808428295,
"grad_norm": 8.266923904418945,
"learning_rate": 9.199676720938886e-06,
"loss": 44.158,
"step": 400
},
{
"epoch": 0.6897441410449365,
"grad_norm": 7.465760231018066,
"learning_rate": 9.194480672256117e-06,
"loss": 43.9078,
"step": 401
},
{
"epoch": 0.6914642012470437,
"grad_norm": 7.522243499755859,
"learning_rate": 9.189269287905849e-06,
"loss": 43.3097,
"step": 402
},
{
"epoch": 0.6931842614491507,
"grad_norm": 6.484007358551025,
"learning_rate": 9.184042586941708e-06,
"loss": 43.9014,
"step": 403
},
{
"epoch": 0.6949043216512578,
"grad_norm": 6.548778533935547,
"learning_rate": 9.178800588473317e-06,
"loss": 43.9104,
"step": 404
},
{
"epoch": 0.6966243818533648,
"grad_norm": 6.578863620758057,
"learning_rate": 9.17354331166623e-06,
"loss": 44.0022,
"step": 405
},
{
"epoch": 0.698344442055472,
"grad_norm": 7.083658695220947,
"learning_rate": 9.168270775741863e-06,
"loss": 43.7902,
"step": 406
},
{
"epoch": 0.700064502257579,
"grad_norm": 7.241711139678955,
"learning_rate": 9.162982999977417e-06,
"loss": 43.9464,
"step": 407
},
{
"epoch": 0.7017845624596861,
"grad_norm": 6.329436779022217,
"learning_rate": 9.157680003705816e-06,
"loss": 44.1072,
"step": 408
},
{
"epoch": 0.7035046226617931,
"grad_norm": 6.435650825500488,
"learning_rate": 9.15236180631563e-06,
"loss": 43.392,
"step": 409
},
{
"epoch": 0.7052246828639003,
"grad_norm": 5.681223392486572,
"learning_rate": 9.14702842725101e-06,
"loss": 44.3148,
"step": 410
},
{
"epoch": 0.7069447430660073,
"grad_norm": 6.71289587020874,
"learning_rate": 9.14167988601161e-06,
"loss": 43.8893,
"step": 411
},
{
"epoch": 0.7086648032681144,
"grad_norm": 6.497440814971924,
"learning_rate": 9.13631620215252e-06,
"loss": 44.1776,
"step": 412
},
{
"epoch": 0.7103848634702215,
"grad_norm": 7.291422367095947,
"learning_rate": 9.130937395284199e-06,
"loss": 43.8195,
"step": 413
},
{
"epoch": 0.7121049236723286,
"grad_norm": 6.935153961181641,
"learning_rate": 9.125543485072386e-06,
"loss": 43.9977,
"step": 414
},
{
"epoch": 0.7138249838744356,
"grad_norm": 6.302245140075684,
"learning_rate": 9.120134491238054e-06,
"loss": 43.677,
"step": 415
},
{
"epoch": 0.7155450440765426,
"grad_norm": 6.205868244171143,
"learning_rate": 9.114710433557314e-06,
"loss": 43.8423,
"step": 416
},
{
"epoch": 0.7172651042786498,
"grad_norm": 5.34831428527832,
"learning_rate": 9.109271331861361e-06,
"loss": 43.6707,
"step": 417
},
{
"epoch": 0.7189851644807568,
"grad_norm": 7.174152374267578,
"learning_rate": 9.103817206036383e-06,
"loss": 43.3579,
"step": 418
},
{
"epoch": 0.7207052246828639,
"grad_norm": 6.666977882385254,
"learning_rate": 9.098348076023506e-06,
"loss": 43.8424,
"step": 419
},
{
"epoch": 0.7224252848849709,
"grad_norm": 7.491025924682617,
"learning_rate": 9.092863961818715e-06,
"loss": 44.4333,
"step": 420
},
{
"epoch": 0.7241453450870781,
"grad_norm": 6.508261680603027,
"learning_rate": 9.087364883472774e-06,
"loss": 43.7001,
"step": 421
},
{
"epoch": 0.7258654052891851,
"grad_norm": 7.418080806732178,
"learning_rate": 9.08185086109116e-06,
"loss": 44.0001,
"step": 422
},
{
"epoch": 0.7275854654912922,
"grad_norm": 6.983603000640869,
"learning_rate": 9.076321914833988e-06,
"loss": 44.6241,
"step": 423
},
{
"epoch": 0.7293055256933992,
"grad_norm": 8.667305946350098,
"learning_rate": 9.070778064915937e-06,
"loss": 44.1089,
"step": 424
},
{
"epoch": 0.7310255858955064,
"grad_norm": 7.419984340667725,
"learning_rate": 9.065219331606182e-06,
"loss": 43.9046,
"step": 425
},
{
"epoch": 0.7327456460976134,
"grad_norm": 7.34318733215332,
"learning_rate": 9.0596457352283e-06,
"loss": 43.794,
"step": 426
},
{
"epoch": 0.7344657062997205,
"grad_norm": 7.931493759155273,
"learning_rate": 9.054057296160221e-06,
"loss": 44.6317,
"step": 427
},
{
"epoch": 0.7361857665018275,
"grad_norm": 6.583981037139893,
"learning_rate": 9.048454034834143e-06,
"loss": 43.5199,
"step": 428
},
{
"epoch": 0.7379058267039347,
"grad_norm": 8.499653816223145,
"learning_rate": 9.042835971736446e-06,
"loss": 43.8616,
"step": 429
},
{
"epoch": 0.7396258869060417,
"grad_norm": 6.757936000823975,
"learning_rate": 9.037203127407642e-06,
"loss": 44.0385,
"step": 430
},
{
"epoch": 0.7413459471081488,
"grad_norm": 8.30978012084961,
"learning_rate": 9.031555522442268e-06,
"loss": 43.4628,
"step": 431
},
{
"epoch": 0.7430660073102558,
"grad_norm": 7.357321262359619,
"learning_rate": 9.025893177488848e-06,
"loss": 43.6677,
"step": 432
},
{
"epoch": 0.744786067512363,
"grad_norm": 8.4613676071167,
"learning_rate": 9.02021611324978e-06,
"loss": 43.2542,
"step": 433
},
{
"epoch": 0.74650612771447,
"grad_norm": 8.782477378845215,
"learning_rate": 9.014524350481287e-06,
"loss": 44.0515,
"step": 434
},
{
"epoch": 0.7482261879165771,
"grad_norm": 6.701351165771484,
"learning_rate": 9.008817909993332e-06,
"loss": 44.0643,
"step": 435
},
{
"epoch": 0.7499462481186842,
"grad_norm": 7.1624884605407715,
"learning_rate": 9.00309681264954e-06,
"loss": 44.0639,
"step": 436
},
{
"epoch": 0.7516663083207912,
"grad_norm": 6.229190349578857,
"learning_rate": 8.997361079367124e-06,
"loss": 43.9093,
"step": 437
},
{
"epoch": 0.7533863685228983,
"grad_norm": 7.180543422698975,
"learning_rate": 8.991610731116808e-06,
"loss": 44.2652,
"step": 438
},
{
"epoch": 0.7551064287250053,
"grad_norm": 6.3346781730651855,
"learning_rate": 8.985845788922753e-06,
"loss": 43.2561,
"step": 439
},
{
"epoch": 0.7568264889271125,
"grad_norm": 7.270414352416992,
"learning_rate": 8.980066273862473e-06,
"loss": 43.9074,
"step": 440
},
{
"epoch": 0.7585465491292195,
"grad_norm": 6.431163311004639,
"learning_rate": 8.974272207066767e-06,
"loss": 43.9343,
"step": 441
},
{
"epoch": 0.7602666093313266,
"grad_norm": 6.415679931640625,
"learning_rate": 8.968463609719636e-06,
"loss": 44.2067,
"step": 442
},
{
"epoch": 0.7619866695334336,
"grad_norm": 6.083033561706543,
"learning_rate": 8.962640503058206e-06,
"loss": 43.7967,
"step": 443
},
{
"epoch": 0.7637067297355408,
"grad_norm": 6.942599773406982,
"learning_rate": 8.956802908372652e-06,
"loss": 43.4928,
"step": 444
},
{
"epoch": 0.7654267899376478,
"grad_norm": 6.515557765960693,
"learning_rate": 8.95095084700612e-06,
"loss": 43.6577,
"step": 445
},
{
"epoch": 0.7671468501397549,
"grad_norm": 7.167238235473633,
"learning_rate": 8.945084340354646e-06,
"loss": 43.4542,
"step": 446
},
{
"epoch": 0.7688669103418619,
"grad_norm": 7.30296516418457,
"learning_rate": 8.939203409867084e-06,
"loss": 43.123,
"step": 447
},
{
"epoch": 0.7705869705439691,
"grad_norm": 7.387278079986572,
"learning_rate": 8.933308077045022e-06,
"loss": 43.6603,
"step": 448
},
{
"epoch": 0.7723070307460761,
"grad_norm": 7.026780128479004,
"learning_rate": 8.927398363442705e-06,
"loss": 44.0497,
"step": 449
},
{
"epoch": 0.7740270909481832,
"grad_norm": 7.03558874130249,
"learning_rate": 8.921474290666955e-06,
"loss": 43.1461,
"step": 450
},
{
"epoch": 0.7757471511502902,
"grad_norm": 8.48353099822998,
"learning_rate": 8.915535880377096e-06,
"loss": 43.6771,
"step": 451
},
{
"epoch": 0.7774672113523974,
"grad_norm": 7.250082015991211,
"learning_rate": 8.909583154284868e-06,
"loss": 43.9369,
"step": 452
},
{
"epoch": 0.7791872715545044,
"grad_norm": 8.958197593688965,
"learning_rate": 8.90361613415436e-06,
"loss": 43.7911,
"step": 453
},
{
"epoch": 0.7809073317566115,
"grad_norm": 8.168319702148438,
"learning_rate": 8.897634841801911e-06,
"loss": 43.3905,
"step": 454
},
{
"epoch": 0.7826273919587186,
"grad_norm": 8.520408630371094,
"learning_rate": 8.891639299096051e-06,
"loss": 43.3708,
"step": 455
},
{
"epoch": 0.7843474521608256,
"grad_norm": 8.194758415222168,
"learning_rate": 8.885629527957407e-06,
"loss": 43.3692,
"step": 456
},
{
"epoch": 0.7860675123629327,
"grad_norm": 7.554206848144531,
"learning_rate": 8.879605550358627e-06,
"loss": 43.5693,
"step": 457
},
{
"epoch": 0.7877875725650397,
"grad_norm": 7.146202087402344,
"learning_rate": 8.873567388324302e-06,
"loss": 43.8261,
"step": 458
},
{
"epoch": 0.7895076327671469,
"grad_norm": 6.808493137359619,
"learning_rate": 8.867515063930881e-06,
"loss": 43.3648,
"step": 459
},
{
"epoch": 0.7912276929692539,
"grad_norm": 7.165658950805664,
"learning_rate": 8.861448599306597e-06,
"loss": 43.0367,
"step": 460
},
{
"epoch": 0.792947753171361,
"grad_norm": 6.527984142303467,
"learning_rate": 8.855368016631377e-06,
"loss": 43.4491,
"step": 461
},
{
"epoch": 0.794667813373468,
"grad_norm": 6.912752628326416,
"learning_rate": 8.849273338136772e-06,
"loss": 43.6405,
"step": 462
},
{
"epoch": 0.7963878735755752,
"grad_norm": 6.334918975830078,
"learning_rate": 8.84316458610586e-06,
"loss": 44.4637,
"step": 463
},
{
"epoch": 0.7981079337776822,
"grad_norm": 7.22133207321167,
"learning_rate": 8.837041782873182e-06,
"loss": 43.2829,
"step": 464
},
{
"epoch": 0.7998279939797893,
"grad_norm": 6.233572006225586,
"learning_rate": 8.83090495082465e-06,
"loss": 43.3993,
"step": 465
},
{
"epoch": 0.8015480541818963,
"grad_norm": 6.681156635284424,
"learning_rate": 8.824754112397467e-06,
"loss": 43.2356,
"step": 466
},
{
"epoch": 0.8032681143840035,
"grad_norm": 7.240959167480469,
"learning_rate": 8.818589290080043e-06,
"loss": 42.8966,
"step": 467
},
{
"epoch": 0.8049881745861105,
"grad_norm": 6.253081798553467,
"learning_rate": 8.812410506411925e-06,
"loss": 43.9822,
"step": 468
},
{
"epoch": 0.8067082347882176,
"grad_norm": 7.541505813598633,
"learning_rate": 8.806217783983693e-06,
"loss": 43.9604,
"step": 469
},
{
"epoch": 0.8084282949903246,
"grad_norm": 7.4928483963012695,
"learning_rate": 8.800011145436893e-06,
"loss": 43.8446,
"step": 470
},
{
"epoch": 0.8101483551924318,
"grad_norm": 6.140499591827393,
"learning_rate": 8.793790613463956e-06,
"loss": 43.6913,
"step": 471
},
{
"epoch": 0.8118684153945388,
"grad_norm": 7.944373607635498,
"learning_rate": 8.787556210808101e-06,
"loss": 43.4474,
"step": 472
},
{
"epoch": 0.8135884755966459,
"grad_norm": 6.9422101974487305,
"learning_rate": 8.781307960263267e-06,
"loss": 43.293,
"step": 473
},
{
"epoch": 0.815308535798753,
"grad_norm": 6.664095878601074,
"learning_rate": 8.77504588467402e-06,
"loss": 43.8227,
"step": 474
},
{
"epoch": 0.81702859600086,
"grad_norm": 7.298461437225342,
"learning_rate": 8.768770006935475e-06,
"loss": 43.7175,
"step": 475
},
{
"epoch": 0.8187486562029671,
"grad_norm": 6.43251895904541,
"learning_rate": 8.762480349993204e-06,
"loss": 43.143,
"step": 476
},
{
"epoch": 0.8204687164050741,
"grad_norm": 6.303859233856201,
"learning_rate": 8.756176936843161e-06,
"loss": 43.7655,
"step": 477
},
{
"epoch": 0.8221887766071813,
"grad_norm": 6.824503421783447,
"learning_rate": 8.749859790531601e-06,
"loss": 43.5909,
"step": 478
},
{
"epoch": 0.8239088368092883,
"grad_norm": 6.232965469360352,
"learning_rate": 8.743528934154982e-06,
"loss": 43.6798,
"step": 479
},
{
"epoch": 0.8256288970113954,
"grad_norm": 6.288873672485352,
"learning_rate": 8.737184390859887e-06,
"loss": 43.4713,
"step": 480
},
{
"epoch": 0.8273489572135024,
"grad_norm": 6.1072306632995605,
"learning_rate": 8.730826183842947e-06,
"loss": 43.4521,
"step": 481
},
{
"epoch": 0.8290690174156096,
"grad_norm": 7.3213701248168945,
"learning_rate": 8.724454336350742e-06,
"loss": 43.9662,
"step": 482
},
{
"epoch": 0.8307890776177166,
"grad_norm": 6.282354354858398,
"learning_rate": 8.718068871679735e-06,
"loss": 44.1781,
"step": 483
},
{
"epoch": 0.8325091378198237,
"grad_norm": 7.692941188812256,
"learning_rate": 8.711669813176165e-06,
"loss": 43.5585,
"step": 484
},
{
"epoch": 0.8342291980219307,
"grad_norm": 6.070176124572754,
"learning_rate": 8.705257184235973e-06,
"loss": 43.843,
"step": 485
},
{
"epoch": 0.8359492582240379,
"grad_norm": 7.584023952484131,
"learning_rate": 8.698831008304723e-06,
"loss": 43.5888,
"step": 486
},
{
"epoch": 0.8376693184261449,
"grad_norm": 8.037973403930664,
"learning_rate": 8.6923913088775e-06,
"loss": 43.4765,
"step": 487
},
{
"epoch": 0.839389378628252,
"grad_norm": 6.745630741119385,
"learning_rate": 8.685938109498839e-06,
"loss": 44.0438,
"step": 488
},
{
"epoch": 0.841109438830359,
"grad_norm": 6.660660743713379,
"learning_rate": 8.679471433762633e-06,
"loss": 43.5884,
"step": 489
},
{
"epoch": 0.8428294990324662,
"grad_norm": 7.205166339874268,
"learning_rate": 8.672991305312042e-06,
"loss": 43.5902,
"step": 490
},
{
"epoch": 0.8445495592345732,
"grad_norm": 6.969662666320801,
"learning_rate": 8.666497747839413e-06,
"loss": 43.339,
"step": 491
},
{
"epoch": 0.8462696194366803,
"grad_norm": 6.3066229820251465,
"learning_rate": 8.659990785086195e-06,
"loss": 43.8102,
"step": 492
},
{
"epoch": 0.8479896796387874,
"grad_norm": 7.804117202758789,
"learning_rate": 8.653470440842847e-06,
"loss": 43.7162,
"step": 493
},
{
"epoch": 0.8497097398408945,
"grad_norm": 6.339798450469971,
"learning_rate": 8.646936738948747e-06,
"loss": 43.3229,
"step": 494
},
{
"epoch": 0.8514298000430015,
"grad_norm": 8.31767749786377,
"learning_rate": 8.64038970329212e-06,
"loss": 43.8772,
"step": 495
},
{
"epoch": 0.8531498602451085,
"grad_norm": 7.365615367889404,
"learning_rate": 8.633829357809937e-06,
"loss": 43.2881,
"step": 496
},
{
"epoch": 0.8548699204472157,
"grad_norm": 8.474952697753906,
"learning_rate": 8.627255726487831e-06,
"loss": 43.7939,
"step": 497
},
{
"epoch": 0.8565899806493227,
"grad_norm": 6.960323333740234,
"learning_rate": 8.620668833360009e-06,
"loss": 43.499,
"step": 498
},
{
"epoch": 0.8583100408514298,
"grad_norm": 7.6085734367370605,
"learning_rate": 8.614068702509169e-06,
"loss": 43.2025,
"step": 499
},
{
"epoch": 0.8600301010535368,
"grad_norm": 5.805634498596191,
"learning_rate": 8.607455358066404e-06,
"loss": 44.0489,
"step": 500
},
{
"epoch": 0.861750161255644,
"grad_norm": 7.452625274658203,
"learning_rate": 8.600828824211122e-06,
"loss": 43.147,
"step": 501
},
{
"epoch": 0.863470221457751,
"grad_norm": 6.4528584480285645,
"learning_rate": 8.594189125170952e-06,
"loss": 43.5228,
"step": 502
},
{
"epoch": 0.8651902816598581,
"grad_norm": 7.375026702880859,
"learning_rate": 8.587536285221656e-06,
"loss": 44.272,
"step": 503
},
{
"epoch": 0.8669103418619651,
"grad_norm": 6.547872543334961,
"learning_rate": 8.580870328687041e-06,
"loss": 43.4759,
"step": 504
},
{
"epoch": 0.8686304020640723,
"grad_norm": 7.088862419128418,
"learning_rate": 8.574191279938872e-06,
"loss": 43.5528,
"step": 505
},
{
"epoch": 0.8703504622661793,
"grad_norm": 6.6691083908081055,
"learning_rate": 8.567499163396777e-06,
"loss": 43.4488,
"step": 506
},
{
"epoch": 0.8720705224682864,
"grad_norm": 7.416652202606201,
"learning_rate": 8.560794003528171e-06,
"loss": 43.5291,
"step": 507
},
{
"epoch": 0.8737905826703934,
"grad_norm": 5.969050884246826,
"learning_rate": 8.554075824848146e-06,
"loss": 43.5905,
"step": 508
},
{
"epoch": 0.8755106428725006,
"grad_norm": 7.501400470733643,
"learning_rate": 8.5473446519194e-06,
"loss": 44.0266,
"step": 509
},
{
"epoch": 0.8772307030746076,
"grad_norm": 6.1612548828125,
"learning_rate": 8.540600509352139e-06,
"loss": 44.2164,
"step": 510
},
{
"epoch": 0.8789507632767147,
"grad_norm": 7.144975662231445,
"learning_rate": 8.533843421803985e-06,
"loss": 43.1628,
"step": 511
},
{
"epoch": 0.8806708234788218,
"grad_norm": 6.905309200286865,
"learning_rate": 8.527073413979894e-06,
"loss": 43.8717,
"step": 512
},
{
"epoch": 0.8823908836809289,
"grad_norm": 7.095192909240723,
"learning_rate": 8.520290510632055e-06,
"loss": 43.612,
"step": 513
},
{
"epoch": 0.8841109438830359,
"grad_norm": 6.804945468902588,
"learning_rate": 8.51349473655981e-06,
"loss": 43.4807,
"step": 514
},
{
"epoch": 0.8858310040851429,
"grad_norm": 8.927321434020996,
"learning_rate": 8.506686116609553e-06,
"loss": 43.5119,
"step": 515
},
{
"epoch": 0.8875510642872501,
"grad_norm": 6.946136474609375,
"learning_rate": 8.499864675674648e-06,
"loss": 43.1227,
"step": 516
},
{
"epoch": 0.8892711244893571,
"grad_norm": 11.003009796142578,
"learning_rate": 8.493030438695336e-06,
"loss": 43.2844,
"step": 517
},
{
"epoch": 0.8909911846914642,
"grad_norm": 9.151321411132812,
"learning_rate": 8.486183430658639e-06,
"loss": 43.9351,
"step": 518
},
{
"epoch": 0.8927112448935712,
"grad_norm": 10.419197082519531,
"learning_rate": 8.479323676598271e-06,
"loss": 43.4703,
"step": 519
},
{
"epoch": 0.8944313050956784,
"grad_norm": 9.973617553710938,
"learning_rate": 8.472451201594556e-06,
"loss": 43.2196,
"step": 520
},
{
"epoch": 0.8961513652977854,
"grad_norm": 7.706507682800293,
"learning_rate": 8.465566030774314e-06,
"loss": 43.3206,
"step": 521
},
{
"epoch": 0.8978714254998925,
"grad_norm": 7.455542087554932,
"learning_rate": 8.458668189310793e-06,
"loss": 43.2529,
"step": 522
},
{
"epoch": 0.8995914857019995,
"grad_norm": 7.595444202423096,
"learning_rate": 8.451757702423566e-06,
"loss": 43.5217,
"step": 523
},
{
"epoch": 0.9013115459041067,
"grad_norm": 7.0413899421691895,
"learning_rate": 8.444834595378434e-06,
"loss": 43.6686,
"step": 524
},
{
"epoch": 0.9030316061062137,
"grad_norm": 7.505328178405762,
"learning_rate": 8.437898893487345e-06,
"loss": 43.508,
"step": 525
},
{
"epoch": 0.9047516663083208,
"grad_norm": 6.636236667633057,
"learning_rate": 8.430950622108292e-06,
"loss": 43.6455,
"step": 526
},
{
"epoch": 0.9064717265104278,
"grad_norm": 8.014334678649902,
"learning_rate": 8.42398980664523e-06,
"loss": 43.9419,
"step": 527
},
{
"epoch": 0.908191786712535,
"grad_norm": 6.861055374145508,
"learning_rate": 8.417016472547968e-06,
"loss": 44.0091,
"step": 528
},
{
"epoch": 0.909911846914642,
"grad_norm": 7.528046607971191,
"learning_rate": 8.41003064531209e-06,
"loss": 43.9412,
"step": 529
},
{
"epoch": 0.9116319071167491,
"grad_norm": 6.380741596221924,
"learning_rate": 8.403032350478857e-06,
"loss": 43.1688,
"step": 530
},
{
"epoch": 0.9133519673188562,
"grad_norm": 8.248329162597656,
"learning_rate": 8.396021613635116e-06,
"loss": 43.241,
"step": 531
}
],
"logging_steps": 1,
"max_steps": 1743,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 59,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 9.448505275235225e+19,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}