LoRA-Llama-3-MLP / trainer_state.json
secretmoon's picture
Upload 9 files
ca63228 verified
raw
history blame
135 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.8527131782945734,
"eval_steps": 500,
"global_step": 774,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003875968992248062,
"grad_norm": 0.2644451856613159,
"learning_rate": 6.6e-06,
"loss": 2.7134,
"step": 1
},
{
"epoch": 0.007751937984496124,
"grad_norm": 0.3978240191936493,
"learning_rate": 1.32e-05,
"loss": 3.1698,
"step": 2
},
{
"epoch": 0.011627906976744186,
"grad_norm": 0.29106518626213074,
"learning_rate": 1.98e-05,
"loss": 2.8065,
"step": 3
},
{
"epoch": 0.015503875968992248,
"grad_norm": 0.26795685291290283,
"learning_rate": 2.64e-05,
"loss": 2.7225,
"step": 4
},
{
"epoch": 0.01937984496124031,
"grad_norm": 0.18800821900367737,
"learning_rate": 3.3e-05,
"loss": 2.3743,
"step": 5
},
{
"epoch": 0.023255813953488372,
"grad_norm": 0.2275751829147339,
"learning_rate": 3.96e-05,
"loss": 2.474,
"step": 6
},
{
"epoch": 0.027131782945736434,
"grad_norm": 0.37303999066352844,
"learning_rate": 4.6200000000000005e-05,
"loss": 2.9455,
"step": 7
},
{
"epoch": 0.031007751937984496,
"grad_norm": 0.5618475079536438,
"learning_rate": 5.28e-05,
"loss": 3.0509,
"step": 8
},
{
"epoch": 0.03488372093023256,
"grad_norm": 0.2064945548772812,
"learning_rate": 5.94e-05,
"loss": 2.3145,
"step": 9
},
{
"epoch": 0.03875968992248062,
"grad_norm": 0.1987352818250656,
"learning_rate": 6.6e-05,
"loss": 2.6002,
"step": 10
},
{
"epoch": 0.04263565891472868,
"grad_norm": 0.20411251485347748,
"learning_rate": 7.26e-05,
"loss": 2.4743,
"step": 11
},
{
"epoch": 0.046511627906976744,
"grad_norm": 0.1682678759098053,
"learning_rate": 7.92e-05,
"loss": 2.2844,
"step": 12
},
{
"epoch": 0.050387596899224806,
"grad_norm": 0.2902089059352875,
"learning_rate": 8.58e-05,
"loss": 2.7268,
"step": 13
},
{
"epoch": 0.05426356589147287,
"grad_norm": 0.34279119968414307,
"learning_rate": 9.240000000000001e-05,
"loss": 2.8172,
"step": 14
},
{
"epoch": 0.05813953488372093,
"grad_norm": 0.25758570432662964,
"learning_rate": 9.9e-05,
"loss": 2.5574,
"step": 15
},
{
"epoch": 0.06201550387596899,
"grad_norm": 0.22630520164966583,
"learning_rate": 0.0001056,
"loss": 2.4886,
"step": 16
},
{
"epoch": 0.06589147286821706,
"grad_norm": 0.24086053669452667,
"learning_rate": 0.0001122,
"loss": 2.1352,
"step": 17
},
{
"epoch": 0.06976744186046512,
"grad_norm": 0.31760409474372864,
"learning_rate": 0.0001188,
"loss": 2.3757,
"step": 18
},
{
"epoch": 0.07364341085271318,
"grad_norm": 0.27383461594581604,
"learning_rate": 0.0001254,
"loss": 2.2567,
"step": 19
},
{
"epoch": 0.07751937984496124,
"grad_norm": 0.25780704617500305,
"learning_rate": 0.000132,
"loss": 2.497,
"step": 20
},
{
"epoch": 0.08139534883720931,
"grad_norm": 0.20427410304546356,
"learning_rate": 0.00013859999999999998,
"loss": 2.2143,
"step": 21
},
{
"epoch": 0.08527131782945736,
"grad_norm": 0.2947242558002472,
"learning_rate": 0.0001452,
"loss": 2.6828,
"step": 22
},
{
"epoch": 0.08914728682170543,
"grad_norm": 0.23378527164459229,
"learning_rate": 0.0001518,
"loss": 2.3352,
"step": 23
},
{
"epoch": 0.09302325581395349,
"grad_norm": 0.23147471249103546,
"learning_rate": 0.0001584,
"loss": 2.2845,
"step": 24
},
{
"epoch": 0.09689922480620156,
"grad_norm": 0.21582217514514923,
"learning_rate": 0.000165,
"loss": 2.5595,
"step": 25
},
{
"epoch": 0.10077519379844961,
"grad_norm": 0.1932438313961029,
"learning_rate": 0.0001716,
"loss": 2.3431,
"step": 26
},
{
"epoch": 0.10465116279069768,
"grad_norm": 0.18576449155807495,
"learning_rate": 0.00017820000000000002,
"loss": 2.3948,
"step": 27
},
{
"epoch": 0.10852713178294573,
"grad_norm": 0.1673685610294342,
"learning_rate": 0.00018480000000000002,
"loss": 2.2775,
"step": 28
},
{
"epoch": 0.1124031007751938,
"grad_norm": 0.17178471386432648,
"learning_rate": 0.0001914,
"loss": 1.9809,
"step": 29
},
{
"epoch": 0.11627906976744186,
"grad_norm": 0.17103907465934753,
"learning_rate": 0.000198,
"loss": 2.2033,
"step": 30
},
{
"epoch": 0.12015503875968993,
"grad_norm": 0.17570005357265472,
"learning_rate": 0.00020459999999999999,
"loss": 2.1358,
"step": 31
},
{
"epoch": 0.12403100775193798,
"grad_norm": 0.16803690791130066,
"learning_rate": 0.0002112,
"loss": 2.164,
"step": 32
},
{
"epoch": 0.12790697674418605,
"grad_norm": 0.17673198878765106,
"learning_rate": 0.0002178,
"loss": 2.2587,
"step": 33
},
{
"epoch": 0.13178294573643412,
"grad_norm": 0.16170482337474823,
"learning_rate": 0.0002244,
"loss": 2.1818,
"step": 34
},
{
"epoch": 0.13565891472868216,
"grad_norm": 0.1507701873779297,
"learning_rate": 0.00023099999999999998,
"loss": 2.3218,
"step": 35
},
{
"epoch": 0.13953488372093023,
"grad_norm": 0.16273871064186096,
"learning_rate": 0.0002376,
"loss": 2.1462,
"step": 36
},
{
"epoch": 0.1434108527131783,
"grad_norm": 0.15038099884986877,
"learning_rate": 0.00024419999999999997,
"loss": 2.1007,
"step": 37
},
{
"epoch": 0.14728682170542637,
"grad_norm": 0.19681762158870697,
"learning_rate": 0.0002508,
"loss": 2.3083,
"step": 38
},
{
"epoch": 0.1511627906976744,
"grad_norm": 0.15614789724349976,
"learning_rate": 0.0002574,
"loss": 2.0723,
"step": 39
},
{
"epoch": 0.15503875968992248,
"grad_norm": 0.14269952476024628,
"learning_rate": 0.000264,
"loss": 2.1325,
"step": 40
},
{
"epoch": 0.15891472868217055,
"grad_norm": 0.16603954136371613,
"learning_rate": 0.00027059999999999996,
"loss": 2.2062,
"step": 41
},
{
"epoch": 0.16279069767441862,
"grad_norm": 0.16374431550502777,
"learning_rate": 0.00027719999999999996,
"loss": 2.0389,
"step": 42
},
{
"epoch": 0.16666666666666666,
"grad_norm": 0.13400350511074066,
"learning_rate": 0.0002838,
"loss": 1.9439,
"step": 43
},
{
"epoch": 0.17054263565891473,
"grad_norm": 0.13477088510990143,
"learning_rate": 0.0002904,
"loss": 2.3466,
"step": 44
},
{
"epoch": 0.1744186046511628,
"grad_norm": 0.14038704335689545,
"learning_rate": 0.000297,
"loss": 2.3133,
"step": 45
},
{
"epoch": 0.17829457364341086,
"grad_norm": 0.14085592329502106,
"learning_rate": 0.0003036,
"loss": 2.3503,
"step": 46
},
{
"epoch": 0.1821705426356589,
"grad_norm": 0.14309249818325043,
"learning_rate": 0.0003102,
"loss": 2.4691,
"step": 47
},
{
"epoch": 0.18604651162790697,
"grad_norm": 0.1526148021221161,
"learning_rate": 0.0003168,
"loss": 2.1829,
"step": 48
},
{
"epoch": 0.18992248062015504,
"grad_norm": 0.15030066668987274,
"learning_rate": 0.0003234,
"loss": 2.1198,
"step": 49
},
{
"epoch": 0.1937984496124031,
"grad_norm": 0.13567374646663666,
"learning_rate": 0.00033,
"loss": 2.3342,
"step": 50
},
{
"epoch": 0.19767441860465115,
"grad_norm": 0.14513419568538666,
"learning_rate": 0.00032999844662654247,
"loss": 2.1416,
"step": 51
},
{
"epoch": 0.20155038759689922,
"grad_norm": 0.1639217734336853,
"learning_rate": 0.00032999378653541785,
"loss": 2.06,
"step": 52
},
{
"epoch": 0.2054263565891473,
"grad_norm": 0.15029025077819824,
"learning_rate": 0.0003299860198143701,
"loss": 2.3455,
"step": 53
},
{
"epoch": 0.20930232558139536,
"grad_norm": 0.15286165475845337,
"learning_rate": 0.00032997514660963685,
"loss": 2.2418,
"step": 54
},
{
"epoch": 0.2131782945736434,
"grad_norm": 0.17834632098674774,
"learning_rate": 0.00032996116712594723,
"loss": 2.3385,
"step": 55
},
{
"epoch": 0.21705426356589147,
"grad_norm": 0.1593712419271469,
"learning_rate": 0.00032994408162651776,
"loss": 2.2248,
"step": 56
},
{
"epoch": 0.22093023255813954,
"grad_norm": 0.13914993405342102,
"learning_rate": 0.00032992389043304726,
"loss": 2.3005,
"step": 57
},
{
"epoch": 0.2248062015503876,
"grad_norm": 0.13964952528476715,
"learning_rate": 0.0003299005939257111,
"loss": 2.284,
"step": 58
},
{
"epoch": 0.22868217054263565,
"grad_norm": 0.151286780834198,
"learning_rate": 0.0003298741925431539,
"loss": 2.1277,
"step": 59
},
{
"epoch": 0.23255813953488372,
"grad_norm": 0.12762194871902466,
"learning_rate": 0.0003298446867824811,
"loss": 2.046,
"step": 60
},
{
"epoch": 0.2364341085271318,
"grad_norm": 0.13711339235305786,
"learning_rate": 0.00032981207719924994,
"loss": 2.1179,
"step": 61
},
{
"epoch": 0.24031007751937986,
"grad_norm": 0.15553320944309235,
"learning_rate": 0.0003297763644074587,
"loss": 2.3921,
"step": 62
},
{
"epoch": 0.2441860465116279,
"grad_norm": 0.14784659445285797,
"learning_rate": 0.0003297375490795353,
"loss": 2.072,
"step": 63
},
{
"epoch": 0.24806201550387597,
"grad_norm": 0.14965656399726868,
"learning_rate": 0.0003296956319463245,
"loss": 2.314,
"step": 64
},
{
"epoch": 0.25193798449612403,
"grad_norm": 0.14986053109169006,
"learning_rate": 0.0003296506137970745,
"loss": 2.16,
"step": 65
},
{
"epoch": 0.2558139534883721,
"grad_norm": 0.14208155870437622,
"learning_rate": 0.0003296024954794215,
"loss": 1.9537,
"step": 66
},
{
"epoch": 0.2596899224806202,
"grad_norm": 0.1486920267343521,
"learning_rate": 0.0003295512778993743,
"loss": 2.1949,
"step": 67
},
{
"epoch": 0.26356589147286824,
"grad_norm": 0.14816056191921234,
"learning_rate": 0.0003294969620212968,
"loss": 2.274,
"step": 68
},
{
"epoch": 0.26744186046511625,
"grad_norm": 0.14654242992401123,
"learning_rate": 0.00032943954886789016,
"loss": 2.3158,
"step": 69
},
{
"epoch": 0.2713178294573643,
"grad_norm": 0.1430758833885193,
"learning_rate": 0.0003293790395201734,
"loss": 2.2095,
"step": 70
},
{
"epoch": 0.2751937984496124,
"grad_norm": 0.13157042860984802,
"learning_rate": 0.00032931543511746306,
"loss": 2.3029,
"step": 71
},
{
"epoch": 0.27906976744186046,
"grad_norm": 0.15158364176750183,
"learning_rate": 0.0003292487368573518,
"loss": 2.4751,
"step": 72
},
{
"epoch": 0.28294573643410853,
"grad_norm": 0.14293590188026428,
"learning_rate": 0.0003291789459956857,
"loss": 2.2384,
"step": 73
},
{
"epoch": 0.2868217054263566,
"grad_norm": 0.13672257959842682,
"learning_rate": 0.00032910606384654086,
"loss": 2.389,
"step": 74
},
{
"epoch": 0.29069767441860467,
"grad_norm": 0.17896883189678192,
"learning_rate": 0.00032903009178219834,
"loss": 2.26,
"step": 75
},
{
"epoch": 0.29457364341085274,
"grad_norm": 0.1485362946987152,
"learning_rate": 0.0003289510312331187,
"loss": 1.9685,
"step": 76
},
{
"epoch": 0.29844961240310075,
"grad_norm": 0.1551189124584198,
"learning_rate": 0.0003288688836879147,
"loss": 2.0234,
"step": 77
},
{
"epoch": 0.3023255813953488,
"grad_norm": 0.14495842158794403,
"learning_rate": 0.0003287836506933235,
"loss": 2.0907,
"step": 78
},
{
"epoch": 0.3062015503875969,
"grad_norm": 0.16551914811134338,
"learning_rate": 0.0003286953338541776,
"loss": 2.1908,
"step": 79
},
{
"epoch": 0.31007751937984496,
"grad_norm": 0.14821241796016693,
"learning_rate": 0.0003286039348333743,
"loss": 1.9845,
"step": 80
},
{
"epoch": 0.313953488372093,
"grad_norm": 0.17918680608272552,
"learning_rate": 0.0003285094553518446,
"loss": 2.4009,
"step": 81
},
{
"epoch": 0.3178294573643411,
"grad_norm": 0.15694545209407806,
"learning_rate": 0.000328411897188521,
"loss": 2.1897,
"step": 82
},
{
"epoch": 0.32170542635658916,
"grad_norm": 0.17306634783744812,
"learning_rate": 0.0003283112621803035,
"loss": 2.2274,
"step": 83
},
{
"epoch": 0.32558139534883723,
"grad_norm": 0.14742298424243927,
"learning_rate": 0.0003282075522220255,
"loss": 2.1474,
"step": 84
},
{
"epoch": 0.32945736434108525,
"grad_norm": 0.17422489821910858,
"learning_rate": 0.0003281007692664178,
"loss": 2.3424,
"step": 85
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.14565351605415344,
"learning_rate": 0.0003279909153240722,
"loss": 2.1367,
"step": 86
},
{
"epoch": 0.3372093023255814,
"grad_norm": 0.16405850648880005,
"learning_rate": 0.00032787799246340304,
"loss": 2.0857,
"step": 87
},
{
"epoch": 0.34108527131782945,
"grad_norm": 0.1722440868616104,
"learning_rate": 0.0003277620028106089,
"loss": 2.0891,
"step": 88
},
{
"epoch": 0.3449612403100775,
"grad_norm": 0.1501605361700058,
"learning_rate": 0.00032764294854963213,
"loss": 2.1899,
"step": 89
},
{
"epoch": 0.3488372093023256,
"grad_norm": 0.14743047952651978,
"learning_rate": 0.0003275208319221179,
"loss": 2.165,
"step": 90
},
{
"epoch": 0.35271317829457366,
"grad_norm": 0.14289091527462006,
"learning_rate": 0.00032739565522737216,
"loss": 2.0733,
"step": 91
},
{
"epoch": 0.35658914728682173,
"grad_norm": 0.13758860528469086,
"learning_rate": 0.00032726742082231787,
"loss": 2.185,
"step": 92
},
{
"epoch": 0.36046511627906974,
"grad_norm": 0.13637179136276245,
"learning_rate": 0.00032713613112145105,
"loss": 2.2052,
"step": 93
},
{
"epoch": 0.3643410852713178,
"grad_norm": 0.14775022864341736,
"learning_rate": 0.0003270017885967952,
"loss": 2.1404,
"step": 94
},
{
"epoch": 0.3682170542635659,
"grad_norm": 0.1494913548231125,
"learning_rate": 0.0003268643957778546,
"loss": 1.9495,
"step": 95
},
{
"epoch": 0.37209302325581395,
"grad_norm": 0.1620800644159317,
"learning_rate": 0.00032672395525156715,
"loss": 2.0199,
"step": 96
},
{
"epoch": 0.375968992248062,
"grad_norm": 0.15806061029434204,
"learning_rate": 0.0003265804696622549,
"loss": 1.9932,
"step": 97
},
{
"epoch": 0.3798449612403101,
"grad_norm": 0.16337577998638153,
"learning_rate": 0.00032643394171157504,
"loss": 2.1522,
"step": 98
},
{
"epoch": 0.38372093023255816,
"grad_norm": 0.1602049618959427,
"learning_rate": 0.0003262843741584684,
"loss": 2.215,
"step": 99
},
{
"epoch": 0.3875968992248062,
"grad_norm": 0.17709355056285858,
"learning_rate": 0.00032613176981910805,
"loss": 1.8921,
"step": 100
},
{
"epoch": 0.39147286821705424,
"grad_norm": 0.14379487931728363,
"learning_rate": 0.0003259761315668458,
"loss": 2.0393,
"step": 101
},
{
"epoch": 0.3953488372093023,
"grad_norm": 0.15646255016326904,
"learning_rate": 0.0003258174623321583,
"loss": 2.2783,
"step": 102
},
{
"epoch": 0.3992248062015504,
"grad_norm": 0.17745360732078552,
"learning_rate": 0.00032565576510259207,
"loss": 2.0816,
"step": 103
},
{
"epoch": 0.40310077519379844,
"grad_norm": 0.14521199464797974,
"learning_rate": 0.0003254910429227069,
"loss": 2.061,
"step": 104
},
{
"epoch": 0.4069767441860465,
"grad_norm": 0.14865247905254364,
"learning_rate": 0.0003253232988940186,
"loss": 2.1272,
"step": 105
},
{
"epoch": 0.4108527131782946,
"grad_norm": 0.15088239312171936,
"learning_rate": 0.0003251525361749409,
"loss": 1.9842,
"step": 106
},
{
"epoch": 0.41472868217054265,
"grad_norm": 0.15884487330913544,
"learning_rate": 0.00032497875798072557,
"loss": 2.2848,
"step": 107
},
{
"epoch": 0.4186046511627907,
"grad_norm": 0.13841639459133148,
"learning_rate": 0.000324801967583402,
"loss": 2.2508,
"step": 108
},
{
"epoch": 0.42248062015503873,
"grad_norm": 0.15195606648921967,
"learning_rate": 0.00032462216831171607,
"loss": 2.2611,
"step": 109
},
{
"epoch": 0.4263565891472868,
"grad_norm": 0.17446410655975342,
"learning_rate": 0.0003244393635510664,
"loss": 2.3199,
"step": 110
},
{
"epoch": 0.43023255813953487,
"grad_norm": 0.1482134312391281,
"learning_rate": 0.0003242535567434419,
"loss": 2.106,
"step": 111
},
{
"epoch": 0.43410852713178294,
"grad_norm": 0.17814427614212036,
"learning_rate": 0.000324064751387356,
"loss": 2.3546,
"step": 112
},
{
"epoch": 0.437984496124031,
"grad_norm": 0.15309078991413116,
"learning_rate": 0.00032387295103778114,
"loss": 1.9944,
"step": 113
},
{
"epoch": 0.4418604651162791,
"grad_norm": 0.14015917479991913,
"learning_rate": 0.00032367815930608177,
"loss": 2.2504,
"step": 114
},
{
"epoch": 0.44573643410852715,
"grad_norm": 0.14815208315849304,
"learning_rate": 0.0003234803798599466,
"loss": 2.1902,
"step": 115
},
{
"epoch": 0.4496124031007752,
"grad_norm": 0.14948517084121704,
"learning_rate": 0.000323279616423319,
"loss": 2.2916,
"step": 116
},
{
"epoch": 0.45348837209302323,
"grad_norm": 0.18051539361476898,
"learning_rate": 0.00032307587277632753,
"loss": 2.2055,
"step": 117
},
{
"epoch": 0.4573643410852713,
"grad_norm": 0.17824648320674896,
"learning_rate": 0.00032286915275521423,
"loss": 1.9562,
"step": 118
},
{
"epoch": 0.46124031007751937,
"grad_norm": 0.15700489282608032,
"learning_rate": 0.0003226594602522626,
"loss": 2.0874,
"step": 119
},
{
"epoch": 0.46511627906976744,
"grad_norm": 0.15389874577522278,
"learning_rate": 0.00032244679921572447,
"loss": 2.314,
"step": 120
},
{
"epoch": 0.4689922480620155,
"grad_norm": 0.15873467922210693,
"learning_rate": 0.0003222311736497454,
"loss": 2.3363,
"step": 121
},
{
"epoch": 0.4728682170542636,
"grad_norm": 0.15486344695091248,
"learning_rate": 0.0003220125876142893,
"loss": 2.2502,
"step": 122
},
{
"epoch": 0.47674418604651164,
"grad_norm": 0.15193265676498413,
"learning_rate": 0.00032179104522506234,
"loss": 2.1667,
"step": 123
},
{
"epoch": 0.4806201550387597,
"grad_norm": 0.17307482659816742,
"learning_rate": 0.00032156655065343487,
"loss": 2.2361,
"step": 124
},
{
"epoch": 0.4844961240310077,
"grad_norm": 0.17050062119960785,
"learning_rate": 0.00032133910812636346,
"loss": 1.9976,
"step": 125
},
{
"epoch": 0.4883720930232558,
"grad_norm": 0.15616333484649658,
"learning_rate": 0.00032110872192631096,
"loss": 2.2286,
"step": 126
},
{
"epoch": 0.49224806201550386,
"grad_norm": 0.15690878033638,
"learning_rate": 0.00032087539639116596,
"loss": 2.2743,
"step": 127
},
{
"epoch": 0.49612403100775193,
"grad_norm": 0.15837594866752625,
"learning_rate": 0.000320639135914161,
"loss": 2.2109,
"step": 128
},
{
"epoch": 0.5,
"grad_norm": 0.16106249392032623,
"learning_rate": 0.0003203999449437902,
"loss": 2.0109,
"step": 129
},
{
"epoch": 0.5038759689922481,
"grad_norm": 0.15444137156009674,
"learning_rate": 0.00032015782798372504,
"loss": 1.8895,
"step": 130
},
{
"epoch": 0.5077519379844961,
"grad_norm": 0.16820791363716125,
"learning_rate": 0.0003199127895927299,
"loss": 2.0923,
"step": 131
},
{
"epoch": 0.5116279069767442,
"grad_norm": 0.17274028062820435,
"learning_rate": 0.0003196648343845761,
"loss": 2.1964,
"step": 132
},
{
"epoch": 0.5155038759689923,
"grad_norm": 0.17962628602981567,
"learning_rate": 0.00031941396702795505,
"loss": 2.2031,
"step": 133
},
{
"epoch": 0.5193798449612403,
"grad_norm": 0.15597763657569885,
"learning_rate": 0.00031916019224639017,
"loss": 1.9934,
"step": 134
},
{
"epoch": 0.5232558139534884,
"grad_norm": 0.15726816654205322,
"learning_rate": 0.00031890351481814844,
"loss": 2.004,
"step": 135
},
{
"epoch": 0.5271317829457365,
"grad_norm": 0.16479064524173737,
"learning_rate": 0.0003186439395761498,
"loss": 2.024,
"step": 136
},
{
"epoch": 0.5310077519379846,
"grad_norm": 0.1560487300157547,
"learning_rate": 0.00031838147140787656,
"loss": 2.2763,
"step": 137
},
{
"epoch": 0.5348837209302325,
"grad_norm": 0.15172263979911804,
"learning_rate": 0.0003181161152552813,
"loss": 2.2485,
"step": 138
},
{
"epoch": 0.5387596899224806,
"grad_norm": 0.14550605416297913,
"learning_rate": 0.00031784787611469377,
"loss": 2.319,
"step": 139
},
{
"epoch": 0.5426356589147286,
"grad_norm": 0.15684114396572113,
"learning_rate": 0.0003175767590367269,
"loss": 2.1815,
"step": 140
},
{
"epoch": 0.5465116279069767,
"grad_norm": 0.15643136203289032,
"learning_rate": 0.00031730276912618146,
"loss": 2.2636,
"step": 141
},
{
"epoch": 0.5503875968992248,
"grad_norm": 0.15671825408935547,
"learning_rate": 0.00031702591154195023,
"loss": 2.2284,
"step": 142
},
{
"epoch": 0.5542635658914729,
"grad_norm": 0.16105671226978302,
"learning_rate": 0.0003167461914969208,
"loss": 2.2209,
"step": 143
},
{
"epoch": 0.5581395348837209,
"grad_norm": 0.15666881203651428,
"learning_rate": 0.0003164636142578775,
"loss": 2.1917,
"step": 144
},
{
"epoch": 0.562015503875969,
"grad_norm": 0.16629591584205627,
"learning_rate": 0.00031617818514540164,
"loss": 2.2236,
"step": 145
},
{
"epoch": 0.5658914728682171,
"grad_norm": 0.15989044308662415,
"learning_rate": 0.00031588990953377227,
"loss": 2.1643,
"step": 146
},
{
"epoch": 0.5697674418604651,
"grad_norm": 0.15829624235630035,
"learning_rate": 0.0003155987928508643,
"loss": 2.0441,
"step": 147
},
{
"epoch": 0.5736434108527132,
"grad_norm": 0.16886377334594727,
"learning_rate": 0.0003153048405780466,
"loss": 2.3204,
"step": 148
},
{
"epoch": 0.5775193798449613,
"grad_norm": 0.14943009614944458,
"learning_rate": 0.0003150080582500786,
"loss": 2.2651,
"step": 149
},
{
"epoch": 0.5813953488372093,
"grad_norm": 0.14913472533226013,
"learning_rate": 0.0003147084514550064,
"loss": 2.0444,
"step": 150
},
{
"epoch": 0.5852713178294574,
"grad_norm": 0.17818570137023926,
"learning_rate": 0.0003144060258340569,
"loss": 2.389,
"step": 151
},
{
"epoch": 0.5891472868217055,
"grad_norm": 0.15725958347320557,
"learning_rate": 0.0003141007870815326,
"loss": 2.0757,
"step": 152
},
{
"epoch": 0.5930232558139535,
"grad_norm": 0.17349454760551453,
"learning_rate": 0.0003137927409447034,
"loss": 2.0407,
"step": 153
},
{
"epoch": 0.5968992248062015,
"grad_norm": 0.16550661623477936,
"learning_rate": 0.0003134818932236989,
"loss": 2.1548,
"step": 154
},
{
"epoch": 0.6007751937984496,
"grad_norm": 0.17092445492744446,
"learning_rate": 0.00031316824977139905,
"loss": 2.1501,
"step": 155
},
{
"epoch": 0.6046511627906976,
"grad_norm": 0.16574212908744812,
"learning_rate": 0.00031285181649332413,
"loss": 2.3126,
"step": 156
},
{
"epoch": 0.6085271317829457,
"grad_norm": 0.18889443576335907,
"learning_rate": 0.00031253259934752324,
"loss": 2.0974,
"step": 157
},
{
"epoch": 0.6124031007751938,
"grad_norm": 0.14138291776180267,
"learning_rate": 0.0003122106043444623,
"loss": 2.1352,
"step": 158
},
{
"epoch": 0.6162790697674418,
"grad_norm": 0.15668374300003052,
"learning_rate": 0.000311885837546911,
"loss": 2.0748,
"step": 159
},
{
"epoch": 0.6201550387596899,
"grad_norm": 0.19118763506412506,
"learning_rate": 0.0003115583050698283,
"loss": 2.126,
"step": 160
},
{
"epoch": 0.624031007751938,
"grad_norm": 0.19029273092746735,
"learning_rate": 0.0003112280130802476,
"loss": 2.2312,
"step": 161
},
{
"epoch": 0.627906976744186,
"grad_norm": 0.16116315126419067,
"learning_rate": 0.0003108949677971607,
"loss": 2.1802,
"step": 162
},
{
"epoch": 0.6317829457364341,
"grad_norm": 0.15257392823696136,
"learning_rate": 0.00031055917549140013,
"loss": 1.8636,
"step": 163
},
{
"epoch": 0.6356589147286822,
"grad_norm": 0.14367957413196564,
"learning_rate": 0.00031022064248552197,
"loss": 2.2663,
"step": 164
},
{
"epoch": 0.6395348837209303,
"grad_norm": 0.155740886926651,
"learning_rate": 0.0003098793751536858,
"loss": 2.0618,
"step": 165
},
{
"epoch": 0.6434108527131783,
"grad_norm": 0.15573202073574066,
"learning_rate": 0.0003095353799215356,
"loss": 2.147,
"step": 166
},
{
"epoch": 0.6472868217054264,
"grad_norm": 0.17215795814990997,
"learning_rate": 0.0003091886632660781,
"loss": 2.381,
"step": 167
},
{
"epoch": 0.6511627906976745,
"grad_norm": 0.16937744617462158,
"learning_rate": 0.0003088392317155612,
"loss": 2.0703,
"step": 168
},
{
"epoch": 0.6550387596899225,
"grad_norm": 0.16653715074062347,
"learning_rate": 0.000308487091849351,
"loss": 2.3338,
"step": 169
},
{
"epoch": 0.6589147286821705,
"grad_norm": 0.1645222008228302,
"learning_rate": 0.0003081322502978077,
"loss": 2.2578,
"step": 170
},
{
"epoch": 0.6627906976744186,
"grad_norm": 0.17217321693897247,
"learning_rate": 0.0003077747137421609,
"loss": 2.3059,
"step": 171
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.15005135536193848,
"learning_rate": 0.000307414488914384,
"loss": 2.3234,
"step": 172
},
{
"epoch": 0.6705426356589147,
"grad_norm": 0.1632111668586731,
"learning_rate": 0.0003070515825970671,
"loss": 2.3237,
"step": 173
},
{
"epoch": 0.6744186046511628,
"grad_norm": 0.16381341218948364,
"learning_rate": 0.0003066860016232894,
"loss": 2.2046,
"step": 174
},
{
"epoch": 0.6782945736434108,
"grad_norm": 0.1917879581451416,
"learning_rate": 0.00030631775287649077,
"loss": 2.2253,
"step": 175
},
{
"epoch": 0.6821705426356589,
"grad_norm": 0.16225776076316833,
"learning_rate": 0.00030594684329034183,
"loss": 2.1078,
"step": 176
},
{
"epoch": 0.686046511627907,
"grad_norm": 0.16708023846149445,
"learning_rate": 0.00030557327984861353,
"loss": 2.3477,
"step": 177
},
{
"epoch": 0.689922480620155,
"grad_norm": 0.1940573751926422,
"learning_rate": 0.00030519706958504573,
"loss": 2.133,
"step": 178
},
{
"epoch": 0.6937984496124031,
"grad_norm": 0.16478034853935242,
"learning_rate": 0.00030481821958321467,
"loss": 1.9538,
"step": 179
},
{
"epoch": 0.6976744186046512,
"grad_norm": 0.16942182183265686,
"learning_rate": 0.00030443673697639955,
"loss": 2.3055,
"step": 180
},
{
"epoch": 0.7015503875968992,
"grad_norm": 0.16306325793266296,
"learning_rate": 0.0003040526289474484,
"loss": 2.0216,
"step": 181
},
{
"epoch": 0.7054263565891473,
"grad_norm": 0.1586439609527588,
"learning_rate": 0.0003036659027286425,
"loss": 2.242,
"step": 182
},
{
"epoch": 0.7093023255813954,
"grad_norm": 0.18067839741706848,
"learning_rate": 0.00030327656560156065,
"loss": 2.201,
"step": 183
},
{
"epoch": 0.7131782945736435,
"grad_norm": 0.17424075305461884,
"learning_rate": 0.00030288462489694176,
"loss": 2.0878,
"step": 184
},
{
"epoch": 0.7170542635658915,
"grad_norm": 0.17749184370040894,
"learning_rate": 0.0003024900879945468,
"loss": 2.0539,
"step": 185
},
{
"epoch": 0.7209302325581395,
"grad_norm": 0.1625218391418457,
"learning_rate": 0.00030209296232302016,
"loss": 2.252,
"step": 186
},
{
"epoch": 0.7248062015503876,
"grad_norm": 0.1539207398891449,
"learning_rate": 0.0003016932553597493,
"loss": 2.261,
"step": 187
},
{
"epoch": 0.7286821705426356,
"grad_norm": 0.1533818393945694,
"learning_rate": 0.00030129097463072455,
"loss": 2.0823,
"step": 188
},
{
"epoch": 0.7325581395348837,
"grad_norm": 0.16169801354408264,
"learning_rate": 0.0003008861277103968,
"loss": 2.3093,
"step": 189
},
{
"epoch": 0.7364341085271318,
"grad_norm": 0.16680897772312164,
"learning_rate": 0.00030047872222153525,
"loss": 1.8175,
"step": 190
},
{
"epoch": 0.7403100775193798,
"grad_norm": 0.15020348131656647,
"learning_rate": 0.0003000687658350839,
"loss": 2.054,
"step": 191
},
{
"epoch": 0.7441860465116279,
"grad_norm": 0.1465124785900116,
"learning_rate": 0.0002996562662700168,
"loss": 1.9802,
"step": 192
},
{
"epoch": 0.748062015503876,
"grad_norm": 0.1666540950536728,
"learning_rate": 0.00029924123129319315,
"loss": 2.2759,
"step": 193
},
{
"epoch": 0.751937984496124,
"grad_norm": 0.20095106959342957,
"learning_rate": 0.0002988236687192107,
"loss": 2.2055,
"step": 194
},
{
"epoch": 0.7558139534883721,
"grad_norm": 0.15560601651668549,
"learning_rate": 0.00029840358641025877,
"loss": 2.2051,
"step": 195
},
{
"epoch": 0.7596899224806202,
"grad_norm": 0.16015595197677612,
"learning_rate": 0.0002979809922759702,
"loss": 2.1635,
"step": 196
},
{
"epoch": 0.7635658914728682,
"grad_norm": 0.15207257866859436,
"learning_rate": 0.00029755589427327237,
"loss": 2.1105,
"step": 197
},
{
"epoch": 0.7674418604651163,
"grad_norm": 0.18203739821910858,
"learning_rate": 0.0002971283004062375,
"loss": 2.0973,
"step": 198
},
{
"epoch": 0.7713178294573644,
"grad_norm": 0.165207177400589,
"learning_rate": 0.00029669821872593187,
"loss": 2.2404,
"step": 199
},
{
"epoch": 0.7751937984496124,
"grad_norm": 0.16326136887073517,
"learning_rate": 0.000296265657330264,
"loss": 2.0473,
"step": 200
},
{
"epoch": 0.7790697674418605,
"grad_norm": 0.17581120133399963,
"learning_rate": 0.0002958306243638327,
"loss": 2.1448,
"step": 201
},
{
"epoch": 0.7829457364341085,
"grad_norm": 0.16953004896640778,
"learning_rate": 0.0002953931280177732,
"loss": 1.9978,
"step": 202
},
{
"epoch": 0.7868217054263565,
"grad_norm": 0.17670084536075592,
"learning_rate": 0.0002949531765296033,
"loss": 2.3193,
"step": 203
},
{
"epoch": 0.7906976744186046,
"grad_norm": 0.16181717813014984,
"learning_rate": 0.000294510778183068,
"loss": 1.9897,
"step": 204
},
{
"epoch": 0.7945736434108527,
"grad_norm": 0.16492588818073273,
"learning_rate": 0.00029406594130798364,
"loss": 2.2115,
"step": 205
},
{
"epoch": 0.7984496124031008,
"grad_norm": 0.1594804972410202,
"learning_rate": 0.00029361867428008104,
"loss": 2.205,
"step": 206
},
{
"epoch": 0.8023255813953488,
"grad_norm": 0.1583336740732193,
"learning_rate": 0.0002931689855208479,
"loss": 1.9226,
"step": 207
},
{
"epoch": 0.8062015503875969,
"grad_norm": 0.18443737924098969,
"learning_rate": 0.00029271688349737,
"loss": 2.3384,
"step": 208
},
{
"epoch": 0.810077519379845,
"grad_norm": 0.17095071077346802,
"learning_rate": 0.00029226237672217196,
"loss": 2.0513,
"step": 209
},
{
"epoch": 0.813953488372093,
"grad_norm": 0.1657685935497284,
"learning_rate": 0.00029180547375305704,
"loss": 2.0806,
"step": 210
},
{
"epoch": 0.8178294573643411,
"grad_norm": 0.16090308129787445,
"learning_rate": 0.00029134618319294573,
"loss": 2.1466,
"step": 211
},
{
"epoch": 0.8217054263565892,
"grad_norm": 0.1775992214679718,
"learning_rate": 0.00029088451368971396,
"loss": 2.084,
"step": 212
},
{
"epoch": 0.8255813953488372,
"grad_norm": 0.16048915684223175,
"learning_rate": 0.00029042047393603034,
"loss": 1.8937,
"step": 213
},
{
"epoch": 0.8294573643410853,
"grad_norm": 0.17156359553337097,
"learning_rate": 0.0002899540726691922,
"loss": 2.2556,
"step": 214
},
{
"epoch": 0.8333333333333334,
"grad_norm": 0.15980640053749084,
"learning_rate": 0.0002894853186709613,
"loss": 1.9849,
"step": 215
},
{
"epoch": 0.8372093023255814,
"grad_norm": 0.15312501788139343,
"learning_rate": 0.00028901422076739854,
"loss": 2.2545,
"step": 216
},
{
"epoch": 0.8410852713178295,
"grad_norm": 0.14746591448783875,
"learning_rate": 0.0002885407878286976,
"loss": 2.236,
"step": 217
},
{
"epoch": 0.8449612403100775,
"grad_norm": 0.16490989923477173,
"learning_rate": 0.0002880650287690181,
"loss": 2.1494,
"step": 218
},
{
"epoch": 0.8488372093023255,
"grad_norm": 0.1589692234992981,
"learning_rate": 0.0002875869525463173,
"loss": 2.1665,
"step": 219
},
{
"epoch": 0.8527131782945736,
"grad_norm": 0.17528380453586578,
"learning_rate": 0.0002871065681621823,
"loss": 1.9999,
"step": 220
},
{
"epoch": 0.8565891472868217,
"grad_norm": 0.1483374536037445,
"learning_rate": 0.0002866238846616597,
"loss": 2.1133,
"step": 221
},
{
"epoch": 0.8604651162790697,
"grad_norm": 0.1600683033466339,
"learning_rate": 0.00028613891113308566,
"loss": 2.061,
"step": 222
},
{
"epoch": 0.8643410852713178,
"grad_norm": 0.16365991532802582,
"learning_rate": 0.000285651656707915,
"loss": 2.0138,
"step": 223
},
{
"epoch": 0.8682170542635659,
"grad_norm": 0.19533555209636688,
"learning_rate": 0.0002851621305605486,
"loss": 2.3357,
"step": 224
},
{
"epoch": 0.872093023255814,
"grad_norm": 0.17096424102783203,
"learning_rate": 0.0002846703419081614,
"loss": 2.1513,
"step": 225
},
{
"epoch": 0.875968992248062,
"grad_norm": 0.1619122326374054,
"learning_rate": 0.00028417630001052844,
"loss": 2.0947,
"step": 226
},
{
"epoch": 0.8798449612403101,
"grad_norm": 0.16593049466609955,
"learning_rate": 0.00028368001416985054,
"loss": 1.9596,
"step": 227
},
{
"epoch": 0.8837209302325582,
"grad_norm": 0.15227988362312317,
"learning_rate": 0.0002831814937305792,
"loss": 1.9498,
"step": 228
},
{
"epoch": 0.8875968992248062,
"grad_norm": 0.16401511430740356,
"learning_rate": 0.0002826807480792408,
"loss": 2.2326,
"step": 229
},
{
"epoch": 0.8914728682170543,
"grad_norm": 0.20172202587127686,
"learning_rate": 0.0002821777866442596,
"loss": 2.2973,
"step": 230
},
{
"epoch": 0.8953488372093024,
"grad_norm": 0.15986618399620056,
"learning_rate": 0.00028167261889578034,
"loss": 1.9597,
"step": 231
},
{
"epoch": 0.8992248062015504,
"grad_norm": 0.16535168886184692,
"learning_rate": 0.0002811652543454899,
"loss": 1.9582,
"step": 232
},
{
"epoch": 0.9031007751937985,
"grad_norm": 0.18099620938301086,
"learning_rate": 0.0002806557025464385,
"loss": 2.3131,
"step": 233
},
{
"epoch": 0.9069767441860465,
"grad_norm": 0.16411282122135162,
"learning_rate": 0.0002801439730928591,
"loss": 2.0324,
"step": 234
},
{
"epoch": 0.9108527131782945,
"grad_norm": 0.1549196094274521,
"learning_rate": 0.00027963007561998765,
"loss": 2.0878,
"step": 235
},
{
"epoch": 0.9147286821705426,
"grad_norm": 0.1656661331653595,
"learning_rate": 0.00027911401980388093,
"loss": 1.9896,
"step": 236
},
{
"epoch": 0.9186046511627907,
"grad_norm": 0.1635042279958725,
"learning_rate": 0.0002785958153612348,
"loss": 1.9988,
"step": 237
},
{
"epoch": 0.9224806201550387,
"grad_norm": 0.18084381520748138,
"learning_rate": 0.0002780754720492012,
"loss": 2.1724,
"step": 238
},
{
"epoch": 0.9263565891472868,
"grad_norm": 0.1592792123556137,
"learning_rate": 0.0002775529996652041,
"loss": 2.1535,
"step": 239
},
{
"epoch": 0.9302325581395349,
"grad_norm": 0.15081603825092316,
"learning_rate": 0.00027702840804675553,
"loss": 2.2441,
"step": 240
},
{
"epoch": 0.9341085271317829,
"grad_norm": 0.1784433275461197,
"learning_rate": 0.00027650170707127,
"loss": 2.2237,
"step": 241
},
{
"epoch": 0.937984496124031,
"grad_norm": 0.15613843500614166,
"learning_rate": 0.00027597290665587855,
"loss": 1.893,
"step": 242
},
{
"epoch": 0.9418604651162791,
"grad_norm": 0.15053103864192963,
"learning_rate": 0.0002754420167572421,
"loss": 2.282,
"step": 243
},
{
"epoch": 0.9457364341085271,
"grad_norm": 0.15055972337722778,
"learning_rate": 0.0002749090473713641,
"loss": 2.0519,
"step": 244
},
{
"epoch": 0.9496124031007752,
"grad_norm": 0.15554535388946533,
"learning_rate": 0.00027437400853340215,
"loss": 1.9656,
"step": 245
},
{
"epoch": 0.9534883720930233,
"grad_norm": 0.16338570415973663,
"learning_rate": 0.00027383691031747885,
"loss": 2.3299,
"step": 246
},
{
"epoch": 0.9573643410852714,
"grad_norm": 0.16029350459575653,
"learning_rate": 0.0002732977628364927,
"loss": 2.0619,
"step": 247
},
{
"epoch": 0.9612403100775194,
"grad_norm": 0.1667211353778839,
"learning_rate": 0.0002727565762419271,
"loss": 2.1308,
"step": 248
},
{
"epoch": 0.9651162790697675,
"grad_norm": 0.1782902479171753,
"learning_rate": 0.0002722133607236595,
"loss": 1.9959,
"step": 249
},
{
"epoch": 0.9689922480620154,
"grad_norm": 0.15829254686832428,
"learning_rate": 0.0002716681265097696,
"loss": 2.1666,
"step": 250
},
{
"epoch": 0.9728682170542635,
"grad_norm": 0.15275566279888153,
"learning_rate": 0.00027112088386634645,
"loss": 2.2501,
"step": 251
},
{
"epoch": 0.9767441860465116,
"grad_norm": 0.17710180580615997,
"learning_rate": 0.00027057164309729557,
"loss": 2.2943,
"step": 252
},
{
"epoch": 0.9806201550387597,
"grad_norm": 0.14490962028503418,
"learning_rate": 0.0002700204145441446,
"loss": 2.1836,
"step": 253
},
{
"epoch": 0.9844961240310077,
"grad_norm": 0.16613595187664032,
"learning_rate": 0.0002694672085858487,
"loss": 2.1681,
"step": 254
},
{
"epoch": 0.9883720930232558,
"grad_norm": 0.15662476420402527,
"learning_rate": 0.0002689120356385952,
"loss": 1.9142,
"step": 255
},
{
"epoch": 0.9922480620155039,
"grad_norm": 0.1657550036907196,
"learning_rate": 0.0002683549061556074,
"loss": 2.3238,
"step": 256
},
{
"epoch": 0.9961240310077519,
"grad_norm": 0.1951078623533249,
"learning_rate": 0.0002677958306269477,
"loss": 2.368,
"step": 257
},
{
"epoch": 1.0,
"grad_norm": 0.14835147559642792,
"learning_rate": 0.00026723481957932025,
"loss": 1.9971,
"step": 258
},
{
"epoch": 1.003875968992248,
"grad_norm": 0.1541881114244461,
"learning_rate": 0.0002666718835758724,
"loss": 2.1023,
"step": 259
},
{
"epoch": 1.0077519379844961,
"grad_norm": 0.1681731790304184,
"learning_rate": 0.00026610703321599643,
"loss": 2.1575,
"step": 260
},
{
"epoch": 1.0116279069767442,
"grad_norm": 0.1625296175479889,
"learning_rate": 0.0002655402791351292,
"loss": 2.0262,
"step": 261
},
{
"epoch": 1.0155038759689923,
"grad_norm": 0.1610299050807953,
"learning_rate": 0.0002649716320045527,
"loss": 2.1875,
"step": 262
},
{
"epoch": 1.0193798449612403,
"grad_norm": 0.16845448315143585,
"learning_rate": 0.00026440110253119235,
"loss": 2.1463,
"step": 263
},
{
"epoch": 1.0232558139534884,
"grad_norm": 0.15429221093654633,
"learning_rate": 0.00026382870145741594,
"loss": 2.2537,
"step": 264
},
{
"epoch": 1.0271317829457365,
"grad_norm": 0.16575424373149872,
"learning_rate": 0.0002632544395608312,
"loss": 2.1336,
"step": 265
},
{
"epoch": 1.0310077519379846,
"grad_norm": 0.15692879259586334,
"learning_rate": 0.0002626783276540828,
"loss": 1.9072,
"step": 266
},
{
"epoch": 1.0348837209302326,
"grad_norm": 0.20094941556453705,
"learning_rate": 0.00026210037658464886,
"loss": 2.1421,
"step": 267
},
{
"epoch": 1.0387596899224807,
"grad_norm": 0.19174852967262268,
"learning_rate": 0.00026152059723463664,
"loss": 1.9939,
"step": 268
},
{
"epoch": 1.0426356589147288,
"grad_norm": 0.17150387167930603,
"learning_rate": 0.00026093900052057774,
"loss": 2.1736,
"step": 269
},
{
"epoch": 1.0465116279069768,
"grad_norm": 0.15520396828651428,
"learning_rate": 0.0002603555973932225,
"loss": 2.0932,
"step": 270
},
{
"epoch": 1.050387596899225,
"grad_norm": 0.1705746054649353,
"learning_rate": 0.0002597703988373336,
"loss": 2.2486,
"step": 271
},
{
"epoch": 1.054263565891473,
"grad_norm": 0.1639339029788971,
"learning_rate": 0.0002591834158714797,
"loss": 2.2351,
"step": 272
},
{
"epoch": 1.058139534883721,
"grad_norm": 0.17867504060268402,
"learning_rate": 0.00025859465954782744,
"loss": 2.25,
"step": 273
},
{
"epoch": 1.062015503875969,
"grad_norm": 0.20789186656475067,
"learning_rate": 0.0002580041409519339,
"loss": 1.9999,
"step": 274
},
{
"epoch": 1.065891472868217,
"grad_norm": 0.15863926708698273,
"learning_rate": 0.0002574118712025374,
"loss": 2.0538,
"step": 275
},
{
"epoch": 1.069767441860465,
"grad_norm": 0.16435660421848297,
"learning_rate": 0.00025681786145134825,
"loss": 2.0152,
"step": 276
},
{
"epoch": 1.073643410852713,
"grad_norm": 0.1803198754787445,
"learning_rate": 0.000256222122882839,
"loss": 2.3157,
"step": 277
},
{
"epoch": 1.0775193798449612,
"grad_norm": 0.19417981803417206,
"learning_rate": 0.0002556246667140337,
"loss": 2.1564,
"step": 278
},
{
"epoch": 1.003875968992248,
"grad_norm": 0.17067064344882965,
"learning_rate": 0.0002550255041942965,
"loss": 1.8682,
"step": 279
},
{
"epoch": 1.0077519379844961,
"grad_norm": 0.20531415939331055,
"learning_rate": 0.0002544246466051203,
"loss": 1.799,
"step": 280
},
{
"epoch": 1.0116279069767442,
"grad_norm": 0.21423804759979248,
"learning_rate": 0.000253822105259914,
"loss": 1.623,
"step": 281
},
{
"epoch": 1.0155038759689923,
"grad_norm": 0.17934457957744598,
"learning_rate": 0.0002532178915037894,
"loss": 1.8665,
"step": 282
},
{
"epoch": 1.0193798449612403,
"grad_norm": 0.19024011492729187,
"learning_rate": 0.0002526120167133479,
"loss": 1.8837,
"step": 283
},
{
"epoch": 1.0232558139534884,
"grad_norm": 0.26798877120018005,
"learning_rate": 0.0002520044922964662,
"loss": 1.8213,
"step": 284
},
{
"epoch": 1.0271317829457365,
"grad_norm": 0.19613595306873322,
"learning_rate": 0.0002513953296920811,
"loss": 1.8714,
"step": 285
},
{
"epoch": 1.0310077519379846,
"grad_norm": 0.17628167569637299,
"learning_rate": 0.0002507845403699748,
"loss": 1.8617,
"step": 286
},
{
"epoch": 1.0348837209302326,
"grad_norm": 0.19648019969463348,
"learning_rate": 0.0002501721358305584,
"loss": 2.1,
"step": 287
},
{
"epoch": 1.0387596899224807,
"grad_norm": 0.21330788731575012,
"learning_rate": 0.0002495581276046557,
"loss": 1.8563,
"step": 288
},
{
"epoch": 1.0426356589147288,
"grad_norm": 0.18956711888313293,
"learning_rate": 0.00024894252725328583,
"loss": 1.769,
"step": 289
},
{
"epoch": 1.0465116279069768,
"grad_norm": 0.17926643788814545,
"learning_rate": 0.00024832534636744566,
"loss": 1.8329,
"step": 290
},
{
"epoch": 1.050387596899225,
"grad_norm": 0.1793479472398758,
"learning_rate": 0.0002477065965678917,
"loss": 1.5659,
"step": 291
},
{
"epoch": 1.054263565891473,
"grad_norm": 0.19575071334838867,
"learning_rate": 0.00024708628950492116,
"loss": 1.7525,
"step": 292
},
{
"epoch": 1.058139534883721,
"grad_norm": 0.20665033161640167,
"learning_rate": 0.00024646443685815247,
"loss": 1.7979,
"step": 293
},
{
"epoch": 1.062015503875969,
"grad_norm": 0.2565125823020935,
"learning_rate": 0.0002458410503363055,
"loss": 1.8589,
"step": 294
},
{
"epoch": 1.0658914728682172,
"grad_norm": 0.18845273554325104,
"learning_rate": 0.0002452161416769813,
"loss": 1.9473,
"step": 295
},
{
"epoch": 1.069767441860465,
"grad_norm": 0.18714243173599243,
"learning_rate": 0.0002445897226464408,
"loss": 1.7174,
"step": 296
},
{
"epoch": 1.073643410852713,
"grad_norm": 0.20173239707946777,
"learning_rate": 0.0002439618050393833,
"loss": 1.7722,
"step": 297
},
{
"epoch": 1.0775193798449612,
"grad_norm": 0.1991068571805954,
"learning_rate": 0.0002433324006787244,
"loss": 1.8291,
"step": 298
},
{
"epoch": 1.0813953488372092,
"grad_norm": 0.20716890692710876,
"learning_rate": 0.00024270152141537367,
"loss": 1.8379,
"step": 299
},
{
"epoch": 1.0852713178294573,
"grad_norm": 0.18439684808254242,
"learning_rate": 0.00024206917912801104,
"loss": 1.7841,
"step": 300
},
{
"epoch": 1.0891472868217054,
"grad_norm": 0.1679239273071289,
"learning_rate": 0.00024143538572286356,
"loss": 1.8962,
"step": 301
},
{
"epoch": 1.0930232558139534,
"grad_norm": 0.1798442304134369,
"learning_rate": 0.00024080015313348086,
"loss": 1.9051,
"step": 302
},
{
"epoch": 1.0968992248062015,
"grad_norm": 0.18058885633945465,
"learning_rate": 0.00024016349332051082,
"loss": 1.9452,
"step": 303
},
{
"epoch": 1.1007751937984496,
"grad_norm": 0.17939876019954681,
"learning_rate": 0.00023952541827147416,
"loss": 1.7577,
"step": 304
},
{
"epoch": 1.1046511627906976,
"grad_norm": 0.1852078139781952,
"learning_rate": 0.00023888594000053855,
"loss": 1.9129,
"step": 305
},
{
"epoch": 1.1085271317829457,
"grad_norm": 0.18097913265228271,
"learning_rate": 0.0002382450705482929,
"loss": 1.8608,
"step": 306
},
{
"epoch": 1.1124031007751938,
"grad_norm": 0.21096573770046234,
"learning_rate": 0.00023760282198152007,
"loss": 2.0269,
"step": 307
},
{
"epoch": 1.1162790697674418,
"grad_norm": 0.18521122634410858,
"learning_rate": 0.0002369592063929702,
"loss": 1.7748,
"step": 308
},
{
"epoch": 1.12015503875969,
"grad_norm": 0.20764781534671783,
"learning_rate": 0.00023631423590113259,
"loss": 1.5756,
"step": 309
},
{
"epoch": 1.124031007751938,
"grad_norm": 0.17593301832675934,
"learning_rate": 0.00023566792265000772,
"loss": 1.7376,
"step": 310
},
{
"epoch": 1.127906976744186,
"grad_norm": 0.1941826045513153,
"learning_rate": 0.00023502027880887874,
"loss": 2.0077,
"step": 311
},
{
"epoch": 1.1317829457364341,
"grad_norm": 0.1869935393333435,
"learning_rate": 0.00023437131657208193,
"loss": 1.7087,
"step": 312
},
{
"epoch": 1.1356589147286822,
"grad_norm": 0.18788675963878632,
"learning_rate": 0.0002337210481587775,
"loss": 1.5735,
"step": 313
},
{
"epoch": 1.1395348837209303,
"grad_norm": 0.2027387022972107,
"learning_rate": 0.0002330694858127193,
"loss": 1.9059,
"step": 314
},
{
"epoch": 1.1434108527131783,
"grad_norm": 0.18780571222305298,
"learning_rate": 0.0002324166418020244,
"loss": 1.7282,
"step": 315
},
{
"epoch": 1.1472868217054264,
"grad_norm": 0.1976962685585022,
"learning_rate": 0.00023176252841894187,
"loss": 2.018,
"step": 316
},
{
"epoch": 1.1511627906976745,
"grad_norm": 0.19255152344703674,
"learning_rate": 0.00023110715797962177,
"loss": 1.729,
"step": 317
},
{
"epoch": 1.1550387596899225,
"grad_norm": 0.18749909102916718,
"learning_rate": 0.00023045054282388273,
"loss": 1.7549,
"step": 318
},
{
"epoch": 1.1589147286821706,
"grad_norm": 0.22557330131530762,
"learning_rate": 0.00022979269531497995,
"loss": 2.0677,
"step": 319
},
{
"epoch": 1.1627906976744187,
"grad_norm": 0.1746273934841156,
"learning_rate": 0.0002291336278393724,
"loss": 1.7589,
"step": 320
},
{
"epoch": 1.1666666666666667,
"grad_norm": 0.17809656262397766,
"learning_rate": 0.00022847335280648943,
"loss": 1.7851,
"step": 321
},
{
"epoch": 1.1705426356589148,
"grad_norm": 0.19532962143421173,
"learning_rate": 0.0002278118826484972,
"loss": 1.8515,
"step": 322
},
{
"epoch": 1.1744186046511629,
"grad_norm": 0.20081496238708496,
"learning_rate": 0.00022714922982006467,
"loss": 1.7351,
"step": 323
},
{
"epoch": 1.178294573643411,
"grad_norm": 0.22169405221939087,
"learning_rate": 0.000226485406798129,
"loss": 1.8978,
"step": 324
},
{
"epoch": 1.1821705426356588,
"grad_norm": 0.1990557610988617,
"learning_rate": 0.00022582042608166063,
"loss": 1.7837,
"step": 325
},
{
"epoch": 1.1860465116279069,
"grad_norm": 0.19927580654621124,
"learning_rate": 0.00022515430019142793,
"loss": 1.7193,
"step": 326
},
{
"epoch": 1.189922480620155,
"grad_norm": 0.19535136222839355,
"learning_rate": 0.00022448704166976166,
"loss": 2.1294,
"step": 327
},
{
"epoch": 1.193798449612403,
"grad_norm": 0.2005845010280609,
"learning_rate": 0.00022381866308031839,
"loss": 1.9261,
"step": 328
},
{
"epoch": 1.197674418604651,
"grad_norm": 0.20202966034412384,
"learning_rate": 0.00022314917700784438,
"loss": 1.8773,
"step": 329
},
{
"epoch": 1.2015503875968991,
"grad_norm": 0.23102688789367676,
"learning_rate": 0.00022247859605793835,
"loss": 1.7446,
"step": 330
},
{
"epoch": 1.2054263565891472,
"grad_norm": 0.19356843829154968,
"learning_rate": 0.00022180693285681419,
"loss": 1.7991,
"step": 331
},
{
"epoch": 1.2093023255813953,
"grad_norm": 0.22885212302207947,
"learning_rate": 0.0002211342000510633,
"loss": 1.6866,
"step": 332
},
{
"epoch": 1.2131782945736433,
"grad_norm": 0.18752378225326538,
"learning_rate": 0.00022046041030741645,
"loss": 1.8063,
"step": 333
},
{
"epoch": 1.2170542635658914,
"grad_norm": 0.18102799355983734,
"learning_rate": 0.00021978557631250505,
"loss": 2.0372,
"step": 334
},
{
"epoch": 1.2209302325581395,
"grad_norm": 0.19595623016357422,
"learning_rate": 0.0002191097107726228,
"loss": 1.9375,
"step": 335
},
{
"epoch": 1.2248062015503876,
"grad_norm": 0.1951194554567337,
"learning_rate": 0.00021843282641348586,
"loss": 1.9015,
"step": 336
},
{
"epoch": 1.2286821705426356,
"grad_norm": 0.19051988422870636,
"learning_rate": 0.00021775493597999359,
"loss": 1.978,
"step": 337
},
{
"epoch": 1.2325581395348837,
"grad_norm": 0.20707543194293976,
"learning_rate": 0.00021707605223598853,
"loss": 1.648,
"step": 338
},
{
"epoch": 1.2364341085271318,
"grad_norm": 0.1884542554616928,
"learning_rate": 0.00021639618796401595,
"loss": 1.8844,
"step": 339
},
{
"epoch": 1.2403100775193798,
"grad_norm": 0.19425931572914124,
"learning_rate": 0.00021571535596508337,
"loss": 1.8605,
"step": 340
},
{
"epoch": 1.244186046511628,
"grad_norm": 0.20685045421123505,
"learning_rate": 0.0002150335690584193,
"loss": 1.865,
"step": 341
},
{
"epoch": 1.248062015503876,
"grad_norm": 0.1994500607252121,
"learning_rate": 0.00021435084008123202,
"loss": 2.01,
"step": 342
},
{
"epoch": 1.251937984496124,
"grad_norm": 0.18918661773204803,
"learning_rate": 0.00021366718188846795,
"loss": 1.9086,
"step": 343
},
{
"epoch": 1.255813953488372,
"grad_norm": 0.18662329018115997,
"learning_rate": 0.0002129826073525693,
"loss": 1.7579,
"step": 344
},
{
"epoch": 1.2596899224806202,
"grad_norm": 0.21391060948371887,
"learning_rate": 0.00021229712936323206,
"loss": 1.8439,
"step": 345
},
{
"epoch": 1.2635658914728682,
"grad_norm": 0.19121584296226501,
"learning_rate": 0.000211610760827163,
"loss": 1.7058,
"step": 346
},
{
"epoch": 1.2674418604651163,
"grad_norm": 0.19377876818180084,
"learning_rate": 0.00021092351466783695,
"loss": 1.7458,
"step": 347
},
{
"epoch": 1.2713178294573644,
"grad_norm": 0.181095689535141,
"learning_rate": 0.00021023540382525313,
"loss": 1.8254,
"step": 348
},
{
"epoch": 1.2751937984496124,
"grad_norm": 0.18665671348571777,
"learning_rate": 0.00020954644125569186,
"loss": 1.8718,
"step": 349
},
{
"epoch": 1.2790697674418605,
"grad_norm": 0.19250045716762543,
"learning_rate": 0.00020885663993147022,
"loss": 1.9262,
"step": 350
},
{
"epoch": 1.2829457364341086,
"grad_norm": 0.17288216948509216,
"learning_rate": 0.00020816601284069818,
"loss": 1.8269,
"step": 351
},
{
"epoch": 1.2868217054263567,
"grad_norm": 0.18719065189361572,
"learning_rate": 0.00020747457298703388,
"loss": 1.6828,
"step": 352
},
{
"epoch": 1.2906976744186047,
"grad_norm": 0.214838445186615,
"learning_rate": 0.00020678233338943861,
"loss": 1.6086,
"step": 353
},
{
"epoch": 1.2945736434108528,
"grad_norm": 0.20970992743968964,
"learning_rate": 0.0002060893070819321,
"loss": 1.8411,
"step": 354
},
{
"epoch": 1.2984496124031009,
"grad_norm": 0.19377633929252625,
"learning_rate": 0.00020539550711334666,
"loss": 1.7452,
"step": 355
},
{
"epoch": 1.302325581395349,
"grad_norm": 0.21270965039730072,
"learning_rate": 0.0002047009465470818,
"loss": 1.8104,
"step": 356
},
{
"epoch": 1.306201550387597,
"grad_norm": 0.20388975739479065,
"learning_rate": 0.00020400563846085804,
"loss": 1.8727,
"step": 357
},
{
"epoch": 1.310077519379845,
"grad_norm": 0.20084132254123688,
"learning_rate": 0.00020330959594647087,
"loss": 1.833,
"step": 358
},
{
"epoch": 1.3139534883720931,
"grad_norm": 0.21278367936611176,
"learning_rate": 0.00020261283210954418,
"loss": 1.8254,
"step": 359
},
{
"epoch": 1.3178294573643412,
"grad_norm": 0.20003218948841095,
"learning_rate": 0.00020191536006928338,
"loss": 1.6369,
"step": 360
},
{
"epoch": 1.3217054263565893,
"grad_norm": 0.1956341415643692,
"learning_rate": 0.00020121719295822856,
"loss": 1.56,
"step": 361
},
{
"epoch": 1.3255813953488373,
"grad_norm": 0.2056959867477417,
"learning_rate": 0.0002005183439220071,
"loss": 1.9608,
"step": 362
},
{
"epoch": 1.3294573643410852,
"grad_norm": 0.20560531318187714,
"learning_rate": 0.00019981882611908616,
"loss": 1.5359,
"step": 363
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.20920397341251373,
"learning_rate": 0.000199118652720525,
"loss": 1.7916,
"step": 364
},
{
"epoch": 1.3372093023255813,
"grad_norm": 0.20703092217445374,
"learning_rate": 0.00019841783690972695,
"loss": 1.7696,
"step": 365
},
{
"epoch": 1.3410852713178294,
"grad_norm": 0.2005213499069214,
"learning_rate": 0.00019771639188219118,
"loss": 1.755,
"step": 366
},
{
"epoch": 1.3449612403100775,
"grad_norm": 0.19476981461048126,
"learning_rate": 0.00019701433084526416,
"loss": 1.9473,
"step": 367
},
{
"epoch": 1.3488372093023255,
"grad_norm": 0.1986457258462906,
"learning_rate": 0.00019631166701789115,
"loss": 1.5943,
"step": 368
},
{
"epoch": 1.3527131782945736,
"grad_norm": 0.1900886595249176,
"learning_rate": 0.0001956084136303671,
"loss": 1.829,
"step": 369
},
{
"epoch": 1.3565891472868217,
"grad_norm": 0.19796966016292572,
"learning_rate": 0.00019490458392408777,
"loss": 1.6652,
"step": 370
},
{
"epoch": 1.3604651162790697,
"grad_norm": 0.19012434780597687,
"learning_rate": 0.00019420019115130022,
"loss": 2.0372,
"step": 371
},
{
"epoch": 1.3643410852713178,
"grad_norm": 0.20295971632003784,
"learning_rate": 0.0001934952485748534,
"loss": 1.7472,
"step": 372
},
{
"epoch": 1.3682170542635659,
"grad_norm": 0.24137261509895325,
"learning_rate": 0.00019278976946794838,
"loss": 1.7415,
"step": 373
},
{
"epoch": 1.372093023255814,
"grad_norm": 0.19063839316368103,
"learning_rate": 0.00019208376711388837,
"loss": 1.6088,
"step": 374
},
{
"epoch": 1.375968992248062,
"grad_norm": 0.21088172495365143,
"learning_rate": 0.00019137725480582884,
"loss": 2.0018,
"step": 375
},
{
"epoch": 1.37984496124031,
"grad_norm": 0.18817083537578583,
"learning_rate": 0.00019067024584652688,
"loss": 1.7041,
"step": 376
},
{
"epoch": 1.3837209302325582,
"grad_norm": 0.19199052453041077,
"learning_rate": 0.000189962753548091,
"loss": 1.7152,
"step": 377
},
{
"epoch": 1.3875968992248062,
"grad_norm": 0.19253146648406982,
"learning_rate": 0.00018925479123173042,
"loss": 1.8655,
"step": 378
},
{
"epoch": 1.3914728682170543,
"grad_norm": 0.19239702820777893,
"learning_rate": 0.00018854637222750418,
"loss": 1.9605,
"step": 379
},
{
"epoch": 1.3953488372093024,
"grad_norm": 0.19195358455181122,
"learning_rate": 0.00018783750987407015,
"loss": 1.7392,
"step": 380
},
{
"epoch": 1.3992248062015504,
"grad_norm": 0.22022220492362976,
"learning_rate": 0.00018712821751843398,
"loss": 1.7857,
"step": 381
},
{
"epoch": 1.4031007751937985,
"grad_norm": 0.20397743582725525,
"learning_rate": 0.0001864185085156978,
"loss": 1.8367,
"step": 382
},
{
"epoch": 1.4069767441860466,
"grad_norm": 0.21266423165798187,
"learning_rate": 0.00018570839622880848,
"loss": 1.7252,
"step": 383
},
{
"epoch": 1.4108527131782946,
"grad_norm": 0.2537650465965271,
"learning_rate": 0.0001849978940283065,
"loss": 1.6755,
"step": 384
},
{
"epoch": 1.4147286821705427,
"grad_norm": 0.2007053792476654,
"learning_rate": 0.00018428701529207367,
"loss": 1.8301,
"step": 385
},
{
"epoch": 1.4186046511627908,
"grad_norm": 0.2502276301383972,
"learning_rate": 0.0001835757734050817,
"loss": 1.8444,
"step": 386
},
{
"epoch": 1.4224806201550386,
"grad_norm": 0.21285481750965118,
"learning_rate": 0.00018286418175913986,
"loss": 1.8921,
"step": 387
},
{
"epoch": 1.4263565891472867,
"grad_norm": 0.192243754863739,
"learning_rate": 0.0001821522537526431,
"loss": 1.7928,
"step": 388
},
{
"epoch": 1.4302325581395348,
"grad_norm": 0.18463876843452454,
"learning_rate": 0.0001814400027903194,
"loss": 1.7785,
"step": 389
},
{
"epoch": 1.4341085271317828,
"grad_norm": 0.19326354563236237,
"learning_rate": 0.00018072744228297784,
"loss": 1.8843,
"step": 390
},
{
"epoch": 1.437984496124031,
"grad_norm": 0.1868479996919632,
"learning_rate": 0.00018001458564725572,
"loss": 1.5857,
"step": 391
},
{
"epoch": 1.441860465116279,
"grad_norm": 0.23313790559768677,
"learning_rate": 0.00017930144630536607,
"loss": 1.8934,
"step": 392
},
{
"epoch": 1.445736434108527,
"grad_norm": 0.2015128880739212,
"learning_rate": 0.00017858803768484497,
"loss": 1.6785,
"step": 393
},
{
"epoch": 1.449612403100775,
"grad_norm": 0.20901648700237274,
"learning_rate": 0.00017787437321829862,
"loss": 1.6072,
"step": 394
},
{
"epoch": 1.4534883720930232,
"grad_norm": 0.21417102217674255,
"learning_rate": 0.0001771604663431506,
"loss": 1.8706,
"step": 395
},
{
"epoch": 1.4573643410852712,
"grad_norm": 0.26056888699531555,
"learning_rate": 0.00017644633050138862,
"loss": 1.8553,
"step": 396
},
{
"epoch": 1.4612403100775193,
"grad_norm": 0.19383443892002106,
"learning_rate": 0.00017573197913931166,
"loss": 1.7848,
"step": 397
},
{
"epoch": 1.4651162790697674,
"grad_norm": 0.2180427759885788,
"learning_rate": 0.00017501742570727666,
"loss": 1.8704,
"step": 398
},
{
"epoch": 1.4689922480620154,
"grad_norm": 0.2105623483657837,
"learning_rate": 0.00017430268365944518,
"loss": 1.6322,
"step": 399
},
{
"epoch": 1.4728682170542635,
"grad_norm": 0.19639936089515686,
"learning_rate": 0.0001735877664535303,
"loss": 1.7703,
"step": 400
},
{
"epoch": 1.4767441860465116,
"grad_norm": 0.18290045857429504,
"learning_rate": 0.00017287268755054305,
"loss": 1.7469,
"step": 401
},
{
"epoch": 1.4806201550387597,
"grad_norm": 0.20623917877674103,
"learning_rate": 0.00017215746041453913,
"loss": 1.7563,
"step": 402
},
{
"epoch": 1.4844961240310077,
"grad_norm": 0.1958610713481903,
"learning_rate": 0.00017144209851236504,
"loss": 1.8197,
"step": 403
},
{
"epoch": 1.4883720930232558,
"grad_norm": 0.1950949728488922,
"learning_rate": 0.00017072661531340502,
"loss": 1.8786,
"step": 404
},
{
"epoch": 1.4922480620155039,
"grad_norm": 0.21500743925571442,
"learning_rate": 0.0001700110242893271,
"loss": 1.7809,
"step": 405
},
{
"epoch": 1.496124031007752,
"grad_norm": 0.2010657638311386,
"learning_rate": 0.00016929533891382945,
"loss": 1.6953,
"step": 406
},
{
"epoch": 1.5,
"grad_norm": 0.19953016936779022,
"learning_rate": 0.00016857957266238688,
"loss": 1.466,
"step": 407
},
{
"epoch": 1.503875968992248,
"grad_norm": 0.19881819188594818,
"learning_rate": 0.00016786373901199684,
"loss": 1.9139,
"step": 408
},
{
"epoch": 1.5077519379844961,
"grad_norm": 0.1984422504901886,
"learning_rate": 0.000167147851440926,
"loss": 1.9282,
"step": 409
},
{
"epoch": 1.5116279069767442,
"grad_norm": 0.19584600627422333,
"learning_rate": 0.00016643192342845602,
"loss": 1.7367,
"step": 410
},
{
"epoch": 1.5155038759689923,
"grad_norm": 0.1895333230495453,
"learning_rate": 0.00016571596845463044,
"loss": 1.8496,
"step": 411
},
{
"epoch": 1.5193798449612403,
"grad_norm": 0.1868915855884552,
"learning_rate": 0.000165,
"loss": 1.7451,
"step": 412
},
{
"epoch": 1.5232558139534884,
"grad_norm": 0.192902609705925,
"learning_rate": 0.00016428403154536962,
"loss": 1.996,
"step": 413
},
{
"epoch": 1.5271317829457365,
"grad_norm": 0.22422343492507935,
"learning_rate": 0.00016356807657154395,
"loss": 1.8168,
"step": 414
},
{
"epoch": 1.5310077519379846,
"grad_norm": 0.1989205926656723,
"learning_rate": 0.00016285214855907406,
"loss": 1.8789,
"step": 415
},
{
"epoch": 1.5348837209302326,
"grad_norm": 0.19881393015384674,
"learning_rate": 0.0001621362609880032,
"loss": 1.732,
"step": 416
},
{
"epoch": 1.5387596899224807,
"grad_norm": 0.197114959359169,
"learning_rate": 0.00016142042733761317,
"loss": 1.7816,
"step": 417
},
{
"epoch": 1.5426356589147288,
"grad_norm": 0.20128700137138367,
"learning_rate": 0.00016070466108617055,
"loss": 1.8697,
"step": 418
},
{
"epoch": 1.5465116279069768,
"grad_norm": 0.2489478439092636,
"learning_rate": 0.00015998897571067292,
"loss": 1.7559,
"step": 419
},
{
"epoch": 1.550387596899225,
"grad_norm": 0.2148464322090149,
"learning_rate": 0.00015927338468659497,
"loss": 1.6178,
"step": 420
},
{
"epoch": 1.554263565891473,
"grad_norm": 0.18926125764846802,
"learning_rate": 0.00015855790148763498,
"loss": 1.8989,
"step": 421
},
{
"epoch": 1.558139534883721,
"grad_norm": 0.2206331193447113,
"learning_rate": 0.00015784253958546092,
"loss": 1.4578,
"step": 422
},
{
"epoch": 1.562015503875969,
"grad_norm": 0.20707029104232788,
"learning_rate": 0.00015712731244945697,
"loss": 1.8484,
"step": 423
},
{
"epoch": 1.5658914728682172,
"grad_norm": 0.2055550217628479,
"learning_rate": 0.00015641223354646977,
"loss": 1.7355,
"step": 424
},
{
"epoch": 1.5697674418604652,
"grad_norm": 0.22906459867954254,
"learning_rate": 0.00015569731634055482,
"loss": 1.8945,
"step": 425
},
{
"epoch": 1.5736434108527133,
"grad_norm": 0.1993347406387329,
"learning_rate": 0.00015498257429272336,
"loss": 1.6523,
"step": 426
},
{
"epoch": 1.5775193798449614,
"grad_norm": 0.20685455203056335,
"learning_rate": 0.00015426802086068833,
"loss": 1.7329,
"step": 427
},
{
"epoch": 1.5813953488372094,
"grad_norm": 0.2264314740896225,
"learning_rate": 0.00015355366949861135,
"loss": 1.7889,
"step": 428
},
{
"epoch": 1.5852713178294575,
"grad_norm": 0.23605959117412567,
"learning_rate": 0.00015283953365684941,
"loss": 1.8501,
"step": 429
},
{
"epoch": 1.5891472868217056,
"grad_norm": 0.1887684017419815,
"learning_rate": 0.0001521256267817014,
"loss": 1.7191,
"step": 430
},
{
"epoch": 1.5930232558139537,
"grad_norm": 0.1952105611562729,
"learning_rate": 0.00015141196231515508,
"loss": 1.7227,
"step": 431
},
{
"epoch": 1.5968992248062015,
"grad_norm": 0.19459757208824158,
"learning_rate": 0.00015069855369463392,
"loss": 1.4825,
"step": 432
},
{
"epoch": 1.6007751937984496,
"grad_norm": 0.20146332681179047,
"learning_rate": 0.0001499854143527443,
"loss": 1.8256,
"step": 433
},
{
"epoch": 1.6046511627906976,
"grad_norm": 0.2057608962059021,
"learning_rate": 0.00014927255771702219,
"loss": 1.6188,
"step": 434
},
{
"epoch": 1.6085271317829457,
"grad_norm": 0.2795553207397461,
"learning_rate": 0.00014855999720968062,
"loss": 1.5248,
"step": 435
},
{
"epoch": 1.6124031007751938,
"grad_norm": 0.20762935280799866,
"learning_rate": 0.00014784774624735695,
"loss": 1.812,
"step": 436
},
{
"epoch": 1.6162790697674418,
"grad_norm": 0.20336000621318817,
"learning_rate": 0.00014713581824086014,
"loss": 2.0534,
"step": 437
},
{
"epoch": 1.62015503875969,
"grad_norm": 0.20982185006141663,
"learning_rate": 0.00014642422659491837,
"loss": 1.708,
"step": 438
},
{
"epoch": 1.624031007751938,
"grad_norm": 0.19309498369693756,
"learning_rate": 0.00014571298470792635,
"loss": 1.7285,
"step": 439
},
{
"epoch": 1.627906976744186,
"grad_norm": 0.2094258964061737,
"learning_rate": 0.00014500210597169353,
"loss": 1.9404,
"step": 440
},
{
"epoch": 1.6317829457364341,
"grad_norm": 0.20995444059371948,
"learning_rate": 0.00014429160377119152,
"loss": 1.8059,
"step": 441
},
{
"epoch": 1.6356589147286822,
"grad_norm": 0.19962961971759796,
"learning_rate": 0.00014358149148430225,
"loss": 1.5697,
"step": 442
},
{
"epoch": 1.6395348837209303,
"grad_norm": 0.2032284438610077,
"learning_rate": 0.000142871782481566,
"loss": 1.8137,
"step": 443
},
{
"epoch": 1.6434108527131783,
"grad_norm": 0.19651709496974945,
"learning_rate": 0.00014216249012592987,
"loss": 1.7584,
"step": 444
},
{
"epoch": 1.6472868217054264,
"grad_norm": 0.20097319781780243,
"learning_rate": 0.00014145362777249587,
"loss": 1.875,
"step": 445
},
{
"epoch": 1.6511627906976745,
"grad_norm": 0.1974637806415558,
"learning_rate": 0.00014074520876826955,
"loss": 1.9157,
"step": 446
},
{
"epoch": 1.6550387596899225,
"grad_norm": 0.1912589967250824,
"learning_rate": 0.00014003724645190898,
"loss": 1.8198,
"step": 447
},
{
"epoch": 1.6589147286821704,
"grad_norm": 0.22384630143642426,
"learning_rate": 0.00013932975415347314,
"loss": 2.0134,
"step": 448
},
{
"epoch": 1.6627906976744184,
"grad_norm": 0.22261658310890198,
"learning_rate": 0.00013862274519417118,
"loss": 1.9937,
"step": 449
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.20569778978824615,
"learning_rate": 0.0001379162328861116,
"loss": 1.7494,
"step": 450
},
{
"epoch": 1.6705426356589146,
"grad_norm": 0.19439737498760223,
"learning_rate": 0.00013721023053205164,
"loss": 1.8848,
"step": 451
},
{
"epoch": 1.6744186046511627,
"grad_norm": 0.1926531344652176,
"learning_rate": 0.0001365047514251466,
"loss": 1.8646,
"step": 452
},
{
"epoch": 1.6782945736434107,
"grad_norm": 0.2144368290901184,
"learning_rate": 0.00013579980884869977,
"loss": 1.8047,
"step": 453
},
{
"epoch": 1.6821705426356588,
"grad_norm": 0.20255987346172333,
"learning_rate": 0.00013509541607591226,
"loss": 1.8394,
"step": 454
},
{
"epoch": 1.6860465116279069,
"grad_norm": 0.1870235800743103,
"learning_rate": 0.00013439158636963292,
"loss": 1.8102,
"step": 455
},
{
"epoch": 1.689922480620155,
"grad_norm": 0.20020346343517303,
"learning_rate": 0.0001336883329821089,
"loss": 1.8804,
"step": 456
},
{
"epoch": 1.693798449612403,
"grad_norm": 0.1906907856464386,
"learning_rate": 0.00013298566915473581,
"loss": 1.9384,
"step": 457
},
{
"epoch": 1.697674418604651,
"grad_norm": 0.2018539309501648,
"learning_rate": 0.00013228360811780882,
"loss": 1.8863,
"step": 458
},
{
"epoch": 1.7015503875968991,
"grad_norm": 0.1981591433286667,
"learning_rate": 0.00013158216309027305,
"loss": 1.9656,
"step": 459
},
{
"epoch": 1.7054263565891472,
"grad_norm": 0.2033264935016632,
"learning_rate": 0.000130881347279475,
"loss": 1.6587,
"step": 460
},
{
"epoch": 1.7093023255813953,
"grad_norm": 0.1924613118171692,
"learning_rate": 0.00013018117388091386,
"loss": 1.6004,
"step": 461
},
{
"epoch": 1.7131782945736433,
"grad_norm": 0.19790911674499512,
"learning_rate": 0.00012948165607799296,
"loss": 1.8981,
"step": 462
},
{
"epoch": 1.7170542635658914,
"grad_norm": 0.20132358372211456,
"learning_rate": 0.0001287828070417715,
"loss": 1.9168,
"step": 463
},
{
"epoch": 1.7209302325581395,
"grad_norm": 0.19085679948329926,
"learning_rate": 0.00012808463993071661,
"loss": 1.6915,
"step": 464
},
{
"epoch": 1.7248062015503876,
"grad_norm": 0.2047281414270401,
"learning_rate": 0.00012738716789045582,
"loss": 1.6943,
"step": 465
},
{
"epoch": 1.7286821705426356,
"grad_norm": 0.21793071925640106,
"learning_rate": 0.00012669040405352916,
"loss": 1.9838,
"step": 466
},
{
"epoch": 1.7325581395348837,
"grad_norm": 0.21391068398952484,
"learning_rate": 0.00012599436153914198,
"loss": 1.8181,
"step": 467
},
{
"epoch": 1.7364341085271318,
"grad_norm": 0.21556377410888672,
"learning_rate": 0.00012529905345291825,
"loss": 1.8422,
"step": 468
},
{
"epoch": 1.7403100775193798,
"grad_norm": 0.2581706941127777,
"learning_rate": 0.00012460449288665337,
"loss": 1.9742,
"step": 469
},
{
"epoch": 1.744186046511628,
"grad_norm": 0.19443494081497192,
"learning_rate": 0.00012391069291806792,
"loss": 1.6538,
"step": 470
},
{
"epoch": 1.748062015503876,
"grad_norm": 0.1977704018354416,
"learning_rate": 0.00012321766661056138,
"loss": 1.5853,
"step": 471
},
{
"epoch": 1.751937984496124,
"grad_norm": 0.21871508657932281,
"learning_rate": 0.00012252542701296612,
"loss": 1.8603,
"step": 472
},
{
"epoch": 1.755813953488372,
"grad_norm": 0.21219857037067413,
"learning_rate": 0.00012183398715930184,
"loss": 2.0966,
"step": 473
},
{
"epoch": 1.7596899224806202,
"grad_norm": 0.2329765260219574,
"learning_rate": 0.00012114336006852986,
"loss": 1.62,
"step": 474
},
{
"epoch": 1.7635658914728682,
"grad_norm": 0.2001282125711441,
"learning_rate": 0.00012045355874430818,
"loss": 1.6761,
"step": 475
},
{
"epoch": 1.7674418604651163,
"grad_norm": 0.21139495074748993,
"learning_rate": 0.0001197645961747469,
"loss": 1.6687,
"step": 476
},
{
"epoch": 1.7713178294573644,
"grad_norm": 0.20851844549179077,
"learning_rate": 0.00011907648533216309,
"loss": 1.7873,
"step": 477
},
{
"epoch": 1.7751937984496124,
"grad_norm": 0.20665644109249115,
"learning_rate": 0.00011838923917283697,
"loss": 2.0133,
"step": 478
},
{
"epoch": 1.7790697674418605,
"grad_norm": 0.21509003639221191,
"learning_rate": 0.00011770287063676794,
"loss": 1.8063,
"step": 479
},
{
"epoch": 1.7829457364341086,
"grad_norm": 0.20424823462963104,
"learning_rate": 0.0001170173926474307,
"loss": 1.7912,
"step": 480
},
{
"epoch": 1.7868217054263567,
"grad_norm": 0.1964096873998642,
"learning_rate": 0.00011633281811153209,
"loss": 1.7892,
"step": 481
},
{
"epoch": 1.7906976744186047,
"grad_norm": 0.19199204444885254,
"learning_rate": 0.00011564915991876793,
"loss": 1.5681,
"step": 482
},
{
"epoch": 1.7945736434108528,
"grad_norm": 0.20553924143314362,
"learning_rate": 0.00011496643094158072,
"loss": 1.7004,
"step": 483
},
{
"epoch": 1.7984496124031009,
"grad_norm": 0.20590077340602875,
"learning_rate": 0.00011428464403491667,
"loss": 1.9121,
"step": 484
},
{
"epoch": 1.802325581395349,
"grad_norm": 0.2159814089536667,
"learning_rate": 0.00011360381203598404,
"loss": 1.7843,
"step": 485
},
{
"epoch": 1.806201550387597,
"grad_norm": 0.19363482296466827,
"learning_rate": 0.00011292394776401152,
"loss": 2.0043,
"step": 486
},
{
"epoch": 1.810077519379845,
"grad_norm": 0.20251871645450592,
"learning_rate": 0.00011224506402000645,
"loss": 1.7657,
"step": 487
},
{
"epoch": 1.8139534883720931,
"grad_norm": 0.2011270970106125,
"learning_rate": 0.0001115671735865142,
"loss": 1.9705,
"step": 488
},
{
"epoch": 1.8178294573643412,
"grad_norm": 0.20194721221923828,
"learning_rate": 0.00011089028922737721,
"loss": 1.7546,
"step": 489
},
{
"epoch": 1.8217054263565893,
"grad_norm": 0.19558173418045044,
"learning_rate": 0.00011021442368749496,
"loss": 1.7589,
"step": 490
},
{
"epoch": 1.8255813953488373,
"grad_norm": 0.1979272961616516,
"learning_rate": 0.00010953958969258363,
"loss": 1.9451,
"step": 491
},
{
"epoch": 1.8294573643410854,
"grad_norm": 0.1853906363248825,
"learning_rate": 0.00010886579994893669,
"loss": 1.6267,
"step": 492
},
{
"epoch": 1.8333333333333335,
"grad_norm": 0.22289901971817017,
"learning_rate": 0.0001081930671431858,
"loss": 1.8587,
"step": 493
},
{
"epoch": 1.8372093023255816,
"grad_norm": 0.1921209841966629,
"learning_rate": 0.00010752140394206169,
"loss": 1.543,
"step": 494
},
{
"epoch": 1.8410852713178296,
"grad_norm": 0.1961943507194519,
"learning_rate": 0.00010685082299215565,
"loss": 1.7071,
"step": 495
},
{
"epoch": 1.8449612403100775,
"grad_norm": 0.19070090353488922,
"learning_rate": 0.00010618133691968159,
"loss": 1.7627,
"step": 496
},
{
"epoch": 1.8488372093023255,
"grad_norm": 0.1972433626651764,
"learning_rate": 0.00010551295833023834,
"loss": 1.9585,
"step": 497
},
{
"epoch": 1.8527131782945736,
"grad_norm": 0.19546210765838623,
"learning_rate": 0.00010484569980857207,
"loss": 1.7804,
"step": 498
},
{
"epoch": 1.8565891472868217,
"grad_norm": 0.18603426218032837,
"learning_rate": 0.00010417957391833937,
"loss": 1.7544,
"step": 499
},
{
"epoch": 1.8604651162790697,
"grad_norm": 0.21424385905265808,
"learning_rate": 0.00010351459320187102,
"loss": 2.0136,
"step": 500
},
{
"epoch": 1.8643410852713178,
"grad_norm": 0.19600240886211395,
"learning_rate": 0.00010285077017993536,
"loss": 2.1243,
"step": 501
},
{
"epoch": 1.8682170542635659,
"grad_norm": 0.19789910316467285,
"learning_rate": 0.00010218811735150283,
"loss": 1.772,
"step": 502
},
{
"epoch": 1.872093023255814,
"grad_norm": 0.24993658065795898,
"learning_rate": 0.0001015266471935106,
"loss": 1.7537,
"step": 503
},
{
"epoch": 1.875968992248062,
"grad_norm": 0.2263271063566208,
"learning_rate": 0.0001008663721606276,
"loss": 1.6373,
"step": 504
},
{
"epoch": 1.87984496124031,
"grad_norm": 0.19725754857063293,
"learning_rate": 0.00010020730468502006,
"loss": 1.5511,
"step": 505
},
{
"epoch": 1.8837209302325582,
"grad_norm": 0.18990269303321838,
"learning_rate": 9.954945717611734e-05,
"loss": 1.9,
"step": 506
},
{
"epoch": 1.8875968992248062,
"grad_norm": 0.20127974450588226,
"learning_rate": 9.889284202037826e-05,
"loss": 1.8697,
"step": 507
},
{
"epoch": 1.8914728682170543,
"grad_norm": 0.19391153752803802,
"learning_rate": 9.823747158105813e-05,
"loss": 1.64,
"step": 508
},
{
"epoch": 1.8953488372093024,
"grad_norm": 0.19547265768051147,
"learning_rate": 9.758335819797565e-05,
"loss": 1.4242,
"step": 509
},
{
"epoch": 1.8992248062015504,
"grad_norm": 0.19835124909877777,
"learning_rate": 9.693051418728067e-05,
"loss": 1.6888,
"step": 510
},
{
"epoch": 1.9031007751937985,
"grad_norm": 0.2248801290988922,
"learning_rate": 9.627895184122254e-05,
"loss": 1.8677,
"step": 511
},
{
"epoch": 1.9069767441860463,
"grad_norm": 0.20312006771564484,
"learning_rate": 9.562868342791809e-05,
"loss": 1.7185,
"step": 512
},
{
"epoch": 1.9108527131782944,
"grad_norm": 0.2256229966878891,
"learning_rate": 9.497972119112128e-05,
"loss": 1.7775,
"step": 513
},
{
"epoch": 1.9147286821705425,
"grad_norm": 0.2004292607307434,
"learning_rate": 9.433207734999223e-05,
"loss": 1.7049,
"step": 514
},
{
"epoch": 1.9186046511627906,
"grad_norm": 0.19018398225307465,
"learning_rate": 9.368576409886741e-05,
"loss": 1.5486,
"step": 515
},
{
"epoch": 1.9224806201550386,
"grad_norm": 0.2169780731201172,
"learning_rate": 9.304079360702984e-05,
"loss": 1.7943,
"step": 516
},
{
"epoch": 1.9263565891472867,
"grad_norm": 0.18505652248859406,
"learning_rate": 9.239717801847988e-05,
"loss": 1.4213,
"step": 517
},
{
"epoch": 1.9302325581395348,
"grad_norm": 0.19745871424674988,
"learning_rate": 9.175492945170716e-05,
"loss": 1.7484,
"step": 518
},
{
"epoch": 1.9341085271317828,
"grad_norm": 0.19631022214889526,
"learning_rate": 9.111405999946145e-05,
"loss": 1.6813,
"step": 519
},
{
"epoch": 1.937984496124031,
"grad_norm": 0.20178832113742828,
"learning_rate": 9.047458172852592e-05,
"loss": 1.797,
"step": 520
},
{
"epoch": 1.941860465116279,
"grad_norm": 0.19533181190490723,
"learning_rate": 8.983650667948915e-05,
"loss": 1.7296,
"step": 521
},
{
"epoch": 1.945736434108527,
"grad_norm": 0.19685396552085876,
"learning_rate": 8.919984686651917e-05,
"loss": 1.9964,
"step": 522
},
{
"epoch": 1.949612403100775,
"grad_norm": 0.21287135779857635,
"learning_rate": 8.856461427713651e-05,
"loss": 1.7688,
"step": 523
},
{
"epoch": 1.9534883720930232,
"grad_norm": 0.19684074819087982,
"learning_rate": 8.793082087198896e-05,
"loss": 1.7588,
"step": 524
},
{
"epoch": 1.9573643410852712,
"grad_norm": 0.20424464344978333,
"learning_rate": 8.729847858462635e-05,
"loss": 1.804,
"step": 525
},
{
"epoch": 1.9612403100775193,
"grad_norm": 0.2178419530391693,
"learning_rate": 8.666759932127563e-05,
"loss": 1.7219,
"step": 526
},
{
"epoch": 1.9651162790697674,
"grad_norm": 0.20854228734970093,
"learning_rate": 8.603819496061677e-05,
"loss": 1.6547,
"step": 527
},
{
"epoch": 1.9689922480620154,
"grad_norm": 0.19146962463855743,
"learning_rate": 8.54102773535592e-05,
"loss": 1.7676,
"step": 528
},
{
"epoch": 1.9728682170542635,
"grad_norm": 0.20440424978733063,
"learning_rate": 8.478385832301868e-05,
"loss": 1.863,
"step": 529
},
{
"epoch": 1.9767441860465116,
"grad_norm": 0.19717960059642792,
"learning_rate": 8.415894966369449e-05,
"loss": 1.9114,
"step": 530
},
{
"epoch": 1.9806201550387597,
"grad_norm": 0.19504162669181824,
"learning_rate": 8.353556314184756e-05,
"loss": 1.8955,
"step": 531
},
{
"epoch": 1.9844961240310077,
"grad_norm": 0.21640577912330627,
"learning_rate": 8.291371049507882e-05,
"loss": 1.5879,
"step": 532
},
{
"epoch": 1.9883720930232558,
"grad_norm": 0.20818272233009338,
"learning_rate": 8.229340343210828e-05,
"loss": 1.6215,
"step": 533
},
{
"epoch": 1.9922480620155039,
"grad_norm": 0.199616938829422,
"learning_rate": 8.167465363255434e-05,
"loss": 1.8259,
"step": 534
},
{
"epoch": 1.996124031007752,
"grad_norm": 0.21627697348594666,
"learning_rate": 8.105747274671419e-05,
"loss": 2.042,
"step": 535
},
{
"epoch": 2.0,
"grad_norm": 0.2383459061384201,
"learning_rate": 8.044187239534429e-05,
"loss": 1.7217,
"step": 536
},
{
"epoch": 2.003875968992248,
"grad_norm": 0.2019493281841278,
"learning_rate": 7.982786416944163e-05,
"loss": 1.6845,
"step": 537
},
{
"epoch": 2.007751937984496,
"grad_norm": 0.2237274944782257,
"learning_rate": 7.92154596300252e-05,
"loss": 1.4392,
"step": 538
},
{
"epoch": 2.011627906976744,
"grad_norm": 0.22651292383670807,
"learning_rate": 7.860467030791895e-05,
"loss": 1.5396,
"step": 539
},
{
"epoch": 2.0155038759689923,
"grad_norm": 0.2347649782896042,
"learning_rate": 7.799550770353385e-05,
"loss": 1.7393,
"step": 540
},
{
"epoch": 2.0193798449612403,
"grad_norm": 0.22472216188907623,
"learning_rate": 7.738798328665212e-05,
"loss": 1.7223,
"step": 541
},
{
"epoch": 2.0232558139534884,
"grad_norm": 0.20871466398239136,
"learning_rate": 7.678210849621058e-05,
"loss": 1.6426,
"step": 542
},
{
"epoch": 2.0271317829457365,
"grad_norm": 0.20463484525680542,
"learning_rate": 7.617789474008606e-05,
"loss": 1.6959,
"step": 543
},
{
"epoch": 2.0310077519379846,
"grad_norm": 0.21245327591896057,
"learning_rate": 7.55753533948797e-05,
"loss": 1.4701,
"step": 544
},
{
"epoch": 2.0348837209302326,
"grad_norm": 0.20299085974693298,
"learning_rate": 7.497449580570352e-05,
"loss": 1.6918,
"step": 545
},
{
"epoch": 2.0387596899224807,
"grad_norm": 0.20907790958881378,
"learning_rate": 7.437533328596635e-05,
"loss": 2.0656,
"step": 546
},
{
"epoch": 2.0426356589147288,
"grad_norm": 0.20104455947875977,
"learning_rate": 7.377787711716099e-05,
"loss": 1.7928,
"step": 547
},
{
"epoch": 2.046511627906977,
"grad_norm": 0.20086194574832916,
"learning_rate": 7.318213854865179e-05,
"loss": 1.6931,
"step": 548
},
{
"epoch": 2.050387596899225,
"grad_norm": 0.19403457641601562,
"learning_rate": 7.258812879746258e-05,
"loss": 1.648,
"step": 549
},
{
"epoch": 2.054263565891473,
"grad_norm": 0.19223634898662567,
"learning_rate": 7.19958590480661e-05,
"loss": 1.7832,
"step": 550
},
{
"epoch": 2.058139534883721,
"grad_norm": 0.2628048360347748,
"learning_rate": 7.140534045217254e-05,
"loss": 1.8131,
"step": 551
},
{
"epoch": 2.062015503875969,
"grad_norm": 0.19258858263492584,
"learning_rate": 7.08165841285204e-05,
"loss": 2.0739,
"step": 552
},
{
"epoch": 2.065891472868217,
"grad_norm": 0.19849108159542084,
"learning_rate": 7.02296011626664e-05,
"loss": 1.5161,
"step": 553
},
{
"epoch": 2.0697674418604652,
"grad_norm": 0.2288181036710739,
"learning_rate": 6.964440260677757e-05,
"loss": 1.8076,
"step": 554
},
{
"epoch": 2.003875968992248,
"grad_norm": 0.2025528997182846,
"learning_rate": 6.906099947942226e-05,
"loss": 1.3829,
"step": 555
},
{
"epoch": 2.007751937984496,
"grad_norm": 0.2283174991607666,
"learning_rate": 6.847940276536336e-05,
"loss": 1.3538,
"step": 556
},
{
"epoch": 2.011627906976744,
"grad_norm": 0.22030316293239594,
"learning_rate": 6.789962341535114e-05,
"loss": 1.4349,
"step": 557
},
{
"epoch": 2.0155038759689923,
"grad_norm": 0.2190942019224167,
"learning_rate": 6.732167234591726e-05,
"loss": 1.3267,
"step": 558
},
{
"epoch": 2.0193798449612403,
"grad_norm": 0.24970054626464844,
"learning_rate": 6.674556043916883e-05,
"loss": 1.4094,
"step": 559
},
{
"epoch": 2.0232558139534884,
"grad_norm": 0.24456344544887543,
"learning_rate": 6.617129854258406e-05,
"loss": 1.129,
"step": 560
},
{
"epoch": 2.0271317829457365,
"grad_norm": 0.27006039023399353,
"learning_rate": 6.559889746880764e-05,
"loss": 1.1812,
"step": 561
},
{
"epoch": 2.0310077519379846,
"grad_norm": 0.27538713812828064,
"learning_rate": 6.502836799544728e-05,
"loss": 1.1389,
"step": 562
},
{
"epoch": 2.0348837209302326,
"grad_norm": 0.3555956482887268,
"learning_rate": 6.445972086487074e-05,
"loss": 1.5623,
"step": 563
},
{
"epoch": 2.0387596899224807,
"grad_norm": 0.35483232140541077,
"learning_rate": 6.389296678400357e-05,
"loss": 1.1459,
"step": 564
},
{
"epoch": 2.0426356589147288,
"grad_norm": 0.36105233430862427,
"learning_rate": 6.332811642412762e-05,
"loss": 1.2068,
"step": 565
},
{
"epoch": 2.046511627906977,
"grad_norm": 0.25484374165534973,
"learning_rate": 6.276518042067982e-05,
"loss": 1.3551,
"step": 566
},
{
"epoch": 2.050387596899225,
"grad_norm": 0.24508428573608398,
"learning_rate": 6.22041693730523e-05,
"loss": 1.3885,
"step": 567
},
{
"epoch": 2.054263565891473,
"grad_norm": 0.24084091186523438,
"learning_rate": 6.164509384439258e-05,
"loss": 1.2559,
"step": 568
},
{
"epoch": 2.058139534883721,
"grad_norm": 0.24399569630622864,
"learning_rate": 6.10879643614048e-05,
"loss": 1.2449,
"step": 569
},
{
"epoch": 2.062015503875969,
"grad_norm": 0.22874587774276733,
"learning_rate": 6.053279141415124e-05,
"loss": 1.2807,
"step": 570
},
{
"epoch": 2.065891472868217,
"grad_norm": 0.23427356779575348,
"learning_rate": 5.997958545585541e-05,
"loss": 1.223,
"step": 571
},
{
"epoch": 2.0697674418604652,
"grad_norm": 0.23317009210586548,
"learning_rate": 5.9428356902704406e-05,
"loss": 1.3503,
"step": 572
},
{
"epoch": 2.0736434108527133,
"grad_norm": 0.2481788992881775,
"learning_rate": 5.887911613365358e-05,
"loss": 1.5248,
"step": 573
},
{
"epoch": 2.0775193798449614,
"grad_norm": 0.2363855540752411,
"learning_rate": 5.833187349023042e-05,
"loss": 1.3902,
"step": 574
},
{
"epoch": 2.0813953488372094,
"grad_norm": 0.21973107755184174,
"learning_rate": 5.778663927634054e-05,
"loss": 1.1951,
"step": 575
},
{
"epoch": 2.0852713178294575,
"grad_norm": 0.24858590960502625,
"learning_rate": 5.724342375807296e-05,
"loss": 1.306,
"step": 576
},
{
"epoch": 2.0891472868217056,
"grad_norm": 0.2570536136627197,
"learning_rate": 5.670223716350734e-05,
"loss": 1.2423,
"step": 577
},
{
"epoch": 2.0930232558139537,
"grad_norm": 0.2506595849990845,
"learning_rate": 5.616308968252116e-05,
"loss": 1.5181,
"step": 578
},
{
"epoch": 2.0968992248062017,
"grad_norm": 0.2839462459087372,
"learning_rate": 5.562599146659789e-05,
"loss": 1.379,
"step": 579
},
{
"epoch": 2.10077519379845,
"grad_norm": 0.245258629322052,
"learning_rate": 5.509095262863592e-05,
"loss": 1.3119,
"step": 580
},
{
"epoch": 2.104651162790698,
"grad_norm": 0.25935980677604675,
"learning_rate": 5.455798324275785e-05,
"loss": 1.2769,
"step": 581
},
{
"epoch": 2.108527131782946,
"grad_norm": 0.2609214186668396,
"learning_rate": 5.4027093344121484e-05,
"loss": 1.3932,
"step": 582
},
{
"epoch": 2.112403100775194,
"grad_norm": 0.26635345816612244,
"learning_rate": 5.3498292928730014e-05,
"loss": 1.3211,
"step": 583
},
{
"epoch": 2.116279069767442,
"grad_norm": 0.28587472438812256,
"learning_rate": 5.2971591953244504e-05,
"loss": 1.1748,
"step": 584
},
{
"epoch": 2.12015503875969,
"grad_norm": 0.26677605509757996,
"learning_rate": 5.244700033479588e-05,
"loss": 1.2278,
"step": 585
},
{
"epoch": 2.124031007751938,
"grad_norm": 0.26765909790992737,
"learning_rate": 5.192452795079885e-05,
"loss": 1.2706,
"step": 586
},
{
"epoch": 2.1279069767441863,
"grad_norm": 0.28276336193084717,
"learning_rate": 5.140418463876519e-05,
"loss": 1.522,
"step": 587
},
{
"epoch": 2.1317829457364343,
"grad_norm": 0.2587178945541382,
"learning_rate": 5.088598019611909e-05,
"loss": 1.2271,
"step": 588
},
{
"epoch": 2.135658914728682,
"grad_norm": 0.24723570048809052,
"learning_rate": 5.036992438001236e-05,
"loss": 1.1214,
"step": 589
},
{
"epoch": 2.13953488372093,
"grad_norm": 0.2772403359413147,
"learning_rate": 4.985602690714091e-05,
"loss": 1.2235,
"step": 590
},
{
"epoch": 2.143410852713178,
"grad_norm": 0.24144676327705383,
"learning_rate": 4.934429745356153e-05,
"loss": 1.1351,
"step": 591
},
{
"epoch": 2.147286821705426,
"grad_norm": 0.24926505982875824,
"learning_rate": 4.883474565451004e-05,
"loss": 1.3695,
"step": 592
},
{
"epoch": 2.1511627906976742,
"grad_norm": 0.2533572316169739,
"learning_rate": 4.832738110421967e-05,
"loss": 1.2713,
"step": 593
},
{
"epoch": 2.1550387596899223,
"grad_norm": 0.2442951202392578,
"learning_rate": 4.7822213355740394e-05,
"loss": 1.4233,
"step": 594
},
{
"epoch": 2.1589147286821704,
"grad_norm": 0.2528780698776245,
"learning_rate": 4.7319251920759175e-05,
"loss": 1.4764,
"step": 595
},
{
"epoch": 2.1627906976744184,
"grad_norm": 0.26493072509765625,
"learning_rate": 4.681850626942078e-05,
"loss": 1.16,
"step": 596
},
{
"epoch": 2.1666666666666665,
"grad_norm": 0.23339848220348358,
"learning_rate": 4.6319985830149517e-05,
"loss": 1.2937,
"step": 597
},
{
"epoch": 2.1705426356589146,
"grad_norm": 0.25760483741760254,
"learning_rate": 4.582369998947161e-05,
"loss": 1.381,
"step": 598
},
{
"epoch": 2.1744186046511627,
"grad_norm": 0.28579002618789673,
"learning_rate": 4.532965809183861e-05,
"loss": 1.3233,
"step": 599
},
{
"epoch": 2.1782945736434107,
"grad_norm": 0.25108829140663147,
"learning_rate": 4.4837869439451415e-05,
"loss": 1.2818,
"step": 600
},
{
"epoch": 2.182170542635659,
"grad_norm": 0.23311974108219147,
"learning_rate": 4.434834329208507e-05,
"loss": 1.096,
"step": 601
},
{
"epoch": 2.186046511627907,
"grad_norm": 0.2580869495868683,
"learning_rate": 4.3861088866914274e-05,
"loss": 1.0973,
"step": 602
},
{
"epoch": 2.189922480620155,
"grad_norm": 0.24833805859088898,
"learning_rate": 4.337611533834032e-05,
"loss": 1.3511,
"step": 603
},
{
"epoch": 2.193798449612403,
"grad_norm": 0.26399341225624084,
"learning_rate": 4.289343183781769e-05,
"loss": 1.218,
"step": 604
},
{
"epoch": 2.197674418604651,
"grad_norm": 0.2563190162181854,
"learning_rate": 4.241304745368273e-05,
"loss": 1.2484,
"step": 605
},
{
"epoch": 2.201550387596899,
"grad_norm": 0.27118462324142456,
"learning_rate": 4.193497123098196e-05,
"loss": 1.2956,
"step": 606
},
{
"epoch": 2.205426356589147,
"grad_norm": 0.25910139083862305,
"learning_rate": 4.14592121713024e-05,
"loss": 1.3615,
"step": 607
},
{
"epoch": 2.2093023255813953,
"grad_norm": 0.2624959945678711,
"learning_rate": 4.098577923260146e-05,
"loss": 1.5,
"step": 608
},
{
"epoch": 2.2131782945736433,
"grad_norm": 0.26731380820274353,
"learning_rate": 4.051468132903872e-05,
"loss": 1.2989,
"step": 609
},
{
"epoch": 2.2170542635658914,
"grad_norm": 0.2650803327560425,
"learning_rate": 4.004592733080782e-05,
"loss": 1.5101,
"step": 610
},
{
"epoch": 2.2209302325581395,
"grad_norm": 0.23880143463611603,
"learning_rate": 3.957952606396964e-05,
"loss": 1.1257,
"step": 611
},
{
"epoch": 2.2248062015503876,
"grad_norm": 0.25078123807907104,
"learning_rate": 3.911548631028602e-05,
"loss": 1.4061,
"step": 612
},
{
"epoch": 2.2286821705426356,
"grad_norm": 0.2630424201488495,
"learning_rate": 3.865381680705422e-05,
"loss": 1.284,
"step": 613
},
{
"epoch": 2.2325581395348837,
"grad_norm": 0.2509412467479706,
"learning_rate": 3.8194526246942975e-05,
"loss": 1.4054,
"step": 614
},
{
"epoch": 2.2364341085271318,
"grad_norm": 0.24037672579288483,
"learning_rate": 3.7737623277828025e-05,
"loss": 1.3992,
"step": 615
},
{
"epoch": 2.24031007751938,
"grad_norm": 0.26423701643943787,
"learning_rate": 3.728311650263008e-05,
"loss": 1.461,
"step": 616
},
{
"epoch": 2.244186046511628,
"grad_norm": 0.25176694989204407,
"learning_rate": 3.683101447915211e-05,
"loss": 1.402,
"step": 617
},
{
"epoch": 2.248062015503876,
"grad_norm": 0.25743353366851807,
"learning_rate": 3.6381325719918976e-05,
"loss": 1.3649,
"step": 618
},
{
"epoch": 2.251937984496124,
"grad_norm": 0.24827177822589874,
"learning_rate": 3.593405869201637e-05,
"loss": 1.4252,
"step": 619
},
{
"epoch": 2.255813953488372,
"grad_norm": 0.2707473039627075,
"learning_rate": 3.548922181693199e-05,
"loss": 1.3534,
"step": 620
},
{
"epoch": 2.25968992248062,
"grad_norm": 0.2554126977920532,
"learning_rate": 3.504682347039667e-05,
"loss": 1.2142,
"step": 621
},
{
"epoch": 2.2635658914728682,
"grad_norm": 0.2582671642303467,
"learning_rate": 3.460687198222681e-05,
"loss": 1.4812,
"step": 622
},
{
"epoch": 2.2674418604651163,
"grad_norm": 0.24931831657886505,
"learning_rate": 3.416937563616733e-05,
"loss": 1.24,
"step": 623
},
{
"epoch": 2.2713178294573644,
"grad_norm": 0.254974901676178,
"learning_rate": 3.373434266973601e-05,
"loss": 1.3225,
"step": 624
},
{
"epoch": 2.2751937984496124,
"grad_norm": 0.25716301798820496,
"learning_rate": 3.330178127406817e-05,
"loss": 1.3147,
"step": 625
},
{
"epoch": 2.2790697674418605,
"grad_norm": 0.25187844038009644,
"learning_rate": 3.2871699593762476e-05,
"loss": 1.2468,
"step": 626
},
{
"epoch": 2.2829457364341086,
"grad_norm": 0.2536908686161041,
"learning_rate": 3.24441057267276e-05,
"loss": 1.3098,
"step": 627
},
{
"epoch": 2.2868217054263567,
"grad_norm": 0.28247135877609253,
"learning_rate": 3.201900772402978e-05,
"loss": 1.3243,
"step": 628
},
{
"epoch": 2.2906976744186047,
"grad_norm": 0.27947402000427246,
"learning_rate": 3.159641358974126e-05,
"loss": 1.272,
"step": 629
},
{
"epoch": 2.294573643410853,
"grad_norm": 0.27251169085502625,
"learning_rate": 3.117633128078931e-05,
"loss": 1.209,
"step": 630
},
{
"epoch": 2.298449612403101,
"grad_norm": 0.2831606864929199,
"learning_rate": 3.0758768706806846e-05,
"loss": 1.3171,
"step": 631
},
{
"epoch": 2.302325581395349,
"grad_norm": 0.33359742164611816,
"learning_rate": 3.0343733729983185e-05,
"loss": 1.2489,
"step": 632
},
{
"epoch": 2.306201550387597,
"grad_norm": 0.3156187832355499,
"learning_rate": 2.9931234164916155e-05,
"loss": 1.3697,
"step": 633
},
{
"epoch": 2.310077519379845,
"grad_norm": 0.24116098880767822,
"learning_rate": 2.952127777846472e-05,
"loss": 1.2184,
"step": 634
},
{
"epoch": 2.313953488372093,
"grad_norm": 0.30879032611846924,
"learning_rate": 2.911387228960322e-05,
"loss": 1.362,
"step": 635
},
{
"epoch": 2.317829457364341,
"grad_norm": 0.25808271765708923,
"learning_rate": 2.8709025369275442e-05,
"loss": 1.2858,
"step": 636
},
{
"epoch": 2.3217054263565893,
"grad_norm": 0.2646956443786621,
"learning_rate": 2.8306744640250702e-05,
"loss": 1.3307,
"step": 637
},
{
"epoch": 2.3255813953488373,
"grad_norm": 0.25518307089805603,
"learning_rate": 2.790703767697985e-05,
"loss": 1.2729,
"step": 638
},
{
"epoch": 2.3294573643410854,
"grad_norm": 0.2450859695672989,
"learning_rate": 2.7509912005453224e-05,
"loss": 1.067,
"step": 639
},
{
"epoch": 2.3333333333333335,
"grad_norm": 0.26299986243247986,
"learning_rate": 2.7115375103058288e-05,
"loss": 1.3176,
"step": 640
},
{
"epoch": 2.3372093023255816,
"grad_norm": 0.2573457956314087,
"learning_rate": 2.672343439843936e-05,
"loss": 1.2987,
"step": 641
},
{
"epoch": 2.3410852713178296,
"grad_norm": 0.25773999094963074,
"learning_rate": 2.6334097271357512e-05,
"loss": 1.3351,
"step": 642
},
{
"epoch": 2.3449612403100777,
"grad_norm": 0.2760131359100342,
"learning_rate": 2.5947371052551607e-05,
"loss": 1.212,
"step": 643
},
{
"epoch": 2.3488372093023258,
"grad_norm": 0.2653418779373169,
"learning_rate": 2.556326302360044e-05,
"loss": 1.2889,
"step": 644
},
{
"epoch": 2.352713178294574,
"grad_norm": 0.26841187477111816,
"learning_rate": 2.5181780416785284e-05,
"loss": 1.2536,
"step": 645
},
{
"epoch": 2.356589147286822,
"grad_norm": 0.25338879227638245,
"learning_rate": 2.4802930414954242e-05,
"loss": 1.2338,
"step": 646
},
{
"epoch": 2.3604651162790695,
"grad_norm": 0.27084293961524963,
"learning_rate": 2.4426720151386478e-05,
"loss": 1.5218,
"step": 647
},
{
"epoch": 2.3643410852713176,
"grad_norm": 0.27807363867759705,
"learning_rate": 2.4053156709658235e-05,
"loss": 1.355,
"step": 648
},
{
"epoch": 2.3682170542635657,
"grad_norm": 0.25639334321022034,
"learning_rate": 2.3682247123509232e-05,
"loss": 1.1381,
"step": 649
},
{
"epoch": 2.3720930232558137,
"grad_norm": 0.27499181032180786,
"learning_rate": 2.3313998376710624e-05,
"loss": 1.3216,
"step": 650
},
{
"epoch": 2.375968992248062,
"grad_norm": 0.2780856490135193,
"learning_rate": 2.294841740293295e-05,
"loss": 1.165,
"step": 651
},
{
"epoch": 2.37984496124031,
"grad_norm": 0.24923618137836456,
"learning_rate": 2.2585511085615996e-05,
"loss": 1.1451,
"step": 652
},
{
"epoch": 2.383720930232558,
"grad_norm": 0.2662268877029419,
"learning_rate": 2.2225286257839073e-05,
"loss": 1.4445,
"step": 653
},
{
"epoch": 2.387596899224806,
"grad_norm": 0.2631353735923767,
"learning_rate": 2.1867749702192334e-05,
"loss": 1.3777,
"step": 654
},
{
"epoch": 2.391472868217054,
"grad_norm": 0.2878567576408386,
"learning_rate": 2.151290815064901e-05,
"loss": 1.2377,
"step": 655
},
{
"epoch": 2.395348837209302,
"grad_norm": 0.2559579908847809,
"learning_rate": 2.1160768284438775e-05,
"loss": 1.3982,
"step": 656
},
{
"epoch": 2.39922480620155,
"grad_norm": 0.2648860514163971,
"learning_rate": 2.0811336733921926e-05,
"loss": 1.237,
"step": 657
},
{
"epoch": 2.4031007751937983,
"grad_norm": 0.2823598384857178,
"learning_rate": 2.046462007846444e-05,
"loss": 1.2538,
"step": 658
},
{
"epoch": 2.4069767441860463,
"grad_norm": 0.2508406639099121,
"learning_rate": 2.0120624846314217e-05,
"loss": 1.3027,
"step": 659
},
{
"epoch": 2.4108527131782944,
"grad_norm": 0.26299890875816345,
"learning_rate": 1.9779357514478066e-05,
"loss": 1.4021,
"step": 660
},
{
"epoch": 2.4147286821705425,
"grad_norm": 0.2663334608078003,
"learning_rate": 1.944082450859986e-05,
"loss": 1.4515,
"step": 661
},
{
"epoch": 2.4186046511627906,
"grad_norm": 0.25878193974494934,
"learning_rate": 1.910503220283934e-05,
"loss": 1.3473,
"step": 662
},
{
"epoch": 2.4224806201550386,
"grad_norm": 0.25262004137039185,
"learning_rate": 1.8771986919752367e-05,
"loss": 1.3389,
"step": 663
},
{
"epoch": 2.4263565891472867,
"grad_norm": 0.26410895586013794,
"learning_rate": 1.844169493017171e-05,
"loss": 1.3108,
"step": 664
},
{
"epoch": 2.4302325581395348,
"grad_norm": 0.24397780001163483,
"learning_rate": 1.8114162453089036e-05,
"loss": 1.2601,
"step": 665
},
{
"epoch": 2.434108527131783,
"grad_norm": 0.24081014096736908,
"learning_rate": 1.778939565553765e-05,
"loss": 1.2056,
"step": 666
},
{
"epoch": 2.437984496124031,
"grad_norm": 0.271356999874115,
"learning_rate": 1.7467400652476762e-05,
"loss": 1.4752,
"step": 667
},
{
"epoch": 2.441860465116279,
"grad_norm": 0.2456904500722885,
"learning_rate": 1.7148183506675864e-05,
"loss": 1.4085,
"step": 668
},
{
"epoch": 2.445736434108527,
"grad_norm": 0.2549789249897003,
"learning_rate": 1.6831750228600952e-05,
"loss": 1.4168,
"step": 669
},
{
"epoch": 2.449612403100775,
"grad_norm": 0.2592543363571167,
"learning_rate": 1.6518106776301112e-05,
"loss": 1.3086,
"step": 670
},
{
"epoch": 2.453488372093023,
"grad_norm": 0.254226952791214,
"learning_rate": 1.620725905529663e-05,
"loss": 1.2994,
"step": 671
},
{
"epoch": 2.4573643410852712,
"grad_norm": 0.2479204684495926,
"learning_rate": 1.589921291846741e-05,
"loss": 1.2711,
"step": 672
},
{
"epoch": 2.4612403100775193,
"grad_norm": 0.2645493447780609,
"learning_rate": 1.5593974165943074e-05,
"loss": 1.3929,
"step": 673
},
{
"epoch": 2.4651162790697674,
"grad_norm": 0.2703893780708313,
"learning_rate": 1.5291548544993655e-05,
"loss": 1.4931,
"step": 674
},
{
"epoch": 2.4689922480620154,
"grad_norm": 0.2592923939228058,
"learning_rate": 1.4991941749921369e-05,
"loss": 1.2048,
"step": 675
},
{
"epoch": 2.4728682170542635,
"grad_norm": 0.2813677489757538,
"learning_rate": 1.4695159421953419e-05,
"loss": 1.2728,
"step": 676
},
{
"epoch": 2.4767441860465116,
"grad_norm": 0.2619397044181824,
"learning_rate": 1.4401207149135698e-05,
"loss": 1.1384,
"step": 677
},
{
"epoch": 2.4806201550387597,
"grad_norm": 0.27888745069503784,
"learning_rate": 1.411009046622775e-05,
"loss": 1.3301,
"step": 678
},
{
"epoch": 2.4844961240310077,
"grad_norm": 0.2559435963630676,
"learning_rate": 1.3821814854598378e-05,
"loss": 1.2627,
"step": 679
},
{
"epoch": 2.488372093023256,
"grad_norm": 0.25446653366088867,
"learning_rate": 1.3536385742122538e-05,
"loss": 1.128,
"step": 680
},
{
"epoch": 2.492248062015504,
"grad_norm": 0.2939082682132721,
"learning_rate": 1.3253808503079129e-05,
"loss": 1.2637,
"step": 681
},
{
"epoch": 2.496124031007752,
"grad_norm": 0.24750369787216187,
"learning_rate": 1.2974088458049774e-05,
"loss": 1.2881,
"step": 682
},
{
"epoch": 2.5,
"grad_norm": 0.25020653009414673,
"learning_rate": 1.2697230873818587e-05,
"loss": 1.3799,
"step": 683
},
{
"epoch": 2.503875968992248,
"grad_norm": 0.25006988644599915,
"learning_rate": 1.2423240963273123e-05,
"loss": 1.2513,
"step": 684
},
{
"epoch": 2.507751937984496,
"grad_norm": 0.27393513917922974,
"learning_rate": 1.2152123885306202e-05,
"loss": 1.2718,
"step": 685
},
{
"epoch": 2.511627906976744,
"grad_norm": 0.2598724961280823,
"learning_rate": 1.1883884744718704e-05,
"loss": 1.4166,
"step": 686
},
{
"epoch": 2.5155038759689923,
"grad_norm": 0.31768569350242615,
"learning_rate": 1.1618528592123451e-05,
"loss": 1.2255,
"step": 687
},
{
"epoch": 2.5193798449612403,
"grad_norm": 0.2585943341255188,
"learning_rate": 1.135606042385021e-05,
"loss": 1.2544,
"step": 688
},
{
"epoch": 2.5232558139534884,
"grad_norm": 0.2472546398639679,
"learning_rate": 1.1096485181851552e-05,
"loss": 1.1121,
"step": 689
},
{
"epoch": 2.5271317829457365,
"grad_norm": 0.26214876770973206,
"learning_rate": 1.0839807753609787e-05,
"loss": 1.3611,
"step": 690
},
{
"epoch": 2.5310077519379846,
"grad_norm": 0.26527321338653564,
"learning_rate": 1.0586032972044979e-05,
"loss": 0.9816,
"step": 691
},
{
"epoch": 2.5348837209302326,
"grad_norm": 0.2589716911315918,
"learning_rate": 1.0335165615423909e-05,
"loss": 1.1664,
"step": 692
},
{
"epoch": 2.5387596899224807,
"grad_norm": 0.26980432868003845,
"learning_rate": 1.0087210407270105e-05,
"loss": 1.5304,
"step": 693
},
{
"epoch": 2.5426356589147288,
"grad_norm": 0.26864027976989746,
"learning_rate": 9.842172016274983e-06,
"loss": 1.4349,
"step": 694
},
{
"epoch": 2.546511627906977,
"grad_norm": 0.24509486556053162,
"learning_rate": 9.600055056209803e-06,
"loss": 1.3234,
"step": 695
},
{
"epoch": 2.550387596899225,
"grad_norm": 0.24848882853984833,
"learning_rate": 9.360864085838973e-06,
"loss": 1.3879,
"step": 696
},
{
"epoch": 2.554263565891473,
"grad_norm": 0.27096468210220337,
"learning_rate": 9.124603608834071e-06,
"loss": 1.4016,
"step": 697
},
{
"epoch": 2.558139534883721,
"grad_norm": 0.24919958412647247,
"learning_rate": 8.891278073688985e-06,
"loss": 1.3114,
"step": 698
},
{
"epoch": 2.562015503875969,
"grad_norm": 0.2680855691432953,
"learning_rate": 8.660891873636498e-06,
"loss": 1.2027,
"step": 699
},
{
"epoch": 2.565891472868217,
"grad_norm": 0.2762061655521393,
"learning_rate": 8.433449346565108e-06,
"loss": 1.4543,
"step": 700
},
{
"epoch": 2.5697674418604652,
"grad_norm": 0.252043753862381,
"learning_rate": 8.208954774937692e-06,
"loss": 1.2538,
"step": 701
},
{
"epoch": 2.5736434108527133,
"grad_norm": 0.27675777673721313,
"learning_rate": 7.987412385710668e-06,
"loss": 1.2478,
"step": 702
},
{
"epoch": 2.5775193798449614,
"grad_norm": 0.25470390915870667,
"learning_rate": 7.768826350254634e-06,
"loss": 1.3053,
"step": 703
},
{
"epoch": 2.5813953488372094,
"grad_norm": 0.25610047578811646,
"learning_rate": 7.553200784275528e-06,
"loss": 1.366,
"step": 704
},
{
"epoch": 2.5852713178294575,
"grad_norm": 0.25024861097335815,
"learning_rate": 7.3405397477373775e-06,
"loss": 1.3907,
"step": 705
},
{
"epoch": 2.5891472868217056,
"grad_norm": 0.2611881196498871,
"learning_rate": 7.130847244785779e-06,
"loss": 1.2982,
"step": 706
},
{
"epoch": 2.5930232558139537,
"grad_norm": 0.2670976519584656,
"learning_rate": 6.9241272236724684e-06,
"loss": 1.4931,
"step": 707
},
{
"epoch": 2.5968992248062017,
"grad_norm": 0.25510528683662415,
"learning_rate": 6.720383576680977e-06,
"loss": 1.3591,
"step": 708
},
{
"epoch": 2.60077519379845,
"grad_norm": 0.3071907162666321,
"learning_rate": 6.519620140053416e-06,
"loss": 1.4327,
"step": 709
},
{
"epoch": 2.604651162790698,
"grad_norm": 0.27962440252304077,
"learning_rate": 6.321840693918205e-06,
"loss": 1.1472,
"step": 710
},
{
"epoch": 2.608527131782946,
"grad_norm": 0.2648572623729706,
"learning_rate": 6.127048962218875e-06,
"loss": 1.5193,
"step": 711
},
{
"epoch": 2.612403100775194,
"grad_norm": 0.2580854892730713,
"learning_rate": 5.9352486126439925e-06,
"loss": 1.4547,
"step": 712
},
{
"epoch": 2.616279069767442,
"grad_norm": 0.26917338371276855,
"learning_rate": 5.746443256558062e-06,
"loss": 1.4236,
"step": 713
},
{
"epoch": 2.62015503875969,
"grad_norm": 0.2621963620185852,
"learning_rate": 5.560636448933566e-06,
"loss": 1.4411,
"step": 714
},
{
"epoch": 2.624031007751938,
"grad_norm": 0.297036737203598,
"learning_rate": 5.377831688283975e-06,
"loss": 1.4621,
"step": 715
},
{
"epoch": 2.6279069767441863,
"grad_norm": 0.2709357440471649,
"learning_rate": 5.198032416597942e-06,
"loss": 1.3117,
"step": 716
},
{
"epoch": 2.6317829457364343,
"grad_norm": 0.2600153982639313,
"learning_rate": 5.021242019274458e-06,
"loss": 1.2619,
"step": 717
},
{
"epoch": 2.6356589147286824,
"grad_norm": 0.28544628620147705,
"learning_rate": 4.847463825059103e-06,
"loss": 1.2316,
"step": 718
},
{
"epoch": 2.6395348837209305,
"grad_norm": 0.25166937708854675,
"learning_rate": 4.6767011059813755e-06,
"loss": 1.2125,
"step": 719
},
{
"epoch": 2.6434108527131785,
"grad_norm": 0.2691721022129059,
"learning_rate": 4.5089570772931255e-06,
"loss": 1.2869,
"step": 720
},
{
"epoch": 2.6472868217054266,
"grad_norm": 0.28785765171051025,
"learning_rate": 4.344234897407919e-06,
"loss": 1.3122,
"step": 721
},
{
"epoch": 2.6511627906976747,
"grad_norm": 0.24798186123371124,
"learning_rate": 4.182537667841703e-06,
"loss": 1.1715,
"step": 722
},
{
"epoch": 2.6550387596899228,
"grad_norm": 0.28408151865005493,
"learning_rate": 4.023868433154223e-06,
"loss": 1.3831,
"step": 723
},
{
"epoch": 2.6589147286821704,
"grad_norm": 0.2505200207233429,
"learning_rate": 3.868230180891944e-06,
"loss": 1.3047,
"step": 724
},
{
"epoch": 2.6627906976744184,
"grad_norm": 0.24836868047714233,
"learning_rate": 3.7156258415315505e-06,
"loss": 1.2543,
"step": 725
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.2672933340072632,
"learning_rate": 3.566058288424942e-06,
"loss": 1.4565,
"step": 726
},
{
"epoch": 2.6705426356589146,
"grad_norm": 0.24122579395771027,
"learning_rate": 3.4195303377450594e-06,
"loss": 1.2132,
"step": 727
},
{
"epoch": 2.6744186046511627,
"grad_norm": 0.31489935517311096,
"learning_rate": 3.2760447484328548e-06,
"loss": 1.2152,
"step": 728
},
{
"epoch": 2.6782945736434107,
"grad_norm": 0.24762488901615143,
"learning_rate": 3.1356042221453766e-06,
"loss": 1.1724,
"step": 729
},
{
"epoch": 2.682170542635659,
"grad_norm": 0.2507333755493164,
"learning_rate": 2.998211403204843e-06,
"loss": 1.4529,
"step": 730
},
{
"epoch": 2.686046511627907,
"grad_norm": 0.27383190393447876,
"learning_rate": 2.863868878548979e-06,
"loss": 1.3259,
"step": 731
},
{
"epoch": 2.689922480620155,
"grad_norm": 0.24372157454490662,
"learning_rate": 2.7325791776821445e-06,
"loss": 1.2469,
"step": 732
},
{
"epoch": 2.693798449612403,
"grad_norm": 0.25599318742752075,
"learning_rate": 2.60434477262785e-06,
"loss": 1.3733,
"step": 733
},
{
"epoch": 2.697674418604651,
"grad_norm": 0.2459976226091385,
"learning_rate": 2.4791680778820455e-06,
"loss": 1.3466,
"step": 734
},
{
"epoch": 2.701550387596899,
"grad_norm": 0.2694660723209381,
"learning_rate": 2.357051450367873e-06,
"loss": 1.5102,
"step": 735
},
{
"epoch": 2.705426356589147,
"grad_norm": 0.27480435371398926,
"learning_rate": 2.2379971893911144e-06,
"loss": 1.4782,
"step": 736
},
{
"epoch": 2.7093023255813953,
"grad_norm": 0.253065288066864,
"learning_rate": 2.122007536596961e-06,
"loss": 1.1246,
"step": 737
},
{
"epoch": 2.7131782945736433,
"grad_norm": 0.269283652305603,
"learning_rate": 2.0090846759278064e-06,
"loss": 1.5083,
"step": 738
},
{
"epoch": 2.7170542635658914,
"grad_norm": 0.2605230510234833,
"learning_rate": 1.8992307335821763e-06,
"loss": 1.2764,
"step": 739
},
{
"epoch": 2.7209302325581395,
"grad_norm": 0.2701490819454193,
"learning_rate": 1.7924477779745368e-06,
"loss": 1.2148,
"step": 740
},
{
"epoch": 2.7248062015503876,
"grad_norm": 0.2751643657684326,
"learning_rate": 1.688737819696533e-06,
"loss": 1.3495,
"step": 741
},
{
"epoch": 2.7286821705426356,
"grad_norm": 0.25778815150260925,
"learning_rate": 1.5881028114790319e-06,
"loss": 1.3935,
"step": 742
},
{
"epoch": 2.7325581395348837,
"grad_norm": 0.26106202602386475,
"learning_rate": 1.4905446481553752e-06,
"loss": 1.3074,
"step": 743
},
{
"epoch": 2.7364341085271318,
"grad_norm": 0.30311137437820435,
"learning_rate": 1.3960651666257135e-06,
"loss": 1.5208,
"step": 744
},
{
"epoch": 2.74031007751938,
"grad_norm": 0.2914187014102936,
"learning_rate": 1.304666145822383e-06,
"loss": 1.2835,
"step": 745
},
{
"epoch": 2.744186046511628,
"grad_norm": 0.2645190358161926,
"learning_rate": 1.2163493066764564e-06,
"loss": 1.4204,
"step": 746
},
{
"epoch": 2.748062015503876,
"grad_norm": 0.3073311448097229,
"learning_rate": 1.1311163120853002e-06,
"loss": 1.4899,
"step": 747
},
{
"epoch": 2.751937984496124,
"grad_norm": 0.25737103819847107,
"learning_rate": 1.0489687668813048e-06,
"loss": 1.2717,
"step": 748
},
{
"epoch": 2.755813953488372,
"grad_norm": 0.27693504095077515,
"learning_rate": 9.6990821780164e-07,
"loss": 1.4203,
"step": 749
},
{
"epoch": 2.75968992248062,
"grad_norm": 0.3176767826080322,
"learning_rate": 8.939361534591472e-07,
"loss": 1.4588,
"step": 750
},
{
"epoch": 2.7635658914728682,
"grad_norm": 0.34483450651168823,
"learning_rate": 8.21054004314275e-07,
"loss": 1.2022,
"step": 751
},
{
"epoch": 2.7674418604651163,
"grad_norm": 0.33260253071784973,
"learning_rate": 7.512631426481869e-07,
"loss": 1.2149,
"step": 752
},
{
"epoch": 2.7713178294573644,
"grad_norm": 0.277261346578598,
"learning_rate": 6.845648825369143e-07,
"loss": 1.311,
"step": 753
},
{
"epoch": 2.7751937984496124,
"grad_norm": 0.32710519433021545,
"learning_rate": 6.209604798265894e-07,
"loss": 1.5075,
"step": 754
},
{
"epoch": 2.7790697674418605,
"grad_norm": 0.2529277205467224,
"learning_rate": 5.604511321098504e-07,
"loss": 1.4405,
"step": 755
},
{
"epoch": 2.7829457364341086,
"grad_norm": 0.273580938577652,
"learning_rate": 5.030379787032185e-07,
"loss": 1.3051,
"step": 756
},
{
"epoch": 2.7868217054263567,
"grad_norm": 0.2600751519203186,
"learning_rate": 4.487221006257197e-07,
"loss": 1.3619,
"step": 757
},
{
"epoch": 2.7906976744186047,
"grad_norm": 0.2721461057662964,
"learning_rate": 3.9750452057847775e-07,
"loss": 1.2126,
"step": 758
},
{
"epoch": 2.794573643410853,
"grad_norm": 0.24929043650627136,
"learning_rate": 3.493862029254979e-07,
"loss": 1.4246,
"step": 759
},
{
"epoch": 2.798449612403101,
"grad_norm": 0.25558316707611084,
"learning_rate": 3.043680536754767e-07,
"loss": 1.4952,
"step": 760
},
{
"epoch": 2.802325581395349,
"grad_norm": 0.3264126181602478,
"learning_rate": 2.624509204647285e-07,
"loss": 1.4849,
"step": 761
},
{
"epoch": 2.806201550387597,
"grad_norm": 0.25953492522239685,
"learning_rate": 2.236355925413036e-07,
"loss": 1.2316,
"step": 762
},
{
"epoch": 2.810077519379845,
"grad_norm": 0.25779008865356445,
"learning_rate": 1.8792280075005829e-07,
"loss": 1.2583,
"step": 763
},
{
"epoch": 2.813953488372093,
"grad_norm": 0.263167142868042,
"learning_rate": 1.5531321751887928e-07,
"loss": 1.507,
"step": 764
},
{
"epoch": 2.817829457364341,
"grad_norm": 0.2614395320415497,
"learning_rate": 1.2580745684609872e-07,
"loss": 1.2681,
"step": 765
},
{
"epoch": 2.8217054263565893,
"grad_norm": 0.2607451379299164,
"learning_rate": 9.940607428888027e-08,
"loss": 1.3721,
"step": 766
},
{
"epoch": 2.8255813953488373,
"grad_norm": 0.27197694778442383,
"learning_rate": 7.610956695275895e-08,
"loss": 1.3152,
"step": 767
},
{
"epoch": 2.8294573643410854,
"grad_norm": 0.25525933504104614,
"learning_rate": 5.591837348228046e-08,
"loss": 1.3361,
"step": 768
},
{
"epoch": 2.8333333333333335,
"grad_norm": 0.2669144868850708,
"learning_rate": 3.883287405277602e-08,
"loss": 1.2546,
"step": 769
},
{
"epoch": 2.8372093023255816,
"grad_norm": 0.2557834982872009,
"learning_rate": 2.4853390363163142e-08,
"loss": 1.1581,
"step": 770
},
{
"epoch": 2.8410852713178296,
"grad_norm": 0.26671895384788513,
"learning_rate": 1.398018562993708e-08,
"loss": 1.2647,
"step": 771
},
{
"epoch": 2.8449612403100772,
"grad_norm": 0.26321831345558167,
"learning_rate": 6.213464582133232e-09,
"loss": 1.2884,
"step": 772
},
{
"epoch": 2.8488372093023253,
"grad_norm": 0.2515983581542969,
"learning_rate": 1.5533734575534641e-09,
"loss": 1.412,
"step": 773
},
{
"epoch": 2.8527131782945734,
"grad_norm": 0.29071158170700073,
"learning_rate": 0.0,
"loss": 1.3587,
"step": 774
}
],
"logging_steps": 1,
"max_steps": 774,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 258,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.6656805872756326e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}