rearc-20k-ablation / trainer_state.json
lillian039's picture
Model save
438f7ca verified
raw
history blame
No virus
52.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9932659932659933,
"eval_steps": 500,
"global_step": 296,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006734006734006734,
"grad_norm": 10.587559734000028,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.3649,
"step": 1
},
{
"epoch": 0.013468013468013467,
"grad_norm": 9.693900007266853,
"learning_rate": 6.666666666666667e-07,
"loss": 0.3686,
"step": 2
},
{
"epoch": 0.020202020202020204,
"grad_norm": 10.061524524104307,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.3574,
"step": 3
},
{
"epoch": 0.026936026936026935,
"grad_norm": 8.726573115555482,
"learning_rate": 1.3333333333333334e-06,
"loss": 0.3702,
"step": 4
},
{
"epoch": 0.03367003367003367,
"grad_norm": 7.3549146274711585,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.309,
"step": 5
},
{
"epoch": 0.04040404040404041,
"grad_norm": 8.093966841100372,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.3129,
"step": 6
},
{
"epoch": 0.04713804713804714,
"grad_norm": 4.842714262161294,
"learning_rate": 2.3333333333333336e-06,
"loss": 0.2599,
"step": 7
},
{
"epoch": 0.05387205387205387,
"grad_norm": 3.6588226035329976,
"learning_rate": 2.666666666666667e-06,
"loss": 0.2195,
"step": 8
},
{
"epoch": 0.06060606060606061,
"grad_norm": 4.446495106085817,
"learning_rate": 3e-06,
"loss": 0.2421,
"step": 9
},
{
"epoch": 0.06734006734006734,
"grad_norm": 2.8148929686162267,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.1974,
"step": 10
},
{
"epoch": 0.07407407407407407,
"grad_norm": 3.100744939706759,
"learning_rate": 3.6666666666666666e-06,
"loss": 0.226,
"step": 11
},
{
"epoch": 0.08080808080808081,
"grad_norm": 5.783874426714631,
"learning_rate": 4.000000000000001e-06,
"loss": 0.244,
"step": 12
},
{
"epoch": 0.08754208754208755,
"grad_norm": 3.931311548892614,
"learning_rate": 4.333333333333334e-06,
"loss": 0.2085,
"step": 13
},
{
"epoch": 0.09427609427609428,
"grad_norm": 3.160745880374999,
"learning_rate": 4.666666666666667e-06,
"loss": 0.193,
"step": 14
},
{
"epoch": 0.10101010101010101,
"grad_norm": 2.593859671610231,
"learning_rate": 5e-06,
"loss": 0.1918,
"step": 15
},
{
"epoch": 0.10774410774410774,
"grad_norm": 2.3411274601291923,
"learning_rate": 5.333333333333334e-06,
"loss": 0.1658,
"step": 16
},
{
"epoch": 0.11447811447811448,
"grad_norm": 2.4713577315394915,
"learning_rate": 5.666666666666667e-06,
"loss": 0.1631,
"step": 17
},
{
"epoch": 0.12121212121212122,
"grad_norm": 1.7868674672717264,
"learning_rate": 6e-06,
"loss": 0.1447,
"step": 18
},
{
"epoch": 0.12794612794612795,
"grad_norm": 2.691562015290166,
"learning_rate": 6.333333333333333e-06,
"loss": 0.1642,
"step": 19
},
{
"epoch": 0.13468013468013468,
"grad_norm": 2.656849943102207,
"learning_rate": 6.666666666666667e-06,
"loss": 0.1679,
"step": 20
},
{
"epoch": 0.1414141414141414,
"grad_norm": 1.8157646932459017,
"learning_rate": 7e-06,
"loss": 0.148,
"step": 21
},
{
"epoch": 0.14814814814814814,
"grad_norm": 1.8536112319855758,
"learning_rate": 7.333333333333333e-06,
"loss": 0.171,
"step": 22
},
{
"epoch": 0.15488215488215487,
"grad_norm": 1.7970217082935407,
"learning_rate": 7.666666666666667e-06,
"loss": 0.1543,
"step": 23
},
{
"epoch": 0.16161616161616163,
"grad_norm": 1.6194461511446632,
"learning_rate": 8.000000000000001e-06,
"loss": 0.1331,
"step": 24
},
{
"epoch": 0.16835016835016836,
"grad_norm": 1.6671876155213552,
"learning_rate": 8.333333333333334e-06,
"loss": 0.1997,
"step": 25
},
{
"epoch": 0.1750841750841751,
"grad_norm": 2.8083673280494987,
"learning_rate": 8.666666666666668e-06,
"loss": 0.149,
"step": 26
},
{
"epoch": 0.18181818181818182,
"grad_norm": 2.3945625699544983,
"learning_rate": 9e-06,
"loss": 0.2017,
"step": 27
},
{
"epoch": 0.18855218855218855,
"grad_norm": 2.547483046750297,
"learning_rate": 9.333333333333334e-06,
"loss": 0.1948,
"step": 28
},
{
"epoch": 0.19528619528619529,
"grad_norm": 2.1650784865978103,
"learning_rate": 9.666666666666667e-06,
"loss": 0.1371,
"step": 29
},
{
"epoch": 0.20202020202020202,
"grad_norm": 3.341335706430969,
"learning_rate": 1e-05,
"loss": 0.1475,
"step": 30
},
{
"epoch": 0.20875420875420875,
"grad_norm": 3.4978479659224306,
"learning_rate": 9.999651284354774e-06,
"loss": 0.1742,
"step": 31
},
{
"epoch": 0.21548821548821548,
"grad_norm": 2.8495339317165573,
"learning_rate": 9.998605186060138e-06,
"loss": 0.1318,
"step": 32
},
{
"epoch": 0.2222222222222222,
"grad_norm": 2.4440237950440045,
"learning_rate": 9.996861851032426e-06,
"loss": 0.1636,
"step": 33
},
{
"epoch": 0.22895622895622897,
"grad_norm": 2.606759351085706,
"learning_rate": 9.99442152244292e-06,
"loss": 0.1655,
"step": 34
},
{
"epoch": 0.2356902356902357,
"grad_norm": 2.4143459809986276,
"learning_rate": 9.991284540683922e-06,
"loss": 0.1492,
"step": 35
},
{
"epoch": 0.24242424242424243,
"grad_norm": 2.2585659926066226,
"learning_rate": 9.98745134332128e-06,
"loss": 0.1426,
"step": 36
},
{
"epoch": 0.24915824915824916,
"grad_norm": 2.2201124689491434,
"learning_rate": 9.98292246503335e-06,
"loss": 0.1692,
"step": 37
},
{
"epoch": 0.2558922558922559,
"grad_norm": 1.8870018630441274,
"learning_rate": 9.97769853753642e-06,
"loss": 0.1139,
"step": 38
},
{
"epoch": 0.26262626262626265,
"grad_norm": 1.571794239015075,
"learning_rate": 9.971780289496585e-06,
"loss": 0.1301,
"step": 39
},
{
"epoch": 0.26936026936026936,
"grad_norm": 1.8448276378849247,
"learning_rate": 9.965168546428122e-06,
"loss": 0.1707,
"step": 40
},
{
"epoch": 0.2760942760942761,
"grad_norm": 1.739556757210725,
"learning_rate": 9.95786423057833e-06,
"loss": 0.1229,
"step": 41
},
{
"epoch": 0.2828282828282828,
"grad_norm": 2.3063563017251667,
"learning_rate": 9.949868360798893e-06,
"loss": 0.1528,
"step": 42
},
{
"epoch": 0.2895622895622896,
"grad_norm": 1.6568054539708137,
"learning_rate": 9.941182052403768e-06,
"loss": 0.1402,
"step": 43
},
{
"epoch": 0.2962962962962963,
"grad_norm": 2.4925535277590254,
"learning_rate": 9.931806517013612e-06,
"loss": 0.1805,
"step": 44
},
{
"epoch": 0.30303030303030304,
"grad_norm": 1.4549013854977153,
"learning_rate": 9.921743062386773e-06,
"loss": 0.1167,
"step": 45
},
{
"epoch": 0.30976430976430974,
"grad_norm": 1.3094244266752977,
"learning_rate": 9.910993092236878e-06,
"loss": 0.125,
"step": 46
},
{
"epoch": 0.3164983164983165,
"grad_norm": 1.7107110369252345,
"learning_rate": 9.899558106037039e-06,
"loss": 0.1509,
"step": 47
},
{
"epoch": 0.32323232323232326,
"grad_norm": 1.2447980159357512,
"learning_rate": 9.887439698810694e-06,
"loss": 0.1101,
"step": 48
},
{
"epoch": 0.32996632996632996,
"grad_norm": 1.4004468828469865,
"learning_rate": 9.874639560909118e-06,
"loss": 0.1526,
"step": 49
},
{
"epoch": 0.3367003367003367,
"grad_norm": 1.7542103007123673,
"learning_rate": 9.861159477775653e-06,
"loss": 0.1446,
"step": 50
},
{
"epoch": 0.3434343434343434,
"grad_norm": 1.2222510349896305,
"learning_rate": 9.847001329696653e-06,
"loss": 0.118,
"step": 51
},
{
"epoch": 0.3501683501683502,
"grad_norm": 1.8427984079271709,
"learning_rate": 9.832167091539215e-06,
"loss": 0.142,
"step": 52
},
{
"epoch": 0.3569023569023569,
"grad_norm": 2.031779034578671,
"learning_rate": 9.816658832475709e-06,
"loss": 0.1314,
"step": 53
},
{
"epoch": 0.36363636363636365,
"grad_norm": 1.6115001478932334,
"learning_rate": 9.800478715695165e-06,
"loss": 0.1389,
"step": 54
},
{
"epoch": 0.37037037037037035,
"grad_norm": 1.3977600344091157,
"learning_rate": 9.783628998101525e-06,
"loss": 0.1223,
"step": 55
},
{
"epoch": 0.3771043771043771,
"grad_norm": 1.2541929470324433,
"learning_rate": 9.766112029998847e-06,
"loss": 0.1188,
"step": 56
},
{
"epoch": 0.3838383838383838,
"grad_norm": 1.1839769901258788,
"learning_rate": 9.747930254763467e-06,
"loss": 0.1445,
"step": 57
},
{
"epoch": 0.39057239057239057,
"grad_norm": 2.1758765383002476,
"learning_rate": 9.729086208503174e-06,
"loss": 0.1361,
"step": 58
},
{
"epoch": 0.39730639730639733,
"grad_norm": 1.308302111463449,
"learning_rate": 9.70958251970347e-06,
"loss": 0.1206,
"step": 59
},
{
"epoch": 0.40404040404040403,
"grad_norm": 1.2548920096786198,
"learning_rate": 9.689421908860928e-06,
"loss": 0.1547,
"step": 60
},
{
"epoch": 0.4107744107744108,
"grad_norm": 1.8106730974081224,
"learning_rate": 9.668607188103708e-06,
"loss": 0.1739,
"step": 61
},
{
"epoch": 0.4175084175084175,
"grad_norm": 1.7503742895673335,
"learning_rate": 9.64714126079933e-06,
"loss": 0.126,
"step": 62
},
{
"epoch": 0.42424242424242425,
"grad_norm": 1.9002464919585416,
"learning_rate": 9.625027121149665e-06,
"loss": 0.1209,
"step": 63
},
{
"epoch": 0.43097643097643096,
"grad_norm": 1.3005333196675646,
"learning_rate": 9.602267853773301e-06,
"loss": 0.1198,
"step": 64
},
{
"epoch": 0.4377104377104377,
"grad_norm": 1.534465904527491,
"learning_rate": 9.578866633275289e-06,
"loss": 0.1118,
"step": 65
},
{
"epoch": 0.4444444444444444,
"grad_norm": 1.2985721012757516,
"learning_rate": 9.554826723804304e-06,
"loss": 0.108,
"step": 66
},
{
"epoch": 0.4511784511784512,
"grad_norm": 1.7026939587233383,
"learning_rate": 9.530151478597366e-06,
"loss": 0.1229,
"step": 67
},
{
"epoch": 0.45791245791245794,
"grad_norm": 1.1948241076690054,
"learning_rate": 9.504844339512096e-06,
"loss": 0.1221,
"step": 68
},
{
"epoch": 0.46464646464646464,
"grad_norm": 1.4429708012011027,
"learning_rate": 9.478908836546629e-06,
"loss": 0.1154,
"step": 69
},
{
"epoch": 0.4713804713804714,
"grad_norm": 1.3983865141656082,
"learning_rate": 9.452348587347224e-06,
"loss": 0.1024,
"step": 70
},
{
"epoch": 0.4781144781144781,
"grad_norm": 1.316559372373498,
"learning_rate": 9.425167296703655e-06,
"loss": 0.1299,
"step": 71
},
{
"epoch": 0.48484848484848486,
"grad_norm": 0.8721397993491847,
"learning_rate": 9.397368756032445e-06,
"loss": 0.0838,
"step": 72
},
{
"epoch": 0.49158249158249157,
"grad_norm": 2.3999305030312006,
"learning_rate": 9.368956842848014e-06,
"loss": 0.1407,
"step": 73
},
{
"epoch": 0.4983164983164983,
"grad_norm": 1.2702381897041746,
"learning_rate": 9.339935520221816e-06,
"loss": 0.1185,
"step": 74
},
{
"epoch": 0.5050505050505051,
"grad_norm": 1.2224610902674924,
"learning_rate": 9.310308836229548e-06,
"loss": 0.1044,
"step": 75
},
{
"epoch": 0.5117845117845118,
"grad_norm": 1.0145648591712295,
"learning_rate": 9.280080923386501e-06,
"loss": 0.0793,
"step": 76
},
{
"epoch": 0.5185185185185185,
"grad_norm": 1.8590770931022098,
"learning_rate": 9.249255998071127e-06,
"loss": 0.1218,
"step": 77
},
{
"epoch": 0.5252525252525253,
"grad_norm": 1.1046753339013544,
"learning_rate": 9.217838359936914e-06,
"loss": 0.0982,
"step": 78
},
{
"epoch": 0.531986531986532,
"grad_norm": 1.5353842647876887,
"learning_rate": 9.185832391312644e-06,
"loss": 0.0976,
"step": 79
},
{
"epoch": 0.5387205387205387,
"grad_norm": 1.6322780255697593,
"learning_rate": 9.153242556591115e-06,
"loss": 0.1073,
"step": 80
},
{
"epoch": 0.5454545454545454,
"grad_norm": 1.4318238106918786,
"learning_rate": 9.120073401606427e-06,
"loss": 0.1535,
"step": 81
},
{
"epoch": 0.5521885521885522,
"grad_norm": 1.2692358298751287,
"learning_rate": 9.08632955299989e-06,
"loss": 0.1162,
"step": 82
},
{
"epoch": 0.5589225589225589,
"grad_norm": 0.9386668388367472,
"learning_rate": 9.052015717574683e-06,
"loss": 0.0819,
"step": 83
},
{
"epoch": 0.5656565656565656,
"grad_norm": 1.578041471845064,
"learning_rate": 9.017136681639307e-06,
"loss": 0.1438,
"step": 84
},
{
"epoch": 0.5723905723905723,
"grad_norm": 1.3747097561675499,
"learning_rate": 8.981697310339972e-06,
"loss": 0.1129,
"step": 85
},
{
"epoch": 0.5791245791245792,
"grad_norm": 1.2857978501576468,
"learning_rate": 8.94570254698197e-06,
"loss": 0.1153,
"step": 86
},
{
"epoch": 0.5858585858585859,
"grad_norm": 1.1071476461695378,
"learning_rate": 8.90915741234015e-06,
"loss": 0.1018,
"step": 87
},
{
"epoch": 0.5925925925925926,
"grad_norm": 1.1979772322060398,
"learning_rate": 8.872067003958597e-06,
"loss": 0.0939,
"step": 88
},
{
"epoch": 0.5993265993265994,
"grad_norm": 1.4043079567849108,
"learning_rate": 8.834436495439588e-06,
"loss": 0.1022,
"step": 89
},
{
"epoch": 0.6060606060606061,
"grad_norm": 1.182762910272131,
"learning_rate": 8.796271135721944e-06,
"loss": 0.1099,
"step": 90
},
{
"epoch": 0.6127946127946128,
"grad_norm": 0.9919737781684954,
"learning_rate": 8.757576248348883e-06,
"loss": 0.101,
"step": 91
},
{
"epoch": 0.6195286195286195,
"grad_norm": 1.0613946248613535,
"learning_rate": 8.71835723072545e-06,
"loss": 0.1074,
"step": 92
},
{
"epoch": 0.6262626262626263,
"grad_norm": 1.1119387505650915,
"learning_rate": 8.67861955336566e-06,
"loss": 0.0999,
"step": 93
},
{
"epoch": 0.632996632996633,
"grad_norm": 1.0663092046509486,
"learning_rate": 8.638368759129433e-06,
"loss": 0.1044,
"step": 94
},
{
"epoch": 0.6397306397306397,
"grad_norm": 1.5025873764978885,
"learning_rate": 8.597610462449441e-06,
"loss": 0.092,
"step": 95
},
{
"epoch": 0.6464646464646465,
"grad_norm": 1.1482831508449691,
"learning_rate": 8.556350348547978e-06,
"loss": 0.099,
"step": 96
},
{
"epoch": 0.6531986531986532,
"grad_norm": 1.1531128286621257,
"learning_rate": 8.514594172643934e-06,
"loss": 0.1122,
"step": 97
},
{
"epoch": 0.6599326599326599,
"grad_norm": 0.92890945229994,
"learning_rate": 8.472347759150044e-06,
"loss": 0.0934,
"step": 98
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.8710829747699658,
"learning_rate": 8.429617000860441e-06,
"loss": 0.0963,
"step": 99
},
{
"epoch": 0.6734006734006734,
"grad_norm": 1.5913618273955719,
"learning_rate": 8.386407858128707e-06,
"loss": 0.1017,
"step": 100
},
{
"epoch": 0.6801346801346801,
"grad_norm": 2.0477106016376423,
"learning_rate": 8.342726358036473e-06,
"loss": 0.1643,
"step": 101
},
{
"epoch": 0.6868686868686869,
"grad_norm": 1.310766070822457,
"learning_rate": 8.298578593552737e-06,
"loss": 0.1129,
"step": 102
},
{
"epoch": 0.6936026936026936,
"grad_norm": 1.0689180927391349,
"learning_rate": 8.253970722683968e-06,
"loss": 0.0805,
"step": 103
},
{
"epoch": 0.7003367003367004,
"grad_norm": 1.021500128946588,
"learning_rate": 8.208908967615159e-06,
"loss": 0.0854,
"step": 104
},
{
"epoch": 0.7070707070707071,
"grad_norm": 1.6146318951430105,
"learning_rate": 8.163399613841903e-06,
"loss": 0.1026,
"step": 105
},
{
"epoch": 0.7138047138047138,
"grad_norm": 1.0430932372821804,
"learning_rate": 8.117449009293668e-06,
"loss": 0.0911,
"step": 106
},
{
"epoch": 0.7205387205387206,
"grad_norm": 0.9627471656747238,
"learning_rate": 8.071063563448341e-06,
"loss": 0.0864,
"step": 107
},
{
"epoch": 0.7272727272727273,
"grad_norm": 0.978325652238678,
"learning_rate": 8.024249746438189e-06,
"loss": 0.0767,
"step": 108
},
{
"epoch": 0.734006734006734,
"grad_norm": 1.5716697095342131,
"learning_rate": 7.977014088147375e-06,
"loss": 0.117,
"step": 109
},
{
"epoch": 0.7407407407407407,
"grad_norm": 0.9226258206362579,
"learning_rate": 7.929363177301124e-06,
"loss": 0.0898,
"step": 110
},
{
"epoch": 0.7474747474747475,
"grad_norm": 0.979018465840073,
"learning_rate": 7.881303660546684e-06,
"loss": 0.096,
"step": 111
},
{
"epoch": 0.7542087542087542,
"grad_norm": 1.405839836354203,
"learning_rate": 7.832842241526212e-06,
"loss": 0.0969,
"step": 112
},
{
"epoch": 0.7609427609427609,
"grad_norm": 1.7228913938880963,
"learning_rate": 7.78398567994171e-06,
"loss": 0.1229,
"step": 113
},
{
"epoch": 0.7676767676767676,
"grad_norm": 1.7211262216477865,
"learning_rate": 7.734740790612137e-06,
"loss": 0.1156,
"step": 114
},
{
"epoch": 0.7744107744107744,
"grad_norm": 1.1471367685176275,
"learning_rate": 7.685114442522831e-06,
"loss": 0.0953,
"step": 115
},
{
"epoch": 0.7811447811447811,
"grad_norm": 1.0484067219863682,
"learning_rate": 7.635113557867395e-06,
"loss": 0.0957,
"step": 116
},
{
"epoch": 0.7878787878787878,
"grad_norm": 1.3643749264979563,
"learning_rate": 7.584745111082128e-06,
"loss": 0.0985,
"step": 117
},
{
"epoch": 0.7946127946127947,
"grad_norm": 1.1409003431687448,
"learning_rate": 7.5340161278732e-06,
"loss": 0.0948,
"step": 118
},
{
"epoch": 0.8013468013468014,
"grad_norm": 1.1845755031970973,
"learning_rate": 7.482933684236654e-06,
"loss": 0.1278,
"step": 119
},
{
"epoch": 0.8080808080808081,
"grad_norm": 1.2048614884456728,
"learning_rate": 7.431504905471407e-06,
"loss": 0.0866,
"step": 120
},
{
"epoch": 0.8148148148148148,
"grad_norm": 1.1024340198116762,
"learning_rate": 7.379736965185369e-06,
"loss": 0.0886,
"step": 121
},
{
"epoch": 0.8215488215488216,
"grad_norm": 1.1875877764472678,
"learning_rate": 7.327637084294818e-06,
"loss": 0.0876,
"step": 122
},
{
"epoch": 0.8282828282828283,
"grad_norm": 0.8892335379224572,
"learning_rate": 7.2752125300171835e-06,
"loss": 0.0829,
"step": 123
},
{
"epoch": 0.835016835016835,
"grad_norm": 1.1503853071251364,
"learning_rate": 7.22247061485738e-06,
"loss": 0.0945,
"step": 124
},
{
"epoch": 0.8417508417508418,
"grad_norm": 0.9116461855370438,
"learning_rate": 7.169418695587791e-06,
"loss": 0.0642,
"step": 125
},
{
"epoch": 0.8484848484848485,
"grad_norm": 0.9169291992757941,
"learning_rate": 7.1160641722221255e-06,
"loss": 0.0762,
"step": 126
},
{
"epoch": 0.8552188552188552,
"grad_norm": 0.8458786678974077,
"learning_rate": 7.062414486983197e-06,
"loss": 0.0782,
"step": 127
},
{
"epoch": 0.8619528619528619,
"grad_norm": 2.0440916025334688,
"learning_rate": 7.008477123264849e-06,
"loss": 0.1654,
"step": 128
},
{
"epoch": 0.8686868686868687,
"grad_norm": 1.2692910029856195,
"learning_rate": 6.954259604588114e-06,
"loss": 0.0993,
"step": 129
},
{
"epoch": 0.8754208754208754,
"grad_norm": 1.0371836145890874,
"learning_rate": 6.8997694935518e-06,
"loss": 0.0792,
"step": 130
},
{
"epoch": 0.8821548821548821,
"grad_norm": 0.8129848414149057,
"learning_rate": 6.845014390777595e-06,
"loss": 0.0631,
"step": 131
},
{
"epoch": 0.8888888888888888,
"grad_norm": 1.0818279823466028,
"learning_rate": 6.7900019338499005e-06,
"loss": 0.084,
"step": 132
},
{
"epoch": 0.8956228956228957,
"grad_norm": 1.2323415448858157,
"learning_rate": 6.734739796250477e-06,
"loss": 0.1037,
"step": 133
},
{
"epoch": 0.9023569023569024,
"grad_norm": 0.9072503259074285,
"learning_rate": 6.6792356862881144e-06,
"loss": 0.1022,
"step": 134
},
{
"epoch": 0.9090909090909091,
"grad_norm": 1.0444248401571283,
"learning_rate": 6.6234973460234184e-06,
"loss": 0.0952,
"step": 135
},
{
"epoch": 0.9158249158249159,
"grad_norm": 0.7905929983463482,
"learning_rate": 6.567532550188908e-06,
"loss": 0.0717,
"step": 136
},
{
"epoch": 0.9225589225589226,
"grad_norm": 0.7434134587240314,
"learning_rate": 6.511349105104534e-06,
"loss": 0.0944,
"step": 137
},
{
"epoch": 0.9292929292929293,
"grad_norm": 0.768332895445934,
"learning_rate": 6.454954847588824e-06,
"loss": 0.0678,
"step": 138
},
{
"epoch": 0.936026936026936,
"grad_norm": 1.0841144826502787,
"learning_rate": 6.398357643865731e-06,
"loss": 0.0845,
"step": 139
},
{
"epoch": 0.9427609427609428,
"grad_norm": 1.0657965630616948,
"learning_rate": 6.341565388467425e-06,
"loss": 0.0845,
"step": 140
},
{
"epoch": 0.9494949494949495,
"grad_norm": 0.7896126816989001,
"learning_rate": 6.284586003133096e-06,
"loss": 0.0922,
"step": 141
},
{
"epoch": 0.9562289562289562,
"grad_norm": 1.4442123017819188,
"learning_rate": 6.227427435703997e-06,
"loss": 0.1182,
"step": 142
},
{
"epoch": 0.9629629629629629,
"grad_norm": 0.9662733291916084,
"learning_rate": 6.170097659014812e-06,
"loss": 0.0915,
"step": 143
},
{
"epoch": 0.9696969696969697,
"grad_norm": 0.9920505746388618,
"learning_rate": 6.112604669781572e-06,
"loss": 0.093,
"step": 144
},
{
"epoch": 0.9764309764309764,
"grad_norm": 1.09207653298068,
"learning_rate": 6.054956487486212e-06,
"loss": 0.0919,
"step": 145
},
{
"epoch": 0.9831649831649831,
"grad_norm": 0.7024022249718929,
"learning_rate": 5.997161153257963e-06,
"loss": 0.0786,
"step": 146
},
{
"epoch": 0.98989898989899,
"grad_norm": 0.8553768429510301,
"learning_rate": 5.939226728751733e-06,
"loss": 0.091,
"step": 147
},
{
"epoch": 0.9966329966329966,
"grad_norm": 0.7368165994452094,
"learning_rate": 5.88116129502361e-06,
"loss": 0.0841,
"step": 148
},
{
"epoch": 0.9966329966329966,
"eval_loss": 0.0851789191365242,
"eval_runtime": 74.3023,
"eval_samples_per_second": 26.89,
"eval_steps_per_second": 0.848,
"step": 148
},
{
"epoch": 1.0033670033670035,
"grad_norm": 1.3624837722625542,
"learning_rate": 5.82297295140367e-06,
"loss": 0.0942,
"step": 149
},
{
"epoch": 1.0101010101010102,
"grad_norm": 0.8091703688887699,
"learning_rate": 5.764669814366231e-06,
"loss": 0.0673,
"step": 150
},
{
"epoch": 1.0168350168350169,
"grad_norm": 1.1689964332508505,
"learning_rate": 5.70626001639771e-06,
"loss": 0.0904,
"step": 151
},
{
"epoch": 1.0235690235690236,
"grad_norm": 0.9720282544537069,
"learning_rate": 5.647751704862263e-06,
"loss": 0.0764,
"step": 152
},
{
"epoch": 1.0303030303030303,
"grad_norm": 0.7814529677090204,
"learning_rate": 5.589153040865333e-06,
"loss": 0.0766,
"step": 153
},
{
"epoch": 1.037037037037037,
"grad_norm": 0.6208562049624894,
"learning_rate": 5.530472198115291e-06,
"loss": 0.0546,
"step": 154
},
{
"epoch": 1.0437710437710437,
"grad_norm": 1.0609611170144315,
"learning_rate": 5.471717361783312e-06,
"loss": 0.0907,
"step": 155
},
{
"epoch": 1.0505050505050506,
"grad_norm": 0.9895862720799716,
"learning_rate": 5.412896727361663e-06,
"loss": 0.0756,
"step": 156
},
{
"epoch": 1.0572390572390573,
"grad_norm": 0.8283511990419397,
"learning_rate": 5.354018499520536e-06,
"loss": 0.0733,
"step": 157
},
{
"epoch": 1.063973063973064,
"grad_norm": 0.7064918637395274,
"learning_rate": 5.2950908909636144e-06,
"loss": 0.0669,
"step": 158
},
{
"epoch": 1.0707070707070707,
"grad_norm": 0.7673084442281245,
"learning_rate": 5.2361221212825175e-06,
"loss": 0.063,
"step": 159
},
{
"epoch": 1.0774410774410774,
"grad_norm": 0.7418675838307957,
"learning_rate": 5.177120415810271e-06,
"loss": 0.0556,
"step": 160
},
{
"epoch": 1.0841750841750841,
"grad_norm": 0.9302869463074369,
"learning_rate": 5.11809400447399e-06,
"loss": 0.0937,
"step": 161
},
{
"epoch": 1.0909090909090908,
"grad_norm": 1.2470452298913122,
"learning_rate": 5.059051120646924e-06,
"loss": 0.0827,
"step": 162
},
{
"epoch": 1.0976430976430978,
"grad_norm": 0.5572520947809881,
"learning_rate": 5e-06,
"loss": 0.0537,
"step": 163
},
{
"epoch": 1.1043771043771045,
"grad_norm": 0.9002485043341543,
"learning_rate": 4.940948879353078e-06,
"loss": 0.088,
"step": 164
},
{
"epoch": 1.1111111111111112,
"grad_norm": 0.7257311476037439,
"learning_rate": 4.8819059955260105e-06,
"loss": 0.0531,
"step": 165
},
{
"epoch": 1.1178451178451179,
"grad_norm": 0.7312827459622674,
"learning_rate": 4.822879584189732e-06,
"loss": 0.0734,
"step": 166
},
{
"epoch": 1.1245791245791246,
"grad_norm": 0.6288896372866642,
"learning_rate": 4.763877878717484e-06,
"loss": 0.0538,
"step": 167
},
{
"epoch": 1.1313131313131313,
"grad_norm": 1.002048278793203,
"learning_rate": 4.704909109036387e-06,
"loss": 0.0843,
"step": 168
},
{
"epoch": 1.138047138047138,
"grad_norm": 0.7883344068499383,
"learning_rate": 4.645981500479466e-06,
"loss": 0.0656,
"step": 169
},
{
"epoch": 1.144781144781145,
"grad_norm": 0.8382010718049546,
"learning_rate": 4.587103272638339e-06,
"loss": 0.0811,
"step": 170
},
{
"epoch": 1.1515151515151516,
"grad_norm": 0.7760482065843696,
"learning_rate": 4.528282638216689e-06,
"loss": 0.0504,
"step": 171
},
{
"epoch": 1.1582491582491583,
"grad_norm": 0.7830043876217352,
"learning_rate": 4.46952780188471e-06,
"loss": 0.0715,
"step": 172
},
{
"epoch": 1.164983164983165,
"grad_norm": 1.0630254048827579,
"learning_rate": 4.410846959134667e-06,
"loss": 0.0947,
"step": 173
},
{
"epoch": 1.1717171717171717,
"grad_norm": 1.0925071520075709,
"learning_rate": 4.352248295137739e-06,
"loss": 0.0879,
"step": 174
},
{
"epoch": 1.1784511784511784,
"grad_norm": 0.6169774600371488,
"learning_rate": 4.293739983602292e-06,
"loss": 0.0611,
"step": 175
},
{
"epoch": 1.1851851851851851,
"grad_norm": 0.9939160789185896,
"learning_rate": 4.23533018563377e-06,
"loss": 0.0883,
"step": 176
},
{
"epoch": 1.1919191919191918,
"grad_norm": 1.1279865240839675,
"learning_rate": 4.17702704859633e-06,
"loss": 0.0913,
"step": 177
},
{
"epoch": 1.1986531986531987,
"grad_norm": 0.9240576378263561,
"learning_rate": 4.118838704976392e-06,
"loss": 0.0766,
"step": 178
},
{
"epoch": 1.2053872053872055,
"grad_norm": 1.2001616277657225,
"learning_rate": 4.06077327124827e-06,
"loss": 0.1161,
"step": 179
},
{
"epoch": 1.2121212121212122,
"grad_norm": 0.8147954747129799,
"learning_rate": 4.002838846742039e-06,
"loss": 0.0735,
"step": 180
},
{
"epoch": 1.2188552188552189,
"grad_norm": 0.7054084380327903,
"learning_rate": 3.94504351251379e-06,
"loss": 0.0594,
"step": 181
},
{
"epoch": 1.2255892255892256,
"grad_norm": 1.1137616381347046,
"learning_rate": 3.887395330218429e-06,
"loss": 0.1081,
"step": 182
},
{
"epoch": 1.2323232323232323,
"grad_norm": 0.671871366714128,
"learning_rate": 3.829902340985189e-06,
"loss": 0.0697,
"step": 183
},
{
"epoch": 1.239057239057239,
"grad_norm": 0.7473723466728038,
"learning_rate": 3.7725725642960047e-06,
"loss": 0.0621,
"step": 184
},
{
"epoch": 1.2457912457912457,
"grad_norm": 0.592505978208727,
"learning_rate": 3.7154139968669043e-06,
"loss": 0.058,
"step": 185
},
{
"epoch": 1.2525252525252526,
"grad_norm": 0.8357383325436504,
"learning_rate": 3.658434611532578e-06,
"loss": 0.0637,
"step": 186
},
{
"epoch": 1.2592592592592593,
"grad_norm": 0.7866708646652821,
"learning_rate": 3.6016423561342707e-06,
"loss": 0.0627,
"step": 187
},
{
"epoch": 1.265993265993266,
"grad_norm": 0.714908923134987,
"learning_rate": 3.545045152411178e-06,
"loss": 0.0632,
"step": 188
},
{
"epoch": 1.2727272727272727,
"grad_norm": 0.5485873177126037,
"learning_rate": 3.4886508948954656e-06,
"loss": 0.0599,
"step": 189
},
{
"epoch": 1.2794612794612794,
"grad_norm": 1.09587990802126,
"learning_rate": 3.4324674498110956e-06,
"loss": 0.0613,
"step": 190
},
{
"epoch": 1.2861952861952861,
"grad_norm": 0.8460192690372663,
"learning_rate": 3.3765026539765832e-06,
"loss": 0.0847,
"step": 191
},
{
"epoch": 1.2929292929292928,
"grad_norm": 0.7434506857298966,
"learning_rate": 3.3207643137118872e-06,
"loss": 0.0592,
"step": 192
},
{
"epoch": 1.2996632996632997,
"grad_norm": 0.7460450427195592,
"learning_rate": 3.2652602037495247e-06,
"loss": 0.0579,
"step": 193
},
{
"epoch": 1.3063973063973064,
"grad_norm": 1.1738509828924104,
"learning_rate": 3.2099980661501016e-06,
"loss": 0.0788,
"step": 194
},
{
"epoch": 1.3131313131313131,
"grad_norm": 0.8215622674250468,
"learning_rate": 3.154985609222405e-06,
"loss": 0.0656,
"step": 195
},
{
"epoch": 1.3198653198653199,
"grad_norm": 0.7211793878206931,
"learning_rate": 3.1002305064482006e-06,
"loss": 0.0643,
"step": 196
},
{
"epoch": 1.3265993265993266,
"grad_norm": 0.975561078244494,
"learning_rate": 3.045740395411886e-06,
"loss": 0.0671,
"step": 197
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.5847136475401894,
"learning_rate": 2.991522876735154e-06,
"loss": 0.0483,
"step": 198
},
{
"epoch": 1.34006734006734,
"grad_norm": 0.7843079810152986,
"learning_rate": 2.9375855130168046e-06,
"loss": 0.0689,
"step": 199
},
{
"epoch": 1.3468013468013469,
"grad_norm": 0.8005399461255156,
"learning_rate": 2.8839358277778758e-06,
"loss": 0.0664,
"step": 200
},
{
"epoch": 1.3535353535353536,
"grad_norm": 0.6772138624158527,
"learning_rate": 2.83058130441221e-06,
"loss": 0.0615,
"step": 201
},
{
"epoch": 1.3602693602693603,
"grad_norm": 0.7191213036719818,
"learning_rate": 2.7775293851426233e-06,
"loss": 0.0774,
"step": 202
},
{
"epoch": 1.367003367003367,
"grad_norm": 0.7605152392156752,
"learning_rate": 2.7247874699828186e-06,
"loss": 0.0928,
"step": 203
},
{
"epoch": 1.3737373737373737,
"grad_norm": 0.6195090121798972,
"learning_rate": 2.6723629157051844e-06,
"loss": 0.0596,
"step": 204
},
{
"epoch": 1.3804713804713804,
"grad_norm": 0.6643501098994239,
"learning_rate": 2.6202630348146323e-06,
"loss": 0.0523,
"step": 205
},
{
"epoch": 1.387205387205387,
"grad_norm": 0.8554198672675404,
"learning_rate": 2.5684950945285937e-06,
"loss": 0.0903,
"step": 206
},
{
"epoch": 1.393939393939394,
"grad_norm": 0.8364523777121037,
"learning_rate": 2.517066315763348e-06,
"loss": 0.0758,
"step": 207
},
{
"epoch": 1.4006734006734007,
"grad_norm": 0.7135150317434039,
"learning_rate": 2.4659838721268005e-06,
"loss": 0.0603,
"step": 208
},
{
"epoch": 1.4074074074074074,
"grad_norm": 0.6321370814051162,
"learning_rate": 2.4152548889178722e-06,
"loss": 0.093,
"step": 209
},
{
"epoch": 1.4141414141414141,
"grad_norm": 0.600961736861407,
"learning_rate": 2.364886442132606e-06,
"loss": 0.0443,
"step": 210
},
{
"epoch": 1.4208754208754208,
"grad_norm": 0.9011898089044157,
"learning_rate": 2.3148855574771706e-06,
"loss": 0.0713,
"step": 211
},
{
"epoch": 1.4276094276094276,
"grad_norm": 0.572568082620369,
"learning_rate": 2.265259209387867e-06,
"loss": 0.0567,
"step": 212
},
{
"epoch": 1.4343434343434343,
"grad_norm": 0.5810025449791688,
"learning_rate": 2.2160143200582906e-06,
"loss": 0.0561,
"step": 213
},
{
"epoch": 1.4410774410774412,
"grad_norm": 0.5830304960007473,
"learning_rate": 2.16715775847379e-06,
"loss": 0.048,
"step": 214
},
{
"epoch": 1.4478114478114479,
"grad_norm": 0.6676748383092154,
"learning_rate": 2.1186963394533165e-06,
"loss": 0.0593,
"step": 215
},
{
"epoch": 1.4545454545454546,
"grad_norm": 0.6977658852399721,
"learning_rate": 2.0706368226988772e-06,
"loss": 0.0594,
"step": 216
},
{
"epoch": 1.4612794612794613,
"grad_norm": 0.506749158721991,
"learning_rate": 2.0229859118526244e-06,
"loss": 0.0556,
"step": 217
},
{
"epoch": 1.468013468013468,
"grad_norm": 0.7142907436715384,
"learning_rate": 1.9757502535618137e-06,
"loss": 0.0524,
"step": 218
},
{
"epoch": 1.4747474747474747,
"grad_norm": 0.6949958378386697,
"learning_rate": 1.928936436551661e-06,
"loss": 0.0666,
"step": 219
},
{
"epoch": 1.4814814814814814,
"grad_norm": 0.7555665471689874,
"learning_rate": 1.8825509907063328e-06,
"loss": 0.0716,
"step": 220
},
{
"epoch": 1.4882154882154883,
"grad_norm": 0.8498899722890332,
"learning_rate": 1.8366003861580966e-06,
"loss": 0.0521,
"step": 221
},
{
"epoch": 1.494949494949495,
"grad_norm": 0.7857371564300432,
"learning_rate": 1.7910910323848435e-06,
"loss": 0.0523,
"step": 222
},
{
"epoch": 1.5016835016835017,
"grad_norm": 0.6313660611784068,
"learning_rate": 1.7460292773160315e-06,
"loss": 0.0485,
"step": 223
},
{
"epoch": 1.5084175084175084,
"grad_norm": 0.790847117411599,
"learning_rate": 1.7014214064472646e-06,
"loss": 0.0808,
"step": 224
},
{
"epoch": 1.5151515151515151,
"grad_norm": 1.1225175609085682,
"learning_rate": 1.6572736419635288e-06,
"loss": 0.0875,
"step": 225
},
{
"epoch": 1.5218855218855218,
"grad_norm": 0.8723669974198242,
"learning_rate": 1.6135921418712959e-06,
"loss": 0.0614,
"step": 226
},
{
"epoch": 1.5286195286195285,
"grad_norm": 0.5803673000386498,
"learning_rate": 1.5703829991395602e-06,
"loss": 0.0521,
"step": 227
},
{
"epoch": 1.5353535353535355,
"grad_norm": 0.5913655803974819,
"learning_rate": 1.5276522408499567e-06,
"loss": 0.0567,
"step": 228
},
{
"epoch": 1.542087542087542,
"grad_norm": 0.5913682879212468,
"learning_rate": 1.4854058273560667e-06,
"loss": 0.0575,
"step": 229
},
{
"epoch": 1.5488215488215489,
"grad_norm": 0.6925238047248119,
"learning_rate": 1.4436496514520253e-06,
"loss": 0.0582,
"step": 230
},
{
"epoch": 1.5555555555555556,
"grad_norm": 0.6395725970290016,
"learning_rate": 1.4023895375505608e-06,
"loss": 0.0605,
"step": 231
},
{
"epoch": 1.5622895622895623,
"grad_norm": 0.5799228625915099,
"learning_rate": 1.361631240870569e-06,
"loss": 0.0572,
"step": 232
},
{
"epoch": 1.569023569023569,
"grad_norm": 1.1307976257157992,
"learning_rate": 1.321380446634342e-06,
"loss": 0.0683,
"step": 233
},
{
"epoch": 1.5757575757575757,
"grad_norm": 0.53126448503362,
"learning_rate": 1.281642769274552e-06,
"loss": 0.044,
"step": 234
},
{
"epoch": 1.5824915824915826,
"grad_norm": 0.5123950135517319,
"learning_rate": 1.242423751651119e-06,
"loss": 0.0804,
"step": 235
},
{
"epoch": 1.589225589225589,
"grad_norm": 0.6099923775928446,
"learning_rate": 1.2037288642780575e-06,
"loss": 0.0408,
"step": 236
},
{
"epoch": 1.595959595959596,
"grad_norm": 0.6248835012817582,
"learning_rate": 1.165563504560413e-06,
"loss": 0.0544,
"step": 237
},
{
"epoch": 1.6026936026936027,
"grad_norm": 0.6510387000168023,
"learning_rate": 1.1279329960414047e-06,
"loss": 0.0815,
"step": 238
},
{
"epoch": 1.6094276094276094,
"grad_norm": 0.6393553382001962,
"learning_rate": 1.0908425876598512e-06,
"loss": 0.0538,
"step": 239
},
{
"epoch": 1.6161616161616161,
"grad_norm": 0.5759287160512089,
"learning_rate": 1.0542974530180327e-06,
"loss": 0.0574,
"step": 240
},
{
"epoch": 1.6228956228956228,
"grad_norm": 0.7684356456332082,
"learning_rate": 1.0183026896600284e-06,
"loss": 0.0704,
"step": 241
},
{
"epoch": 1.6296296296296298,
"grad_norm": 0.7092834204839467,
"learning_rate": 9.82863318360695e-07,
"loss": 0.0688,
"step": 242
},
{
"epoch": 1.6363636363636362,
"grad_norm": 0.8525774187670475,
"learning_rate": 9.479842824253182e-07,
"loss": 0.058,
"step": 243
},
{
"epoch": 1.6430976430976432,
"grad_norm": 0.494283557256197,
"learning_rate": 9.136704470001101e-07,
"loss": 0.0455,
"step": 244
},
{
"epoch": 1.6498316498316499,
"grad_norm": 0.5072796987904575,
"learning_rate": 8.799265983935734e-07,
"loss": 0.0573,
"step": 245
},
{
"epoch": 1.6565656565656566,
"grad_norm": 0.6544970574576625,
"learning_rate": 8.46757443408886e-07,
"loss": 0.0606,
"step": 246
},
{
"epoch": 1.6632996632996633,
"grad_norm": 0.5325045241743026,
"learning_rate": 8.141676086873574e-07,
"loss": 0.0476,
"step": 247
},
{
"epoch": 1.67003367003367,
"grad_norm": 0.571954858670152,
"learning_rate": 7.821616400630866e-07,
"loss": 0.0553,
"step": 248
},
{
"epoch": 1.676767676767677,
"grad_norm": 0.7124362240685431,
"learning_rate": 7.507440019288742e-07,
"loss": 0.0648,
"step": 249
},
{
"epoch": 1.6835016835016834,
"grad_norm": 0.8074026482316966,
"learning_rate": 7.199190766135001e-07,
"loss": 0.0677,
"step": 250
},
{
"epoch": 1.6902356902356903,
"grad_norm": 0.49821428047955796,
"learning_rate": 6.896911637704534e-07,
"loss": 0.0414,
"step": 251
},
{
"epoch": 1.696969696969697,
"grad_norm": 0.6996554021349439,
"learning_rate": 6.600644797781847e-07,
"loss": 0.0666,
"step": 252
},
{
"epoch": 1.7037037037037037,
"grad_norm": 0.5640914695952624,
"learning_rate": 6.310431571519865e-07,
"loss": 0.076,
"step": 253
},
{
"epoch": 1.7104377104377104,
"grad_norm": 0.5394553999404826,
"learning_rate": 6.026312439675553e-07,
"loss": 0.0498,
"step": 254
},
{
"epoch": 1.7171717171717171,
"grad_norm": 0.47737348342117025,
"learning_rate": 5.748327032963464e-07,
"loss": 0.0446,
"step": 255
},
{
"epoch": 1.723905723905724,
"grad_norm": 0.5870399819271779,
"learning_rate": 5.476514126527771e-07,
"loss": 0.085,
"step": 256
},
{
"epoch": 1.7306397306397305,
"grad_norm": 0.6059552958362416,
"learning_rate": 5.210911634533722e-07,
"loss": 0.0628,
"step": 257
},
{
"epoch": 1.7373737373737375,
"grad_norm": 0.4612493978879547,
"learning_rate": 4.951556604879049e-07,
"loss": 0.045,
"step": 258
},
{
"epoch": 1.7441077441077442,
"grad_norm": 0.9732831834165256,
"learning_rate": 4.698485214026349e-07,
"loss": 0.0719,
"step": 259
},
{
"epoch": 1.7508417508417509,
"grad_norm": 0.7084062235433236,
"learning_rate": 4.4517327619569784e-07,
"loss": 0.0451,
"step": 260
},
{
"epoch": 1.7575757575757576,
"grad_norm": 0.5150893466864267,
"learning_rate": 4.211333667247125e-07,
"loss": 0.054,
"step": 261
},
{
"epoch": 1.7643097643097643,
"grad_norm": 0.7197694152938333,
"learning_rate": 3.9773214622669974e-07,
"loss": 0.0528,
"step": 262
},
{
"epoch": 1.7710437710437712,
"grad_norm": 0.7937673202702573,
"learning_rate": 3.7497287885033763e-07,
"loss": 0.0551,
"step": 263
},
{
"epoch": 1.7777777777777777,
"grad_norm": 0.7807725839915187,
"learning_rate": 3.528587392006716e-07,
"loss": 0.0503,
"step": 264
},
{
"epoch": 1.7845117845117846,
"grad_norm": 0.5505115937056173,
"learning_rate": 3.313928118962906e-07,
"loss": 0.0623,
"step": 265
},
{
"epoch": 1.791245791245791,
"grad_norm": 0.6073147135377622,
"learning_rate": 3.105780911390738e-07,
"loss": 0.0674,
"step": 266
},
{
"epoch": 1.797979797979798,
"grad_norm": 0.4386889000592362,
"learning_rate": 2.904174802965293e-07,
"loss": 0.0337,
"step": 267
},
{
"epoch": 1.8047138047138047,
"grad_norm": 0.5371720109059138,
"learning_rate": 2.7091379149682683e-07,
"loss": 0.056,
"step": 268
},
{
"epoch": 1.8114478114478114,
"grad_norm": 0.4897357275916543,
"learning_rate": 2.520697452365345e-07,
"loss": 0.0442,
"step": 269
},
{
"epoch": 1.8181818181818183,
"grad_norm": 0.6143440216760881,
"learning_rate": 2.3388797000115427e-07,
"loss": 0.0644,
"step": 270
},
{
"epoch": 1.8249158249158248,
"grad_norm": 0.6184051449052006,
"learning_rate": 2.163710018984766e-07,
"loss": 0.0494,
"step": 271
},
{
"epoch": 1.8316498316498318,
"grad_norm": 0.5174263740286721,
"learning_rate": 1.9952128430483718e-07,
"loss": 0.0733,
"step": 272
},
{
"epoch": 1.8383838383838382,
"grad_norm": 0.5711547890486989,
"learning_rate": 1.8334116752429243e-07,
"loss": 0.0546,
"step": 273
},
{
"epoch": 1.8451178451178452,
"grad_norm": 0.543912361463854,
"learning_rate": 1.6783290846078714e-07,
"loss": 0.0467,
"step": 274
},
{
"epoch": 1.8518518518518519,
"grad_norm": 0.6068680622136619,
"learning_rate": 1.5299867030334815e-07,
"loss": 0.0569,
"step": 275
},
{
"epoch": 1.8585858585858586,
"grad_norm": 0.5927823999637648,
"learning_rate": 1.388405222243472e-07,
"loss": 0.0533,
"step": 276
},
{
"epoch": 1.8653198653198653,
"grad_norm": 0.47694438499884656,
"learning_rate": 1.253604390908819e-07,
"loss": 0.0564,
"step": 277
},
{
"epoch": 1.872053872053872,
"grad_norm": 0.5019664348313649,
"learning_rate": 1.1256030118930727e-07,
"loss": 0.0421,
"step": 278
},
{
"epoch": 1.878787878787879,
"grad_norm": 0.5244072558220748,
"learning_rate": 1.0044189396296144e-07,
"loss": 0.0443,
"step": 279
},
{
"epoch": 1.8855218855218854,
"grad_norm": 0.5694979771175117,
"learning_rate": 8.900690776312282e-08,
"loss": 0.0519,
"step": 280
},
{
"epoch": 1.8922558922558923,
"grad_norm": 0.5383345898694687,
"learning_rate": 7.825693761322861e-08,
"loss": 0.0645,
"step": 281
},
{
"epoch": 1.898989898989899,
"grad_norm": 0.58996301402648,
"learning_rate": 6.819348298638839e-08,
"loss": 0.0534,
"step": 282
},
{
"epoch": 1.9057239057239057,
"grad_norm": 0.4519126591505613,
"learning_rate": 5.881794759623194e-08,
"loss": 0.0397,
"step": 283
},
{
"epoch": 1.9124579124579124,
"grad_norm": 0.459397062343545,
"learning_rate": 5.013163920110864e-08,
"loss": 0.0562,
"step": 284
},
{
"epoch": 1.9191919191919191,
"grad_norm": 0.6692683465122977,
"learning_rate": 4.21357694216723e-08,
"loss": 0.0536,
"step": 285
},
{
"epoch": 1.925925925925926,
"grad_norm": 0.5515458851174179,
"learning_rate": 3.483145357187967e-08,
"loss": 0.0709,
"step": 286
},
{
"epoch": 1.9326599326599325,
"grad_norm": 0.7465197849269988,
"learning_rate": 2.8219710503416543e-08,
"loss": 0.0656,
"step": 287
},
{
"epoch": 1.9393939393939394,
"grad_norm": 0.5871951635424402,
"learning_rate": 2.230146246358256e-08,
"loss": 0.0667,
"step": 288
},
{
"epoch": 1.9461279461279462,
"grad_norm": 0.4842964057678575,
"learning_rate": 1.7077534966650767e-08,
"loss": 0.0408,
"step": 289
},
{
"epoch": 1.9528619528619529,
"grad_norm": 0.5849783564235355,
"learning_rate": 1.2548656678721404e-08,
"loss": 0.067,
"step": 290
},
{
"epoch": 1.9595959595959596,
"grad_norm": 0.6264347265804303,
"learning_rate": 8.715459316078756e-09,
"loss": 0.0492,
"step": 291
},
{
"epoch": 1.9663299663299663,
"grad_norm": 0.5290968473577989,
"learning_rate": 5.578477557081074e-09,
"loss": 0.0543,
"step": 292
},
{
"epoch": 1.9730639730639732,
"grad_norm": 0.4771694715604212,
"learning_rate": 3.1381489675746946e-09,
"loss": 0.0419,
"step": 293
},
{
"epoch": 1.9797979797979797,
"grad_norm": 0.7387620656079369,
"learning_rate": 1.3948139398628492e-09,
"loss": 0.0613,
"step": 294
},
{
"epoch": 1.9865319865319866,
"grad_norm": 0.5259388355087269,
"learning_rate": 3.487156452258722e-10,
"loss": 0.0405,
"step": 295
},
{
"epoch": 1.9932659932659933,
"grad_norm": 0.7981313513015809,
"learning_rate": 0.0,
"loss": 0.0774,
"step": 296
},
{
"epoch": 1.9932659932659933,
"eval_loss": 0.06412597745656967,
"eval_runtime": 70.2541,
"eval_samples_per_second": 28.44,
"eval_steps_per_second": 0.897,
"step": 296
},
{
"epoch": 1.9932659932659933,
"step": 296,
"total_flos": 1.5329682729109094e+17,
"train_loss": 0.098479394486325,
"train_runtime": 5160.5268,
"train_samples_per_second": 7.355,
"train_steps_per_second": 0.057
}
],
"logging_steps": 1,
"max_steps": 296,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.5329682729109094e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}