{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 2001, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0014992503748125937, "grad_norm": 0.027869803162182624, "learning_rate": 1.4925373134328358e-06, "loss": 0.0328, "step": 1 }, { "epoch": 0.0029985007496251873, "grad_norm": 0.03034633120050999, "learning_rate": 2.9850746268656716e-06, "loss": 0.0521, "step": 2 }, { "epoch": 0.004497751124437781, "grad_norm": 0.033013192126779396, "learning_rate": 4.477611940298508e-06, "loss": 0.0481, "step": 3 }, { "epoch": 0.005997001499250375, "grad_norm": 0.02964454530876554, "learning_rate": 5.970149253731343e-06, "loss": 0.0348, "step": 4 }, { "epoch": 0.0074962518740629685, "grad_norm": 0.04545013411384094, "learning_rate": 7.4626865671641785e-06, "loss": 0.0634, "step": 5 }, { "epoch": 0.008995502248875561, "grad_norm": 0.037827090237667615, "learning_rate": 8.955223880597016e-06, "loss": 0.0511, "step": 6 }, { "epoch": 0.010494752623688156, "grad_norm": 0.04024496430992649, "learning_rate": 1.0447761194029851e-05, "loss": 0.0902, "step": 7 }, { "epoch": 0.01199400299850075, "grad_norm": 0.02784723970702267, "learning_rate": 1.1940298507462686e-05, "loss": 0.0346, "step": 8 }, { "epoch": 0.013493253373313344, "grad_norm": 0.04073024185229626, "learning_rate": 1.3432835820895523e-05, "loss": 0.0551, "step": 9 }, { "epoch": 0.014992503748125937, "grad_norm": 0.03872731587077291, "learning_rate": 1.4925373134328357e-05, "loss": 0.0457, "step": 10 }, { "epoch": 0.01649175412293853, "grad_norm": 0.036983272179927376, "learning_rate": 1.6417910447761194e-05, "loss": 0.0466, "step": 11 }, { "epoch": 0.017991004497751123, "grad_norm": 0.03450782898126308, "learning_rate": 1.791044776119403e-05, "loss": 0.0402, "step": 12 }, { "epoch": 0.019490254872563718, "grad_norm": 0.0392817087888048, "learning_rate": 1.9402985074626868e-05, "loss": 0.0679, "step": 13 }, { "epoch": 0.020989505247376312, "grad_norm": 0.044590345998053095, "learning_rate": 2.0895522388059702e-05, "loss": 0.052, "step": 14 }, { "epoch": 0.022488755622188907, "grad_norm": 0.0356527349265605, "learning_rate": 2.238805970149254e-05, "loss": 0.0468, "step": 15 }, { "epoch": 0.0239880059970015, "grad_norm": 0.03207250371592397, "learning_rate": 2.3880597014925373e-05, "loss": 0.037, "step": 16 }, { "epoch": 0.025487256371814093, "grad_norm": 0.03631797856767525, "learning_rate": 2.537313432835821e-05, "loss": 0.0606, "step": 17 }, { "epoch": 0.026986506746626688, "grad_norm": 0.042515444732217036, "learning_rate": 2.6865671641791047e-05, "loss": 0.0527, "step": 18 }, { "epoch": 0.02848575712143928, "grad_norm": 0.03859829754974758, "learning_rate": 2.835820895522388e-05, "loss": 0.0577, "step": 19 }, { "epoch": 0.029985007496251874, "grad_norm": 0.04677384374732077, "learning_rate": 2.9850746268656714e-05, "loss": 0.056, "step": 20 }, { "epoch": 0.031484257871064465, "grad_norm": 0.04751009104400587, "learning_rate": 3.1343283582089554e-05, "loss": 0.0623, "step": 21 }, { "epoch": 0.03298350824587706, "grad_norm": 0.029581915526506828, "learning_rate": 3.283582089552239e-05, "loss": 0.0468, "step": 22 }, { "epoch": 0.034482758620689655, "grad_norm": 0.03579505353381419, "learning_rate": 3.432835820895522e-05, "loss": 0.0587, "step": 23 }, { "epoch": 0.035982008995502246, "grad_norm": 0.04751415291633006, "learning_rate": 3.582089552238806e-05, "loss": 0.1013, "step": 24 }, { "epoch": 0.037481259370314844, "grad_norm": 0.04449100113636761, "learning_rate": 3.73134328358209e-05, "loss": 0.0661, "step": 25 }, { "epoch": 0.038980509745127435, "grad_norm": 0.042733677260132046, "learning_rate": 3.8805970149253736e-05, "loss": 0.0885, "step": 26 }, { "epoch": 0.04047976011994003, "grad_norm": 0.02626605876758245, "learning_rate": 4.029850746268657e-05, "loss": 0.0423, "step": 27 }, { "epoch": 0.041979010494752625, "grad_norm": 0.03414998677776291, "learning_rate": 4.1791044776119404e-05, "loss": 0.0484, "step": 28 }, { "epoch": 0.043478260869565216, "grad_norm": 0.023465307923170507, "learning_rate": 4.328358208955224e-05, "loss": 0.0249, "step": 29 }, { "epoch": 0.044977511244377814, "grad_norm": 0.03238010472154882, "learning_rate": 4.477611940298508e-05, "loss": 0.0544, "step": 30 }, { "epoch": 0.046476761619190406, "grad_norm": 0.031312824540825665, "learning_rate": 4.626865671641791e-05, "loss": 0.046, "step": 31 }, { "epoch": 0.047976011994003, "grad_norm": 0.03219959967609331, "learning_rate": 4.7761194029850745e-05, "loss": 0.058, "step": 32 }, { "epoch": 0.049475262368815595, "grad_norm": 0.05253387545719208, "learning_rate": 4.9253731343283586e-05, "loss": 0.0935, "step": 33 }, { "epoch": 0.050974512743628186, "grad_norm": 0.04906400811846751, "learning_rate": 5.074626865671642e-05, "loss": 0.0592, "step": 34 }, { "epoch": 0.05247376311844078, "grad_norm": 0.028913111785186865, "learning_rate": 5.223880597014925e-05, "loss": 0.0422, "step": 35 }, { "epoch": 0.053973013493253376, "grad_norm": 0.036628261788675495, "learning_rate": 5.373134328358209e-05, "loss": 0.056, "step": 36 }, { "epoch": 0.05547226386806597, "grad_norm": 0.03723553635398014, "learning_rate": 5.5223880597014934e-05, "loss": 0.0514, "step": 37 }, { "epoch": 0.05697151424287856, "grad_norm": 0.030481090328732696, "learning_rate": 5.671641791044776e-05, "loss": 0.0505, "step": 38 }, { "epoch": 0.05847076461769116, "grad_norm": 0.06319461018053213, "learning_rate": 5.82089552238806e-05, "loss": 0.0932, "step": 39 }, { "epoch": 0.05997001499250375, "grad_norm": 0.04375744045965157, "learning_rate": 5.970149253731343e-05, "loss": 0.052, "step": 40 }, { "epoch": 0.06146926536731634, "grad_norm": 0.02716485170444979, "learning_rate": 6.119402985074628e-05, "loss": 0.0394, "step": 41 }, { "epoch": 0.06296851574212893, "grad_norm": 0.03380373227357743, "learning_rate": 6.268656716417911e-05, "loss": 0.0444, "step": 42 }, { "epoch": 0.06446776611694154, "grad_norm": 0.023682114706538047, "learning_rate": 6.417910447761194e-05, "loss": 0.0318, "step": 43 }, { "epoch": 0.06596701649175413, "grad_norm": 0.03717832489837232, "learning_rate": 6.567164179104478e-05, "loss": 0.0542, "step": 44 }, { "epoch": 0.06746626686656672, "grad_norm": 0.03849400793350328, "learning_rate": 6.716417910447762e-05, "loss": 0.066, "step": 45 }, { "epoch": 0.06896551724137931, "grad_norm": 0.039557900524499395, "learning_rate": 6.865671641791044e-05, "loss": 0.0599, "step": 46 }, { "epoch": 0.0704647676161919, "grad_norm": 0.04183512559026415, "learning_rate": 7.014925373134329e-05, "loss": 0.0559, "step": 47 }, { "epoch": 0.07196401799100449, "grad_norm": 0.023819063243281605, "learning_rate": 7.164179104477612e-05, "loss": 0.0334, "step": 48 }, { "epoch": 0.0734632683658171, "grad_norm": 0.036181513868464375, "learning_rate": 7.313432835820896e-05, "loss": 0.053, "step": 49 }, { "epoch": 0.07496251874062969, "grad_norm": 0.038961606945728214, "learning_rate": 7.46268656716418e-05, "loss": 0.0469, "step": 50 }, { "epoch": 0.07646176911544228, "grad_norm": 0.047892810949924564, "learning_rate": 7.611940298507463e-05, "loss": 0.0585, "step": 51 }, { "epoch": 0.07796101949025487, "grad_norm": 0.06267828987629875, "learning_rate": 7.761194029850747e-05, "loss": 0.0671, "step": 52 }, { "epoch": 0.07946026986506746, "grad_norm": 0.027197819167560006, "learning_rate": 7.910447761194029e-05, "loss": 0.0384, "step": 53 }, { "epoch": 0.08095952023988005, "grad_norm": 0.062135219556912424, "learning_rate": 8.059701492537314e-05, "loss": 0.0724, "step": 54 }, { "epoch": 0.08245877061469266, "grad_norm": 0.04159649809292002, "learning_rate": 8.208955223880597e-05, "loss": 0.0528, "step": 55 }, { "epoch": 0.08395802098950525, "grad_norm": 0.04806451958909231, "learning_rate": 8.358208955223881e-05, "loss": 0.0522, "step": 56 }, { "epoch": 0.08545727136431784, "grad_norm": 0.0328648844097681, "learning_rate": 8.507462686567164e-05, "loss": 0.0529, "step": 57 }, { "epoch": 0.08695652173913043, "grad_norm": 0.03499304511129161, "learning_rate": 8.656716417910447e-05, "loss": 0.0457, "step": 58 }, { "epoch": 0.08845577211394302, "grad_norm": 0.03632609832686457, "learning_rate": 8.805970149253732e-05, "loss": 0.0511, "step": 59 }, { "epoch": 0.08995502248875563, "grad_norm": 0.03286113779513323, "learning_rate": 8.955223880597016e-05, "loss": 0.0605, "step": 60 }, { "epoch": 0.09145427286356822, "grad_norm": 0.05626334745781654, "learning_rate": 9.104477611940299e-05, "loss": 0.0591, "step": 61 }, { "epoch": 0.09295352323838081, "grad_norm": 0.04452407950397946, "learning_rate": 9.253731343283582e-05, "loss": 0.0769, "step": 62 }, { "epoch": 0.0944527736131934, "grad_norm": 0.03625308210216529, "learning_rate": 9.402985074626867e-05, "loss": 0.0468, "step": 63 }, { "epoch": 0.095952023988006, "grad_norm": 0.029791030013327606, "learning_rate": 9.552238805970149e-05, "loss": 0.0584, "step": 64 }, { "epoch": 0.09745127436281859, "grad_norm": 0.03780831906836486, "learning_rate": 9.701492537313434e-05, "loss": 0.0333, "step": 65 }, { "epoch": 0.09895052473763119, "grad_norm": 0.032397150776764055, "learning_rate": 9.850746268656717e-05, "loss": 0.0447, "step": 66 }, { "epoch": 0.10044977511244378, "grad_norm": 0.03634540437911119, "learning_rate": 0.0001, "loss": 0.0602, "step": 67 }, { "epoch": 0.10194902548725637, "grad_norm": 0.025903306639537493, "learning_rate": 0.00010149253731343284, "loss": 0.0334, "step": 68 }, { "epoch": 0.10344827586206896, "grad_norm": 0.05692249372761759, "learning_rate": 0.00010298507462686569, "loss": 0.0897, "step": 69 }, { "epoch": 0.10494752623688156, "grad_norm": 0.04193568812333293, "learning_rate": 0.0001044776119402985, "loss": 0.035, "step": 70 }, { "epoch": 0.10644677661169415, "grad_norm": 0.03978190062644848, "learning_rate": 0.00010597014925373134, "loss": 0.0442, "step": 71 }, { "epoch": 0.10794602698650675, "grad_norm": 0.041177480827902055, "learning_rate": 0.00010746268656716419, "loss": 0.0695, "step": 72 }, { "epoch": 0.10944527736131934, "grad_norm": 0.044042393308047885, "learning_rate": 0.00010895522388059702, "loss": 0.061, "step": 73 }, { "epoch": 0.11094452773613193, "grad_norm": 0.027180434415964242, "learning_rate": 0.00011044776119402987, "loss": 0.0316, "step": 74 }, { "epoch": 0.11244377811094453, "grad_norm": 0.03300393578125543, "learning_rate": 0.00011194029850746269, "loss": 0.0587, "step": 75 }, { "epoch": 0.11394302848575712, "grad_norm": 0.028528364848481542, "learning_rate": 0.00011343283582089552, "loss": 0.0352, "step": 76 }, { "epoch": 0.11544227886056972, "grad_norm": 0.030828012521923, "learning_rate": 0.00011492537313432837, "loss": 0.0481, "step": 77 }, { "epoch": 0.11694152923538231, "grad_norm": 0.04083313566006794, "learning_rate": 0.0001164179104477612, "loss": 0.0604, "step": 78 }, { "epoch": 0.1184407796101949, "grad_norm": 0.025320856260068247, "learning_rate": 0.00011791044776119405, "loss": 0.0425, "step": 79 }, { "epoch": 0.1199400299850075, "grad_norm": 0.025013025974113073, "learning_rate": 0.00011940298507462686, "loss": 0.0485, "step": 80 }, { "epoch": 0.12143928035982009, "grad_norm": 0.02829433473565749, "learning_rate": 0.0001208955223880597, "loss": 0.0475, "step": 81 }, { "epoch": 0.12293853073463268, "grad_norm": 0.04936619613683576, "learning_rate": 0.00012238805970149255, "loss": 0.0766, "step": 82 }, { "epoch": 0.12443778110944528, "grad_norm": 0.06517476388380389, "learning_rate": 0.00012388059701492538, "loss": 0.0801, "step": 83 }, { "epoch": 0.12593703148425786, "grad_norm": 0.045016042881905784, "learning_rate": 0.00012537313432835822, "loss": 0.051, "step": 84 }, { "epoch": 0.12743628185907047, "grad_norm": 0.03884803395791614, "learning_rate": 0.00012686567164179105, "loss": 0.0594, "step": 85 }, { "epoch": 0.12893553223388307, "grad_norm": 0.028365426279383703, "learning_rate": 0.00012835820895522389, "loss": 0.0436, "step": 86 }, { "epoch": 0.13043478260869565, "grad_norm": 0.030909522037005376, "learning_rate": 0.00012985074626865672, "loss": 0.0348, "step": 87 }, { "epoch": 0.13193403298350825, "grad_norm": 0.0651832129849852, "learning_rate": 0.00013134328358208955, "loss": 0.0687, "step": 88 }, { "epoch": 0.13343328335832083, "grad_norm": 0.043557085072315455, "learning_rate": 0.0001328358208955224, "loss": 0.0473, "step": 89 }, { "epoch": 0.13493253373313344, "grad_norm": 0.028845028778833763, "learning_rate": 0.00013432835820895525, "loss": 0.0387, "step": 90 }, { "epoch": 0.136431784107946, "grad_norm": 0.04331404459497082, "learning_rate": 0.00013582089552238805, "loss": 0.0619, "step": 91 }, { "epoch": 0.13793103448275862, "grad_norm": 0.03133711164003795, "learning_rate": 0.0001373134328358209, "loss": 0.0315, "step": 92 }, { "epoch": 0.13943028485757122, "grad_norm": 0.04502206273984752, "learning_rate": 0.00013880597014925375, "loss": 0.0575, "step": 93 }, { "epoch": 0.1409295352323838, "grad_norm": 0.032917699035736186, "learning_rate": 0.00014029850746268658, "loss": 0.0398, "step": 94 }, { "epoch": 0.1424287856071964, "grad_norm": 0.04144739884037838, "learning_rate": 0.00014179104477611942, "loss": 0.0569, "step": 95 }, { "epoch": 0.14392803598200898, "grad_norm": 0.03128670188478762, "learning_rate": 0.00014328358208955225, "loss": 0.039, "step": 96 }, { "epoch": 0.1454272863568216, "grad_norm": 0.044978141143991145, "learning_rate": 0.00014477611940298508, "loss": 0.0607, "step": 97 }, { "epoch": 0.1469265367316342, "grad_norm": 0.049596838088832966, "learning_rate": 0.00014626865671641792, "loss": 0.053, "step": 98 }, { "epoch": 0.14842578710644677, "grad_norm": 0.03497503808856175, "learning_rate": 0.00014776119402985075, "loss": 0.0513, "step": 99 }, { "epoch": 0.14992503748125938, "grad_norm": 0.03494679451699245, "learning_rate": 0.0001492537313432836, "loss": 0.032, "step": 100 }, { "epoch": 0.15142428785607195, "grad_norm": 0.0572890539993266, "learning_rate": 0.00015074626865671642, "loss": 0.0665, "step": 101 }, { "epoch": 0.15292353823088456, "grad_norm": 0.03898694652987806, "learning_rate": 0.00015223880597014925, "loss": 0.0374, "step": 102 }, { "epoch": 0.15442278860569716, "grad_norm": 0.04199773920037622, "learning_rate": 0.00015373134328358208, "loss": 0.0461, "step": 103 }, { "epoch": 0.15592203898050974, "grad_norm": 0.057829415794355014, "learning_rate": 0.00015522388059701495, "loss": 0.0755, "step": 104 }, { "epoch": 0.15742128935532235, "grad_norm": 0.050148540491503274, "learning_rate": 0.00015671641791044778, "loss": 0.0507, "step": 105 }, { "epoch": 0.15892053973013492, "grad_norm": 0.04386621541760065, "learning_rate": 0.00015820895522388059, "loss": 0.0452, "step": 106 }, { "epoch": 0.16041979010494753, "grad_norm": 0.03554877367553605, "learning_rate": 0.00015970149253731345, "loss": 0.0428, "step": 107 }, { "epoch": 0.1619190404797601, "grad_norm": 0.11161835133361775, "learning_rate": 0.00016119402985074628, "loss": 0.0681, "step": 108 }, { "epoch": 0.1634182908545727, "grad_norm": 0.04645296691253517, "learning_rate": 0.00016268656716417911, "loss": 0.0538, "step": 109 }, { "epoch": 0.16491754122938532, "grad_norm": 0.02803028019747501, "learning_rate": 0.00016417910447761195, "loss": 0.0272, "step": 110 }, { "epoch": 0.1664167916041979, "grad_norm": 0.03415893742686059, "learning_rate": 0.00016567164179104478, "loss": 0.0254, "step": 111 }, { "epoch": 0.1679160419790105, "grad_norm": 0.05809631567240609, "learning_rate": 0.00016716417910447761, "loss": 0.0611, "step": 112 }, { "epoch": 0.16941529235382308, "grad_norm": 0.05603865215472538, "learning_rate": 0.00016865671641791045, "loss": 0.0394, "step": 113 }, { "epoch": 0.17091454272863568, "grad_norm": 0.045375774062388276, "learning_rate": 0.00017014925373134328, "loss": 0.0538, "step": 114 }, { "epoch": 0.1724137931034483, "grad_norm": 0.03960126146668028, "learning_rate": 0.00017164179104477614, "loss": 0.0423, "step": 115 }, { "epoch": 0.17391304347826086, "grad_norm": 0.037163479499863115, "learning_rate": 0.00017313432835820895, "loss": 0.0481, "step": 116 }, { "epoch": 0.17541229385307347, "grad_norm": 0.039738692107982816, "learning_rate": 0.00017462686567164178, "loss": 0.0345, "step": 117 }, { "epoch": 0.17691154422788605, "grad_norm": 0.03833182770101284, "learning_rate": 0.00017611940298507464, "loss": 0.0402, "step": 118 }, { "epoch": 0.17841079460269865, "grad_norm": 0.047548298417749796, "learning_rate": 0.00017761194029850748, "loss": 0.0542, "step": 119 }, { "epoch": 0.17991004497751126, "grad_norm": 0.033462493688485306, "learning_rate": 0.0001791044776119403, "loss": 0.0303, "step": 120 }, { "epoch": 0.18140929535232383, "grad_norm": 0.08098170529150021, "learning_rate": 0.00018059701492537314, "loss": 0.0802, "step": 121 }, { "epoch": 0.18290854572713644, "grad_norm": 0.031967576513773445, "learning_rate": 0.00018208955223880598, "loss": 0.0291, "step": 122 }, { "epoch": 0.18440779610194902, "grad_norm": 0.07620934156104307, "learning_rate": 0.0001835820895522388, "loss": 0.0737, "step": 123 }, { "epoch": 0.18590704647676162, "grad_norm": 0.05731449355361744, "learning_rate": 0.00018507462686567165, "loss": 0.0504, "step": 124 }, { "epoch": 0.1874062968515742, "grad_norm": 0.024809079949469532, "learning_rate": 0.00018656716417910448, "loss": 0.0302, "step": 125 }, { "epoch": 0.1889055472263868, "grad_norm": 0.03621287076524117, "learning_rate": 0.00018805970149253734, "loss": 0.0557, "step": 126 }, { "epoch": 0.1904047976011994, "grad_norm": 0.0495647456653575, "learning_rate": 0.00018955223880597015, "loss": 0.0547, "step": 127 }, { "epoch": 0.191904047976012, "grad_norm": 0.03485762608266449, "learning_rate": 0.00019104477611940298, "loss": 0.041, "step": 128 }, { "epoch": 0.1934032983508246, "grad_norm": 0.059266654938214335, "learning_rate": 0.00019253731343283584, "loss": 0.0909, "step": 129 }, { "epoch": 0.19490254872563717, "grad_norm": 0.08172903239953559, "learning_rate": 0.00019402985074626867, "loss": 0.1242, "step": 130 }, { "epoch": 0.19640179910044978, "grad_norm": 0.05505103331536185, "learning_rate": 0.0001955223880597015, "loss": 0.08, "step": 131 }, { "epoch": 0.19790104947526238, "grad_norm": 0.03964613621186028, "learning_rate": 0.00019701492537313434, "loss": 0.0465, "step": 132 }, { "epoch": 0.19940029985007496, "grad_norm": 0.04992604106896031, "learning_rate": 0.00019850746268656718, "loss": 0.0503, "step": 133 }, { "epoch": 0.20089955022488756, "grad_norm": 0.049289471075707976, "learning_rate": 0.0002, "loss": 0.0636, "step": 134 }, { "epoch": 0.20239880059970014, "grad_norm": 0.07369356962323476, "learning_rate": 0.00019999965730559848, "loss": 0.0732, "step": 135 }, { "epoch": 0.20389805097451275, "grad_norm": 0.047181541580997785, "learning_rate": 0.00019999862922474268, "loss": 0.0606, "step": 136 }, { "epoch": 0.20539730134932535, "grad_norm": 0.03887614440789255, "learning_rate": 0.00019999691576447898, "loss": 0.036, "step": 137 }, { "epoch": 0.20689655172413793, "grad_norm": 0.05259010448351047, "learning_rate": 0.00019999451693655123, "loss": 0.0631, "step": 138 }, { "epoch": 0.20839580209895053, "grad_norm": 0.04113206912303282, "learning_rate": 0.00019999143275740072, "loss": 0.052, "step": 139 }, { "epoch": 0.2098950524737631, "grad_norm": 0.04177105607090762, "learning_rate": 0.00019998766324816607, "loss": 0.0404, "step": 140 }, { "epoch": 0.21139430284857572, "grad_norm": 0.08334534578136313, "learning_rate": 0.0001999832084346831, "loss": 0.0897, "step": 141 }, { "epoch": 0.2128935532233883, "grad_norm": 0.04268046411364645, "learning_rate": 0.00019997806834748456, "loss": 0.0427, "step": 142 }, { "epoch": 0.2143928035982009, "grad_norm": 0.03598851664774015, "learning_rate": 0.00019997224302180006, "loss": 0.0302, "step": 143 }, { "epoch": 0.2158920539730135, "grad_norm": 0.05555632220145266, "learning_rate": 0.00019996573249755572, "loss": 0.0564, "step": 144 }, { "epoch": 0.21739130434782608, "grad_norm": 0.06965699310085238, "learning_rate": 0.00019995853681937397, "loss": 0.071, "step": 145 }, { "epoch": 0.21889055472263869, "grad_norm": 0.07939241103941817, "learning_rate": 0.00019995065603657316, "loss": 0.0907, "step": 146 }, { "epoch": 0.22038980509745126, "grad_norm": 0.050804495965929154, "learning_rate": 0.0001999420902031673, "loss": 0.048, "step": 147 }, { "epoch": 0.22188905547226387, "grad_norm": 0.0590827091873444, "learning_rate": 0.00019993283937786563, "loss": 0.0705, "step": 148 }, { "epoch": 0.22338830584707647, "grad_norm": 0.06023624941065103, "learning_rate": 0.0001999229036240723, "loss": 0.0527, "step": 149 }, { "epoch": 0.22488755622188905, "grad_norm": 0.03486164684165847, "learning_rate": 0.00019991228300988585, "loss": 0.0442, "step": 150 }, { "epoch": 0.22638680659670166, "grad_norm": 0.051399897395934625, "learning_rate": 0.00019990097760809876, "loss": 0.0975, "step": 151 }, { "epoch": 0.22788605697151423, "grad_norm": 0.03483187240018451, "learning_rate": 0.00019988898749619702, "loss": 0.0366, "step": 152 }, { "epoch": 0.22938530734632684, "grad_norm": 0.10145745147250426, "learning_rate": 0.00019987631275635948, "loss": 0.0709, "step": 153 }, { "epoch": 0.23088455772113944, "grad_norm": 0.0738726285389296, "learning_rate": 0.0001998629534754574, "loss": 0.1201, "step": 154 }, { "epoch": 0.23238380809595202, "grad_norm": 0.04853210541031046, "learning_rate": 0.00019984890974505381, "loss": 0.0628, "step": 155 }, { "epoch": 0.23388305847076463, "grad_norm": 0.05279238629760989, "learning_rate": 0.00019983418166140285, "loss": 0.0395, "step": 156 }, { "epoch": 0.2353823088455772, "grad_norm": 0.02918089043238743, "learning_rate": 0.00019981876932544917, "loss": 0.0295, "step": 157 }, { "epoch": 0.2368815592203898, "grad_norm": 0.06770346724424872, "learning_rate": 0.00019980267284282717, "loss": 0.0638, "step": 158 }, { "epoch": 0.2383808095952024, "grad_norm": 0.04833446518136561, "learning_rate": 0.00019978589232386035, "loss": 0.0519, "step": 159 }, { "epoch": 0.239880059970015, "grad_norm": 0.05186898701512186, "learning_rate": 0.00019976842788356055, "loss": 0.0478, "step": 160 }, { "epoch": 0.2413793103448276, "grad_norm": 0.03886911667428048, "learning_rate": 0.00019975027964162702, "loss": 0.0371, "step": 161 }, { "epoch": 0.24287856071964017, "grad_norm": 0.05937400672895507, "learning_rate": 0.00019973144772244582, "loss": 0.0645, "step": 162 }, { "epoch": 0.24437781109445278, "grad_norm": 0.046719542118188725, "learning_rate": 0.00019971193225508879, "loss": 0.0428, "step": 163 }, { "epoch": 0.24587706146926536, "grad_norm": 0.06610712938750751, "learning_rate": 0.0001996917333733128, "loss": 0.0482, "step": 164 }, { "epoch": 0.24737631184407796, "grad_norm": 0.050519212755809846, "learning_rate": 0.0001996708512155587, "loss": 0.0402, "step": 165 }, { "epoch": 0.24887556221889057, "grad_norm": 0.047755461648101266, "learning_rate": 0.00019964928592495045, "loss": 0.048, "step": 166 }, { "epoch": 0.25037481259370314, "grad_norm": 0.04539061348669229, "learning_rate": 0.00019962703764929413, "loss": 0.0539, "step": 167 }, { "epoch": 0.2518740629685157, "grad_norm": 0.044079218147863274, "learning_rate": 0.00019960410654107697, "loss": 0.0338, "step": 168 }, { "epoch": 0.25337331334332835, "grad_norm": 0.15940447460493573, "learning_rate": 0.0001995804927574662, "loss": 0.1907, "step": 169 }, { "epoch": 0.25487256371814093, "grad_norm": 0.14142827538567665, "learning_rate": 0.00019955619646030802, "loss": 0.1205, "step": 170 }, { "epoch": 0.2563718140929535, "grad_norm": 0.044572048202989434, "learning_rate": 0.00019953121781612657, "loss": 0.0398, "step": 171 }, { "epoch": 0.25787106446776614, "grad_norm": 0.06581882047357028, "learning_rate": 0.00019950555699612267, "loss": 0.0598, "step": 172 }, { "epoch": 0.2593703148425787, "grad_norm": 0.058819185905422434, "learning_rate": 0.00019947921417617267, "loss": 0.0356, "step": 173 }, { "epoch": 0.2608695652173913, "grad_norm": 0.07245731651154715, "learning_rate": 0.00019945218953682734, "loss": 0.0554, "step": 174 }, { "epoch": 0.2623688155922039, "grad_norm": 0.0365300697607886, "learning_rate": 0.00019942448326331055, "loss": 0.0274, "step": 175 }, { "epoch": 0.2638680659670165, "grad_norm": 0.06298974261981975, "learning_rate": 0.000199396095545518, "loss": 0.0607, "step": 176 }, { "epoch": 0.2653673163418291, "grad_norm": 0.11713127968683351, "learning_rate": 0.00019936702657801587, "loss": 0.1058, "step": 177 }, { "epoch": 0.26686656671664166, "grad_norm": 0.06199101074736703, "learning_rate": 0.00019933727656003963, "loss": 0.0491, "step": 178 }, { "epoch": 0.2683658170914543, "grad_norm": 0.06866049851783687, "learning_rate": 0.00019930684569549264, "loss": 0.0536, "step": 179 }, { "epoch": 0.2698650674662669, "grad_norm": 0.09605253534909008, "learning_rate": 0.00019927573419294456, "loss": 0.07, "step": 180 }, { "epoch": 0.27136431784107945, "grad_norm": 0.08723233109826037, "learning_rate": 0.00019924394226563018, "loss": 0.0728, "step": 181 }, { "epoch": 0.272863568215892, "grad_norm": 0.08669179137921693, "learning_rate": 0.0001992114701314478, "loss": 0.0691, "step": 182 }, { "epoch": 0.27436281859070466, "grad_norm": 0.05913118898122669, "learning_rate": 0.0001991783180129578, "loss": 0.066, "step": 183 }, { "epoch": 0.27586206896551724, "grad_norm": 0.051263995779607295, "learning_rate": 0.00019914448613738106, "loss": 0.0404, "step": 184 }, { "epoch": 0.2773613193403298, "grad_norm": 0.05477400514379348, "learning_rate": 0.0001991099747365975, "loss": 0.0427, "step": 185 }, { "epoch": 0.27886056971514245, "grad_norm": 0.06920109883069289, "learning_rate": 0.00019907478404714436, "loss": 0.0474, "step": 186 }, { "epoch": 0.280359820089955, "grad_norm": 0.03826038209148185, "learning_rate": 0.00019903891431021477, "loss": 0.0419, "step": 187 }, { "epoch": 0.2818590704647676, "grad_norm": 0.10991221849444072, "learning_rate": 0.00019900236577165576, "loss": 0.0858, "step": 188 }, { "epoch": 0.28335832083958024, "grad_norm": 0.031681176003979745, "learning_rate": 0.00019896513868196704, "loss": 0.035, "step": 189 }, { "epoch": 0.2848575712143928, "grad_norm": 0.0519531999433836, "learning_rate": 0.00019892723329629887, "loss": 0.0464, "step": 190 }, { "epoch": 0.2863568215892054, "grad_norm": 0.05571628339788405, "learning_rate": 0.0001988886498744505, "loss": 0.0518, "step": 191 }, { "epoch": 0.28785607196401797, "grad_norm": 0.06914430596733225, "learning_rate": 0.00019884938868086835, "loss": 0.0865, "step": 192 }, { "epoch": 0.2893553223388306, "grad_norm": 0.07723650037620701, "learning_rate": 0.00019880944998464434, "loss": 0.07, "step": 193 }, { "epoch": 0.2908545727136432, "grad_norm": 0.03868436237329038, "learning_rate": 0.00019876883405951377, "loss": 0.0405, "step": 194 }, { "epoch": 0.29235382308845576, "grad_norm": 0.03616727905905788, "learning_rate": 0.00019872754118385367, "loss": 0.0353, "step": 195 }, { "epoch": 0.2938530734632684, "grad_norm": 0.06407436416080277, "learning_rate": 0.00019868557164068074, "loss": 0.0518, "step": 196 }, { "epoch": 0.29535232383808097, "grad_norm": 0.0885276429432359, "learning_rate": 0.00019864292571764955, "loss": 0.071, "step": 197 }, { "epoch": 0.29685157421289354, "grad_norm": 0.06517887639083755, "learning_rate": 0.0001985996037070505, "loss": 0.0472, "step": 198 }, { "epoch": 0.2983508245877061, "grad_norm": 0.03262350195450329, "learning_rate": 0.00019855560590580778, "loss": 0.0357, "step": 199 }, { "epoch": 0.29985007496251875, "grad_norm": 0.06352918641533918, "learning_rate": 0.0001985109326154774, "loss": 0.0636, "step": 200 }, { "epoch": 0.30134932533733133, "grad_norm": 0.06275349247389327, "learning_rate": 0.0001984655841422451, "loss": 0.0746, "step": 201 }, { "epoch": 0.3028485757121439, "grad_norm": 0.04177370401732468, "learning_rate": 0.0001984195607969242, "loss": 0.0483, "step": 202 }, { "epoch": 0.30434782608695654, "grad_norm": 0.09421608937730566, "learning_rate": 0.00019837286289495361, "loss": 0.0731, "step": 203 }, { "epoch": 0.3058470764617691, "grad_norm": 0.06309582839396509, "learning_rate": 0.0001983254907563955, "loss": 0.0806, "step": 204 }, { "epoch": 0.3073463268365817, "grad_norm": 0.043040241993361705, "learning_rate": 0.00019827744470593314, "loss": 0.042, "step": 205 }, { "epoch": 0.30884557721139433, "grad_norm": 0.07592171510511146, "learning_rate": 0.0001982287250728689, "loss": 0.0672, "step": 206 }, { "epoch": 0.3103448275862069, "grad_norm": 0.04998022421825792, "learning_rate": 0.00019817933219112158, "loss": 0.0598, "step": 207 }, { "epoch": 0.3118440779610195, "grad_norm": 0.06445241515366612, "learning_rate": 0.0001981292663992245, "loss": 0.042, "step": 208 }, { "epoch": 0.31334332833583206, "grad_norm": 0.05511385292159757, "learning_rate": 0.00019807852804032305, "loss": 0.0679, "step": 209 }, { "epoch": 0.3148425787106447, "grad_norm": 0.10911657904245481, "learning_rate": 0.00019802711746217218, "loss": 0.0992, "step": 210 }, { "epoch": 0.31634182908545727, "grad_norm": 0.04607736230949591, "learning_rate": 0.00019797503501713427, "loss": 0.0386, "step": 211 }, { "epoch": 0.31784107946026985, "grad_norm": 0.055336925627864554, "learning_rate": 0.00019792228106217658, "loss": 0.0482, "step": 212 }, { "epoch": 0.3193403298350825, "grad_norm": 0.0367000018321132, "learning_rate": 0.0001978688559588688, "loss": 0.0468, "step": 213 }, { "epoch": 0.32083958020989506, "grad_norm": 0.03475056457429854, "learning_rate": 0.00019781476007338058, "loss": 0.0303, "step": 214 }, { "epoch": 0.32233883058470764, "grad_norm": 0.06222566098960902, "learning_rate": 0.0001977599937764791, "loss": 0.0643, "step": 215 }, { "epoch": 0.3238380809595202, "grad_norm": 0.08945096189461864, "learning_rate": 0.0001977045574435264, "loss": 0.1043, "step": 216 }, { "epoch": 0.32533733133433285, "grad_norm": 0.056441998030072674, "learning_rate": 0.00019764845145447689, "loss": 0.0486, "step": 217 }, { "epoch": 0.3268365817091454, "grad_norm": 0.062483693636514896, "learning_rate": 0.00019759167619387476, "loss": 0.0606, "step": 218 }, { "epoch": 0.328335832083958, "grad_norm": 0.11698538531516464, "learning_rate": 0.00019753423205085127, "loss": 0.0507, "step": 219 }, { "epoch": 0.32983508245877063, "grad_norm": 0.07112373998312807, "learning_rate": 0.0001974761194191222, "loss": 0.0504, "step": 220 }, { "epoch": 0.3313343328335832, "grad_norm": 0.04204879182316876, "learning_rate": 0.00019741733869698495, "loss": 0.0457, "step": 221 }, { "epoch": 0.3328335832083958, "grad_norm": 0.047081193320726994, "learning_rate": 0.00019735789028731604, "loss": 0.0562, "step": 222 }, { "epoch": 0.3343328335832084, "grad_norm": 0.046700620652771065, "learning_rate": 0.0001972977745975682, "loss": 0.0387, "step": 223 }, { "epoch": 0.335832083958021, "grad_norm": 0.05658161180730498, "learning_rate": 0.00019723699203976766, "loss": 0.075, "step": 224 }, { "epoch": 0.3373313343328336, "grad_norm": 0.04886519491828949, "learning_rate": 0.00019717554303051127, "loss": 0.0438, "step": 225 }, { "epoch": 0.33883058470764615, "grad_norm": 0.061882324104049265, "learning_rate": 0.00019711342799096361, "loss": 0.107, "step": 226 }, { "epoch": 0.3403298350824588, "grad_norm": 0.05128064620446334, "learning_rate": 0.00019705064734685425, "loss": 0.0499, "step": 227 }, { "epoch": 0.34182908545727136, "grad_norm": 0.037511885925608925, "learning_rate": 0.00019698720152847468, "loss": 0.0354, "step": 228 }, { "epoch": 0.34332833583208394, "grad_norm": 0.054518869510984926, "learning_rate": 0.00019692309097067546, "loss": 0.0623, "step": 229 }, { "epoch": 0.3448275862068966, "grad_norm": 0.04111726041766873, "learning_rate": 0.0001968583161128631, "loss": 0.0399, "step": 230 }, { "epoch": 0.34632683658170915, "grad_norm": 0.05016557394284265, "learning_rate": 0.0001967928773989973, "loss": 0.0399, "step": 231 }, { "epoch": 0.34782608695652173, "grad_norm": 0.054953298674171765, "learning_rate": 0.0001967267752775877, "loss": 0.0453, "step": 232 }, { "epoch": 0.3493253373313343, "grad_norm": 0.04128614741818805, "learning_rate": 0.00019666001020169073, "loss": 0.0437, "step": 233 }, { "epoch": 0.35082458770614694, "grad_norm": 0.0711824564667931, "learning_rate": 0.00019659258262890683, "loss": 0.0849, "step": 234 }, { "epoch": 0.3523238380809595, "grad_norm": 0.042020248359914844, "learning_rate": 0.00019652449302137704, "loss": 0.0386, "step": 235 }, { "epoch": 0.3538230884557721, "grad_norm": 0.046481177796938444, "learning_rate": 0.00019645574184577982, "loss": 0.0383, "step": 236 }, { "epoch": 0.3553223388305847, "grad_norm": 0.08165767843883412, "learning_rate": 0.0001963863295733281, "loss": 0.042, "step": 237 }, { "epoch": 0.3568215892053973, "grad_norm": 0.038903250983456907, "learning_rate": 0.00019631625667976583, "loss": 0.0341, "step": 238 }, { "epoch": 0.3583208395802099, "grad_norm": 0.04053657985889596, "learning_rate": 0.00019624552364536473, "loss": 0.0348, "step": 239 }, { "epoch": 0.3598200899550225, "grad_norm": 0.04790719363132042, "learning_rate": 0.00019617413095492114, "loss": 0.0386, "step": 240 }, { "epoch": 0.3613193403298351, "grad_norm": 0.053683248937234807, "learning_rate": 0.00019610207909775255, "loss": 0.0714, "step": 241 }, { "epoch": 0.36281859070464767, "grad_norm": 0.038744106761683335, "learning_rate": 0.0001960293685676943, "loss": 0.0456, "step": 242 }, { "epoch": 0.36431784107946025, "grad_norm": 0.05154181261803494, "learning_rate": 0.00019595599986309628, "loss": 0.0581, "step": 243 }, { "epoch": 0.3658170914542729, "grad_norm": 0.04824861446452559, "learning_rate": 0.0001958819734868193, "loss": 0.0706, "step": 244 }, { "epoch": 0.36731634182908546, "grad_norm": 0.03758555173738341, "learning_rate": 0.00019580728994623195, "loss": 0.0305, "step": 245 }, { "epoch": 0.36881559220389803, "grad_norm": 0.03835047499522843, "learning_rate": 0.00019573194975320673, "loss": 0.032, "step": 246 }, { "epoch": 0.37031484257871067, "grad_norm": 0.03475293565575896, "learning_rate": 0.000195655953424117, "loss": 0.0376, "step": 247 }, { "epoch": 0.37181409295352325, "grad_norm": 0.04896052744776511, "learning_rate": 0.00019557930147983302, "loss": 0.0398, "step": 248 }, { "epoch": 0.3733133433283358, "grad_norm": 0.04797369299914517, "learning_rate": 0.0001955019944457187, "loss": 0.0454, "step": 249 }, { "epoch": 0.3748125937031484, "grad_norm": 0.07781910571114095, "learning_rate": 0.0001954240328516277, "loss": 0.0698, "step": 250 }, { "epoch": 0.37631184407796103, "grad_norm": 0.04338750887929876, "learning_rate": 0.0001953454172319001, "loss": 0.0456, "step": 251 }, { "epoch": 0.3778110944527736, "grad_norm": 0.04454900811666872, "learning_rate": 0.00019526614812535864, "loss": 0.0282, "step": 252 }, { "epoch": 0.3793103448275862, "grad_norm": 0.05266395324382892, "learning_rate": 0.0001951862260753048, "loss": 0.0672, "step": 253 }, { "epoch": 0.3808095952023988, "grad_norm": 0.057494318285817976, "learning_rate": 0.00019510565162951537, "loss": 0.0731, "step": 254 }, { "epoch": 0.3823088455772114, "grad_norm": 0.04735808897837354, "learning_rate": 0.00019502442534023858, "loss": 0.0456, "step": 255 }, { "epoch": 0.383808095952024, "grad_norm": 0.1006133724153584, "learning_rate": 0.0001949425477641904, "loss": 0.0872, "step": 256 }, { "epoch": 0.3853073463268366, "grad_norm": 0.02920756237692682, "learning_rate": 0.00019486001946255046, "loss": 0.0309, "step": 257 }, { "epoch": 0.3868065967016492, "grad_norm": 0.04990145786446464, "learning_rate": 0.0001947768410009586, "loss": 0.033, "step": 258 }, { "epoch": 0.38830584707646176, "grad_norm": 0.050701365717084494, "learning_rate": 0.0001946930129495106, "loss": 0.0367, "step": 259 }, { "epoch": 0.38980509745127434, "grad_norm": 0.03698562204942488, "learning_rate": 0.00019460853588275454, "loss": 0.034, "step": 260 }, { "epoch": 0.391304347826087, "grad_norm": 0.07391856974626904, "learning_rate": 0.00019452341037968682, "loss": 0.0489, "step": 261 }, { "epoch": 0.39280359820089955, "grad_norm": 0.05717733621263349, "learning_rate": 0.00019443763702374812, "loss": 0.0448, "step": 262 }, { "epoch": 0.39430284857571213, "grad_norm": 0.044077421896152096, "learning_rate": 0.00019435121640281938, "loss": 0.0289, "step": 263 }, { "epoch": 0.39580209895052476, "grad_norm": 0.07710766944383475, "learning_rate": 0.00019426414910921787, "loss": 0.062, "step": 264 }, { "epoch": 0.39730134932533734, "grad_norm": 0.06386202461201915, "learning_rate": 0.00019417643573969303, "loss": 0.0545, "step": 265 }, { "epoch": 0.3988005997001499, "grad_norm": 0.035388492021471624, "learning_rate": 0.00019408807689542257, "loss": 0.0338, "step": 266 }, { "epoch": 0.4002998500749625, "grad_norm": 0.10927492127823649, "learning_rate": 0.00019399907318200802, "loss": 0.0893, "step": 267 }, { "epoch": 0.4017991004497751, "grad_norm": 0.06012238034598342, "learning_rate": 0.0001939094252094709, "loss": 0.0527, "step": 268 }, { "epoch": 0.4032983508245877, "grad_norm": 0.03391978832858692, "learning_rate": 0.00019381913359224842, "loss": 0.0374, "step": 269 }, { "epoch": 0.4047976011994003, "grad_norm": 0.059440210121053584, "learning_rate": 0.00019372819894918915, "loss": 0.0709, "step": 270 }, { "epoch": 0.4062968515742129, "grad_norm": 0.05188231618768116, "learning_rate": 0.000193636621903549, "loss": 0.0683, "step": 271 }, { "epoch": 0.4077961019490255, "grad_norm": 0.04719133608642286, "learning_rate": 0.00019354440308298675, "loss": 0.0477, "step": 272 }, { "epoch": 0.40929535232383807, "grad_norm": 0.07053359056135224, "learning_rate": 0.00019345154311955985, "loss": 0.0917, "step": 273 }, { "epoch": 0.4107946026986507, "grad_norm": 0.03312060738445713, "learning_rate": 0.00019335804264972018, "loss": 0.0342, "step": 274 }, { "epoch": 0.4122938530734633, "grad_norm": 0.05192332591417718, "learning_rate": 0.00019326390231430942, "loss": 0.0621, "step": 275 }, { "epoch": 0.41379310344827586, "grad_norm": 0.06103626606110473, "learning_rate": 0.0001931691227585549, "loss": 0.0544, "step": 276 }, { "epoch": 0.41529235382308843, "grad_norm": 0.035426733024569466, "learning_rate": 0.0001930737046320651, "loss": 0.0386, "step": 277 }, { "epoch": 0.41679160419790107, "grad_norm": 0.03649997902640354, "learning_rate": 0.00019297764858882514, "loss": 0.0299, "step": 278 }, { "epoch": 0.41829085457271364, "grad_norm": 0.044934938886158564, "learning_rate": 0.00019288095528719243, "loss": 0.043, "step": 279 }, { "epoch": 0.4197901049475262, "grad_norm": 0.056516397554641565, "learning_rate": 0.000192783625389892, "loss": 0.0491, "step": 280 }, { "epoch": 0.42128935532233885, "grad_norm": 0.03896533163368287, "learning_rate": 0.00019268565956401208, "loss": 0.0358, "step": 281 }, { "epoch": 0.42278860569715143, "grad_norm": 0.04117282780665747, "learning_rate": 0.0001925870584809995, "loss": 0.046, "step": 282 }, { "epoch": 0.424287856071964, "grad_norm": 0.3604260889224257, "learning_rate": 0.00019248782281665498, "loss": 0.1119, "step": 283 }, { "epoch": 0.4257871064467766, "grad_norm": 0.04720518482542838, "learning_rate": 0.0001923879532511287, "loss": 0.0427, "step": 284 }, { "epoch": 0.4272863568215892, "grad_norm": 0.04236046102866595, "learning_rate": 0.00019228745046891542, "loss": 0.026, "step": 285 }, { "epoch": 0.4287856071964018, "grad_norm": 0.06797148449245759, "learning_rate": 0.00019218631515885006, "loss": 0.0449, "step": 286 }, { "epoch": 0.4302848575712144, "grad_norm": 0.07816741282791993, "learning_rate": 0.00019208454801410266, "loss": 0.0696, "step": 287 }, { "epoch": 0.431784107946027, "grad_norm": 0.05675040921166316, "learning_rate": 0.00019198214973217378, "loss": 0.0434, "step": 288 }, { "epoch": 0.4332833583208396, "grad_norm": 0.06237548709245271, "learning_rate": 0.00019187912101488984, "loss": 0.0696, "step": 289 }, { "epoch": 0.43478260869565216, "grad_norm": 0.07002008527028293, "learning_rate": 0.00019177546256839812, "loss": 0.1012, "step": 290 }, { "epoch": 0.4362818590704648, "grad_norm": 0.04893068129720385, "learning_rate": 0.000191671175103162, "loss": 0.0409, "step": 291 }, { "epoch": 0.43778110944527737, "grad_norm": 0.04601357359317086, "learning_rate": 0.00019156625933395614, "loss": 0.0519, "step": 292 }, { "epoch": 0.43928035982008995, "grad_norm": 0.05337895694784327, "learning_rate": 0.00019146071597986138, "loss": 0.0372, "step": 293 }, { "epoch": 0.4407796101949025, "grad_norm": 0.03995143875131533, "learning_rate": 0.0001913545457642601, "loss": 0.0368, "step": 294 }, { "epoch": 0.44227886056971516, "grad_norm": 0.09850621381875037, "learning_rate": 0.00019124774941483107, "loss": 0.0761, "step": 295 }, { "epoch": 0.44377811094452774, "grad_norm": 0.038785537167516496, "learning_rate": 0.00019114032766354453, "loss": 0.041, "step": 296 }, { "epoch": 0.4452773613193403, "grad_norm": 0.066120691647694, "learning_rate": 0.00019103228124665712, "loss": 0.0747, "step": 297 }, { "epoch": 0.44677661169415295, "grad_norm": 0.11147859567607986, "learning_rate": 0.00019092361090470688, "loss": 0.0508, "step": 298 }, { "epoch": 0.4482758620689655, "grad_norm": 0.09831159493729676, "learning_rate": 0.00019081431738250814, "loss": 0.0562, "step": 299 }, { "epoch": 0.4497751124437781, "grad_norm": 0.07264006526009004, "learning_rate": 0.0001907044014291465, "loss": 0.0557, "step": 300 }, { "epoch": 0.4512743628185907, "grad_norm": 0.07996560398829375, "learning_rate": 0.0001905938637979736, "loss": 0.0482, "step": 301 }, { "epoch": 0.4527736131934033, "grad_norm": 0.0686136548292036, "learning_rate": 0.00019048270524660196, "loss": 0.0503, "step": 302 }, { "epoch": 0.4542728635682159, "grad_norm": 0.045762027803397255, "learning_rate": 0.00019037092653689987, "loss": 0.0385, "step": 303 }, { "epoch": 0.45577211394302847, "grad_norm": 0.06709124946128789, "learning_rate": 0.00019025852843498607, "loss": 0.0678, "step": 304 }, { "epoch": 0.4572713643178411, "grad_norm": 0.05610882393237923, "learning_rate": 0.00019014551171122457, "loss": 0.0482, "step": 305 }, { "epoch": 0.4587706146926537, "grad_norm": 0.06482964052411654, "learning_rate": 0.00019003187714021938, "loss": 0.0531, "step": 306 }, { "epoch": 0.46026986506746626, "grad_norm": 0.07856511913650002, "learning_rate": 0.00018991762550080906, "loss": 0.0531, "step": 307 }, { "epoch": 0.4617691154422789, "grad_norm": 0.08727827018786473, "learning_rate": 0.00018980275757606157, "loss": 0.0696, "step": 308 }, { "epoch": 0.46326836581709147, "grad_norm": 0.054829790468602986, "learning_rate": 0.00018968727415326884, "loss": 0.0522, "step": 309 }, { "epoch": 0.46476761619190404, "grad_norm": 0.0533722415191346, "learning_rate": 0.0001895711760239413, "loss": 0.0455, "step": 310 }, { "epoch": 0.4662668665667166, "grad_norm": 0.0404960196283668, "learning_rate": 0.0001894544639838025, "loss": 0.0432, "step": 311 }, { "epoch": 0.46776611694152925, "grad_norm": 0.044028946635248145, "learning_rate": 0.00018933713883278376, "loss": 0.0425, "step": 312 }, { "epoch": 0.46926536731634183, "grad_norm": 0.06293150170038057, "learning_rate": 0.00018921920137501849, "loss": 0.0508, "step": 313 }, { "epoch": 0.4707646176911544, "grad_norm": 0.06810220996600896, "learning_rate": 0.0001891006524188368, "loss": 0.0581, "step": 314 }, { "epoch": 0.47226386806596704, "grad_norm": 0.06959342447747006, "learning_rate": 0.00018898149277675997, "loss": 0.0725, "step": 315 }, { "epoch": 0.4737631184407796, "grad_norm": 0.061826777912191176, "learning_rate": 0.0001888617232654949, "loss": 0.0598, "step": 316 }, { "epoch": 0.4752623688155922, "grad_norm": 0.066670766072721, "learning_rate": 0.00018874134470592835, "loss": 0.0387, "step": 317 }, { "epoch": 0.4767616191904048, "grad_norm": 0.05284481473477404, "learning_rate": 0.00018862035792312147, "loss": 0.0353, "step": 318 }, { "epoch": 0.4782608695652174, "grad_norm": 0.03546158471045893, "learning_rate": 0.0001884987637463042, "loss": 0.0353, "step": 319 }, { "epoch": 0.47976011994003, "grad_norm": 0.03695047012775239, "learning_rate": 0.00018837656300886937, "loss": 0.033, "step": 320 }, { "epoch": 0.48125937031484256, "grad_norm": 0.09946531845424002, "learning_rate": 0.00018825375654836713, "loss": 0.068, "step": 321 }, { "epoch": 0.4827586206896552, "grad_norm": 0.0525789693495926, "learning_rate": 0.0001881303452064992, "loss": 0.0603, "step": 322 }, { "epoch": 0.48425787106446777, "grad_norm": 0.061737118720852, "learning_rate": 0.00018800632982911322, "loss": 0.0606, "step": 323 }, { "epoch": 0.48575712143928035, "grad_norm": 0.036553688772480226, "learning_rate": 0.00018788171126619653, "loss": 0.0299, "step": 324 }, { "epoch": 0.487256371814093, "grad_norm": 0.04923145685241048, "learning_rate": 0.00018775649037187093, "loss": 0.0475, "step": 325 }, { "epoch": 0.48875562218890556, "grad_norm": 0.06336340360281395, "learning_rate": 0.00018763066800438636, "loss": 0.0689, "step": 326 }, { "epoch": 0.49025487256371814, "grad_norm": 0.044393763008170536, "learning_rate": 0.00018750424502611527, "loss": 0.0413, "step": 327 }, { "epoch": 0.4917541229385307, "grad_norm": 0.04138603160844389, "learning_rate": 0.00018737722230354655, "loss": 0.0445, "step": 328 }, { "epoch": 0.49325337331334335, "grad_norm": 0.08068404550662704, "learning_rate": 0.00018724960070727972, "loss": 0.0507, "step": 329 }, { "epoch": 0.4947526236881559, "grad_norm": 0.05448885894958413, "learning_rate": 0.00018712138111201895, "loss": 0.0581, "step": 330 }, { "epoch": 0.4962518740629685, "grad_norm": 0.0541786649288355, "learning_rate": 0.00018699256439656697, "loss": 0.05, "step": 331 }, { "epoch": 0.49775112443778113, "grad_norm": 0.09333363889478805, "learning_rate": 0.00018686315144381913, "loss": 0.0634, "step": 332 }, { "epoch": 0.4992503748125937, "grad_norm": 0.06046721622165493, "learning_rate": 0.00018673314314075732, "loss": 0.0393, "step": 333 }, { "epoch": 0.5007496251874063, "grad_norm": 0.22984713729553005, "learning_rate": 0.00018660254037844388, "loss": 0.1195, "step": 334 }, { "epoch": 0.5022488755622189, "grad_norm": 0.03840858370215027, "learning_rate": 0.0001864713440520155, "loss": 0.038, "step": 335 }, { "epoch": 0.5037481259370314, "grad_norm": 0.0460739408926456, "learning_rate": 0.00018633955506067718, "loss": 0.0277, "step": 336 }, { "epoch": 0.5052473763118441, "grad_norm": 0.058586326621381794, "learning_rate": 0.00018620717430769586, "loss": 0.0418, "step": 337 }, { "epoch": 0.5067466266866567, "grad_norm": 0.0691170892823961, "learning_rate": 0.0001860742027003944, "loss": 0.0611, "step": 338 }, { "epoch": 0.5082458770614693, "grad_norm": 0.054450391696896736, "learning_rate": 0.0001859406411501453, "loss": 0.0389, "step": 339 }, { "epoch": 0.5097451274362819, "grad_norm": 0.05989403502334338, "learning_rate": 0.00018580649057236447, "loss": 0.0656, "step": 340 }, { "epoch": 0.5112443778110944, "grad_norm": 0.05812418432414411, "learning_rate": 0.00018567175188650498, "loss": 0.0516, "step": 341 }, { "epoch": 0.512743628185907, "grad_norm": 0.1024472803798587, "learning_rate": 0.00018553642601605068, "loss": 0.0489, "step": 342 }, { "epoch": 0.5142428785607196, "grad_norm": 0.06418199886991582, "learning_rate": 0.00018540051388850993, "loss": 0.0641, "step": 343 }, { "epoch": 0.5157421289355323, "grad_norm": 0.05354265643676874, "learning_rate": 0.00018526401643540922, "loss": 0.0342, "step": 344 }, { "epoch": 0.5172413793103449, "grad_norm": 0.060234343510686926, "learning_rate": 0.00018512693459228684, "loss": 0.0441, "step": 345 }, { "epoch": 0.5187406296851574, "grad_norm": 0.05502477922945836, "learning_rate": 0.00018498926929868642, "loss": 0.044, "step": 346 }, { "epoch": 0.52023988005997, "grad_norm": 0.05316566519953199, "learning_rate": 0.00018485102149815038, "loss": 0.0407, "step": 347 }, { "epoch": 0.5217391304347826, "grad_norm": 0.08251700451700679, "learning_rate": 0.00018471219213821375, "loss": 0.0475, "step": 348 }, { "epoch": 0.5232383808095952, "grad_norm": 0.044332263163009075, "learning_rate": 0.00018457278217039736, "loss": 0.0339, "step": 349 }, { "epoch": 0.5247376311844077, "grad_norm": 0.06598873060964301, "learning_rate": 0.00018443279255020152, "loss": 0.0623, "step": 350 }, { "epoch": 0.5262368815592204, "grad_norm": 0.09432658184269146, "learning_rate": 0.00018429222423709947, "loss": 0.0681, "step": 351 }, { "epoch": 0.527736131934033, "grad_norm": 0.048137123345461545, "learning_rate": 0.00018415107819453062, "loss": 0.0379, "step": 352 }, { "epoch": 0.5292353823088456, "grad_norm": 0.048726928699549, "learning_rate": 0.0001840093553898942, "loss": 0.0517, "step": 353 }, { "epoch": 0.5307346326836582, "grad_norm": 0.03886288744382217, "learning_rate": 0.00018386705679454242, "loss": 0.0309, "step": 354 }, { "epoch": 0.5322338830584707, "grad_norm": 0.04487311032788653, "learning_rate": 0.0001837241833837739, "loss": 0.0294, "step": 355 }, { "epoch": 0.5337331334332833, "grad_norm": 0.03798327068043393, "learning_rate": 0.00018358073613682706, "loss": 0.031, "step": 356 }, { "epoch": 0.5352323838080959, "grad_norm": 0.050882637106700504, "learning_rate": 0.00018343671603687317, "loss": 0.039, "step": 357 }, { "epoch": 0.5367316341829086, "grad_norm": 0.044214998902001985, "learning_rate": 0.00018329212407100994, "loss": 0.0333, "step": 358 }, { "epoch": 0.5382308845577212, "grad_norm": 0.06411346885005922, "learning_rate": 0.00018314696123025454, "loss": 0.066, "step": 359 }, { "epoch": 0.5397301349325337, "grad_norm": 0.042800963513781314, "learning_rate": 0.00018300122850953675, "loss": 0.0303, "step": 360 }, { "epoch": 0.5412293853073463, "grad_norm": 0.051572469675010446, "learning_rate": 0.00018285492690769237, "loss": 0.0459, "step": 361 }, { "epoch": 0.5427286356821589, "grad_norm": 0.06594725487157567, "learning_rate": 0.00018270805742745617, "loss": 0.0554, "step": 362 }, { "epoch": 0.5442278860569715, "grad_norm": 0.08966668405728594, "learning_rate": 0.00018256062107545518, "loss": 0.0504, "step": 363 }, { "epoch": 0.545727136431784, "grad_norm": 0.05705712395021745, "learning_rate": 0.00018241261886220154, "loss": 0.0542, "step": 364 }, { "epoch": 0.5472263868065967, "grad_norm": 0.046841858897893975, "learning_rate": 0.000182264051802086, "loss": 0.046, "step": 365 }, { "epoch": 0.5487256371814093, "grad_norm": 0.05043321212035937, "learning_rate": 0.00018211492091337042, "loss": 0.0515, "step": 366 }, { "epoch": 0.5502248875562219, "grad_norm": 0.05643206918749337, "learning_rate": 0.00018196522721818128, "loss": 0.0398, "step": 367 }, { "epoch": 0.5517241379310345, "grad_norm": 0.10718050662244094, "learning_rate": 0.00018181497174250236, "loss": 0.0234, "step": 368 }, { "epoch": 0.553223388305847, "grad_norm": 0.07740102244845123, "learning_rate": 0.00018166415551616792, "loss": 0.0523, "step": 369 }, { "epoch": 0.5547226386806596, "grad_norm": 0.06095682806386288, "learning_rate": 0.00018151277957285543, "loss": 0.0434, "step": 370 }, { "epoch": 0.5562218890554723, "grad_norm": 0.06932915829661977, "learning_rate": 0.00018136084495007872, "loss": 0.0531, "step": 371 }, { "epoch": 0.5577211394302849, "grad_norm": 0.04400477443557516, "learning_rate": 0.00018120835268918063, "loss": 0.0387, "step": 372 }, { "epoch": 0.5592203898050975, "grad_norm": 0.05568457669143035, "learning_rate": 0.00018105530383532607, "loss": 0.0359, "step": 373 }, { "epoch": 0.56071964017991, "grad_norm": 0.09375047939362961, "learning_rate": 0.00018090169943749476, "loss": 0.0613, "step": 374 }, { "epoch": 0.5622188905547226, "grad_norm": 0.09342104698058959, "learning_rate": 0.000180747540548474, "loss": 0.0508, "step": 375 }, { "epoch": 0.5637181409295352, "grad_norm": 0.06980360470280776, "learning_rate": 0.00018059282822485158, "loss": 0.0588, "step": 376 }, { "epoch": 0.5652173913043478, "grad_norm": 0.04834116988503355, "learning_rate": 0.00018043756352700846, "loss": 0.0407, "step": 377 }, { "epoch": 0.5667166416791605, "grad_norm": 0.0559433504674298, "learning_rate": 0.00018028174751911146, "loss": 0.0487, "step": 378 }, { "epoch": 0.568215892053973, "grad_norm": 0.04299769695781603, "learning_rate": 0.00018012538126910608, "loss": 0.0289, "step": 379 }, { "epoch": 0.5697151424287856, "grad_norm": 0.061484741377732285, "learning_rate": 0.00017996846584870908, "loss": 0.0545, "step": 380 }, { "epoch": 0.5712143928035982, "grad_norm": 0.04596560988589707, "learning_rate": 0.00017981100233340115, "loss": 0.0381, "step": 381 }, { "epoch": 0.5727136431784108, "grad_norm": 0.03818801182308544, "learning_rate": 0.00017965299180241963, "loss": 0.0348, "step": 382 }, { "epoch": 0.5742128935532234, "grad_norm": 0.050232900629482914, "learning_rate": 0.000179494435338751, "loss": 0.0265, "step": 383 }, { "epoch": 0.5757121439280359, "grad_norm": 0.0740772370322547, "learning_rate": 0.00017933533402912354, "loss": 0.0629, "step": 384 }, { "epoch": 0.5772113943028486, "grad_norm": 0.03602072499212665, "learning_rate": 0.00017917568896399973, "loss": 0.0303, "step": 385 }, { "epoch": 0.5787106446776612, "grad_norm": 0.04732712114588865, "learning_rate": 0.00017901550123756906, "loss": 0.0533, "step": 386 }, { "epoch": 0.5802098950524738, "grad_norm": 0.0438295195621145, "learning_rate": 0.0001788547719477402, "loss": 0.033, "step": 387 }, { "epoch": 0.5817091454272864, "grad_norm": 0.045622862481622464, "learning_rate": 0.00017869350219613375, "loss": 0.0436, "step": 388 }, { "epoch": 0.5832083958020989, "grad_norm": 0.04964994949451908, "learning_rate": 0.00017853169308807448, "loss": 0.0375, "step": 389 }, { "epoch": 0.5847076461769115, "grad_norm": 0.05078556363206241, "learning_rate": 0.000178369345732584, "loss": 0.0345, "step": 390 }, { "epoch": 0.5862068965517241, "grad_norm": 0.05074730082524994, "learning_rate": 0.00017820646124237278, "loss": 0.0364, "step": 391 }, { "epoch": 0.5877061469265368, "grad_norm": 0.04603113674846868, "learning_rate": 0.000178043040733833, "loss": 0.0302, "step": 392 }, { "epoch": 0.5892053973013494, "grad_norm": 0.058291821623076764, "learning_rate": 0.00017787908532703042, "loss": 0.0538, "step": 393 }, { "epoch": 0.5907046476761619, "grad_norm": 0.06209286221938147, "learning_rate": 0.0001777145961456971, "loss": 0.0437, "step": 394 }, { "epoch": 0.5922038980509745, "grad_norm": 0.035508544978000475, "learning_rate": 0.00017754957431722346, "loss": 0.0361, "step": 395 }, { "epoch": 0.5937031484257871, "grad_norm": 0.045021307147078936, "learning_rate": 0.00017738402097265064, "loss": 0.0314, "step": 396 }, { "epoch": 0.5952023988005997, "grad_norm": 0.045117529238231965, "learning_rate": 0.00017721793724666268, "loss": 0.0286, "step": 397 }, { "epoch": 0.5967016491754122, "grad_norm": 0.04284437809907477, "learning_rate": 0.00017705132427757895, "loss": 0.0316, "step": 398 }, { "epoch": 0.5982008995502249, "grad_norm": 0.05363788525676639, "learning_rate": 0.00017688418320734598, "loss": 0.0575, "step": 399 }, { "epoch": 0.5997001499250375, "grad_norm": 0.039147054068539446, "learning_rate": 0.00017671651518153, "loss": 0.0298, "step": 400 }, { "epoch": 0.6011994002998501, "grad_norm": 0.047103082273761164, "learning_rate": 0.00017654832134930882, "loss": 0.0332, "step": 401 }, { "epoch": 0.6026986506746627, "grad_norm": 0.03434268067696955, "learning_rate": 0.00017637960286346425, "loss": 0.0224, "step": 402 }, { "epoch": 0.6041979010494752, "grad_norm": 0.05227048698626022, "learning_rate": 0.00017621036088037378, "loss": 0.0518, "step": 403 }, { "epoch": 0.6056971514242878, "grad_norm": 0.18818259332861206, "learning_rate": 0.0001760405965600031, "loss": 0.0758, "step": 404 }, { "epoch": 0.6071964017991005, "grad_norm": 0.05885735282398511, "learning_rate": 0.00017587031106589782, "loss": 0.0361, "step": 405 }, { "epoch": 0.6086956521739131, "grad_norm": 0.08962878091622765, "learning_rate": 0.00017569950556517566, "loss": 0.0523, "step": 406 }, { "epoch": 0.6101949025487257, "grad_norm": 0.07406938968015182, "learning_rate": 0.00017552818122851838, "loss": 0.0372, "step": 407 }, { "epoch": 0.6116941529235382, "grad_norm": 0.0589541107505507, "learning_rate": 0.0001753563392301638, "loss": 0.0518, "step": 408 }, { "epoch": 0.6131934032983508, "grad_norm": 0.0493031824176551, "learning_rate": 0.00017518398074789775, "loss": 0.0404, "step": 409 }, { "epoch": 0.6146926536731634, "grad_norm": 0.060948463938890016, "learning_rate": 0.00017501110696304596, "loss": 0.0434, "step": 410 }, { "epoch": 0.616191904047976, "grad_norm": 0.05214572496276531, "learning_rate": 0.00017483771906046602, "loss": 0.0384, "step": 411 }, { "epoch": 0.6176911544227887, "grad_norm": 0.05632303506730485, "learning_rate": 0.00017466381822853915, "loss": 0.032, "step": 412 }, { "epoch": 0.6191904047976012, "grad_norm": 0.06363985932901699, "learning_rate": 0.00017448940565916222, "loss": 0.0532, "step": 413 }, { "epoch": 0.6206896551724138, "grad_norm": 0.045655772233298454, "learning_rate": 0.00017431448254773944, "loss": 0.0335, "step": 414 }, { "epoch": 0.6221889055472264, "grad_norm": 0.05496397917447876, "learning_rate": 0.00017413905009317423, "loss": 0.0432, "step": 415 }, { "epoch": 0.623688155922039, "grad_norm": 0.0542650773216564, "learning_rate": 0.000173963109497861, "loss": 0.0299, "step": 416 }, { "epoch": 0.6251874062968515, "grad_norm": 0.053416335915512446, "learning_rate": 0.00017378666196767685, "loss": 0.0384, "step": 417 }, { "epoch": 0.6266866566716641, "grad_norm": 0.053317686220209964, "learning_rate": 0.00017360970871197346, "loss": 0.0355, "step": 418 }, { "epoch": 0.6281859070464768, "grad_norm": 0.058288847449129474, "learning_rate": 0.00017343225094356855, "loss": 0.052, "step": 419 }, { "epoch": 0.6296851574212894, "grad_norm": 0.042009952684143474, "learning_rate": 0.0001732542898787379, "loss": 0.0452, "step": 420 }, { "epoch": 0.631184407796102, "grad_norm": 0.04265378963968585, "learning_rate": 0.00017307582673720663, "loss": 0.0314, "step": 421 }, { "epoch": 0.6326836581709145, "grad_norm": 0.0473560166231087, "learning_rate": 0.00017289686274214118, "loss": 0.0481, "step": 422 }, { "epoch": 0.6341829085457271, "grad_norm": 0.04316822944816478, "learning_rate": 0.00017271739912014068, "loss": 0.035, "step": 423 }, { "epoch": 0.6356821589205397, "grad_norm": 0.04821805663840275, "learning_rate": 0.00017253743710122875, "loss": 0.0353, "step": 424 }, { "epoch": 0.6371814092953523, "grad_norm": 0.05766150770706173, "learning_rate": 0.00017235697791884494, "loss": 0.0518, "step": 425 }, { "epoch": 0.638680659670165, "grad_norm": 0.09048982025586255, "learning_rate": 0.00017217602280983623, "loss": 0.0477, "step": 426 }, { "epoch": 0.6401799100449775, "grad_norm": 0.04016296087640653, "learning_rate": 0.00017199457301444868, "loss": 0.0284, "step": 427 }, { "epoch": 0.6416791604197901, "grad_norm": 0.07151231389242098, "learning_rate": 0.00017181262977631888, "loss": 0.0534, "step": 428 }, { "epoch": 0.6431784107946027, "grad_norm": 0.0521164376523254, "learning_rate": 0.00017163019434246547, "loss": 0.0413, "step": 429 }, { "epoch": 0.6446776611694153, "grad_norm": 0.06650764446753712, "learning_rate": 0.00017144726796328034, "loss": 0.0532, "step": 430 }, { "epoch": 0.6461769115442278, "grad_norm": 0.05631153402061295, "learning_rate": 0.00017126385189252053, "loss": 0.04, "step": 431 }, { "epoch": 0.6476761619190404, "grad_norm": 0.07908703283560595, "learning_rate": 0.00017107994738729926, "loss": 0.0546, "step": 432 }, { "epoch": 0.6491754122938531, "grad_norm": 0.053983476730184235, "learning_rate": 0.00017089555570807736, "loss": 0.0428, "step": 433 }, { "epoch": 0.6506746626686657, "grad_norm": 0.04223485749838613, "learning_rate": 0.00017071067811865476, "loss": 0.039, "step": 434 }, { "epoch": 0.6521739130434783, "grad_norm": 0.05289448745150468, "learning_rate": 0.00017052531588616178, "loss": 0.0369, "step": 435 }, { "epoch": 0.6536731634182908, "grad_norm": 0.04254188142407335, "learning_rate": 0.00017033947028105039, "loss": 0.0482, "step": 436 }, { "epoch": 0.6551724137931034, "grad_norm": 0.03642799958156187, "learning_rate": 0.0001701531425770856, "loss": 0.0227, "step": 437 }, { "epoch": 0.656671664167916, "grad_norm": 0.044290202162954866, "learning_rate": 0.00016996633405133655, "loss": 0.0317, "step": 438 }, { "epoch": 0.6581709145427287, "grad_norm": 0.03328713625145707, "learning_rate": 0.00016977904598416803, "loss": 0.0193, "step": 439 }, { "epoch": 0.6596701649175413, "grad_norm": 0.06825872867733464, "learning_rate": 0.00016959127965923142, "loss": 0.0524, "step": 440 }, { "epoch": 0.6611694152923538, "grad_norm": 0.05737656538911755, "learning_rate": 0.00016940303636345618, "loss": 0.0384, "step": 441 }, { "epoch": 0.6626686656671664, "grad_norm": 0.05217169033909698, "learning_rate": 0.0001692143173870407, "loss": 0.0398, "step": 442 }, { "epoch": 0.664167916041979, "grad_norm": 0.08029746938521902, "learning_rate": 0.00016902512402344373, "loss": 0.0996, "step": 443 }, { "epoch": 0.6656671664167916, "grad_norm": 0.05023109850591579, "learning_rate": 0.0001688354575693754, "loss": 0.0444, "step": 444 }, { "epoch": 0.6671664167916042, "grad_norm": 0.05007534396139674, "learning_rate": 0.00016864531932478838, "loss": 0.0495, "step": 445 }, { "epoch": 0.6686656671664168, "grad_norm": 0.06850886410730117, "learning_rate": 0.00016845471059286887, "loss": 0.0491, "step": 446 }, { "epoch": 0.6701649175412294, "grad_norm": 0.0709049222949615, "learning_rate": 0.00016826363268002782, "loss": 0.086, "step": 447 }, { "epoch": 0.671664167916042, "grad_norm": 0.07625185163942867, "learning_rate": 0.0001680720868958918, "loss": 0.0612, "step": 448 }, { "epoch": 0.6731634182908546, "grad_norm": 0.03903647995204454, "learning_rate": 0.0001678800745532942, "loss": 0.0286, "step": 449 }, { "epoch": 0.6746626686656672, "grad_norm": 0.04875227954727952, "learning_rate": 0.00016768759696826608, "loss": 0.0452, "step": 450 }, { "epoch": 0.6761619190404797, "grad_norm": 0.045509076234092845, "learning_rate": 0.0001674946554600273, "loss": 0.0401, "step": 451 }, { "epoch": 0.6776611694152923, "grad_norm": 0.038954973560474074, "learning_rate": 0.00016730125135097735, "loss": 0.0428, "step": 452 }, { "epoch": 0.679160419790105, "grad_norm": 0.05064178738289288, "learning_rate": 0.00016710738596668632, "loss": 0.0481, "step": 453 }, { "epoch": 0.6806596701649176, "grad_norm": 0.03652262635297044, "learning_rate": 0.00016691306063588583, "loss": 0.0341, "step": 454 }, { "epoch": 0.6821589205397302, "grad_norm": 0.0469218301221786, "learning_rate": 0.00016671827669045998, "loss": 0.0388, "step": 455 }, { "epoch": 0.6836581709145427, "grad_norm": 0.05748273212579017, "learning_rate": 0.00016652303546543608, "loss": 0.0424, "step": 456 }, { "epoch": 0.6851574212893553, "grad_norm": 0.029734098462886188, "learning_rate": 0.00016632733829897566, "loss": 0.0221, "step": 457 }, { "epoch": 0.6866566716641679, "grad_norm": 0.03162041588945904, "learning_rate": 0.00016613118653236518, "loss": 0.0284, "step": 458 }, { "epoch": 0.6881559220389805, "grad_norm": 0.0770559271143547, "learning_rate": 0.00016593458151000688, "loss": 0.0647, "step": 459 }, { "epoch": 0.6896551724137931, "grad_norm": 0.0806809324128278, "learning_rate": 0.0001657375245794096, "loss": 0.0507, "step": 460 }, { "epoch": 0.6911544227886057, "grad_norm": 0.0623039707244334, "learning_rate": 0.0001655400170911794, "loss": 0.0443, "step": 461 }, { "epoch": 0.6926536731634183, "grad_norm": 0.04046812960090323, "learning_rate": 0.00016534206039901057, "loss": 0.047, "step": 462 }, { "epoch": 0.6941529235382309, "grad_norm": 0.09437551047944975, "learning_rate": 0.00016514365585967603, "loss": 0.0626, "step": 463 }, { "epoch": 0.6956521739130435, "grad_norm": 0.07162777105996744, "learning_rate": 0.00016494480483301836, "loss": 0.0503, "step": 464 }, { "epoch": 0.697151424287856, "grad_norm": 0.04109223749911729, "learning_rate": 0.00016474550868194023, "loss": 0.0315, "step": 465 }, { "epoch": 0.6986506746626686, "grad_norm": 0.06077734184621873, "learning_rate": 0.00016454576877239507, "loss": 0.0422, "step": 466 }, { "epoch": 0.7001499250374813, "grad_norm": 0.19884111549839148, "learning_rate": 0.0001643455864733779, "loss": 0.0926, "step": 467 }, { "epoch": 0.7016491754122939, "grad_norm": 0.04642784378882251, "learning_rate": 0.00016414496315691581, "loss": 0.0385, "step": 468 }, { "epoch": 0.7031484257871065, "grad_norm": 0.4393223814451174, "learning_rate": 0.00016394390019805848, "loss": 0.0993, "step": 469 }, { "epoch": 0.704647676161919, "grad_norm": 0.07482643775181062, "learning_rate": 0.000163742398974869, "loss": 0.0489, "step": 470 }, { "epoch": 0.7061469265367316, "grad_norm": 0.04885944436540397, "learning_rate": 0.0001635404608684141, "loss": 0.0348, "step": 471 }, { "epoch": 0.7076461769115442, "grad_norm": 0.049664390973816655, "learning_rate": 0.000163338087262755, "loss": 0.0391, "step": 472 }, { "epoch": 0.7091454272863568, "grad_norm": 0.09885583055337595, "learning_rate": 0.00016313527954493778, "loss": 0.0642, "step": 473 }, { "epoch": 0.7106446776611695, "grad_norm": 0.06584763072102895, "learning_rate": 0.00016293203910498376, "loss": 0.0446, "step": 474 }, { "epoch": 0.712143928035982, "grad_norm": 0.052090571949195315, "learning_rate": 0.00016272836733588017, "loss": 0.0449, "step": 475 }, { "epoch": 0.7136431784107946, "grad_norm": 0.060369511284079136, "learning_rate": 0.00016252426563357055, "loss": 0.0498, "step": 476 }, { "epoch": 0.7151424287856072, "grad_norm": 0.07350090323203447, "learning_rate": 0.00016231973539694504, "loss": 0.0594, "step": 477 }, { "epoch": 0.7166416791604198, "grad_norm": 0.07477778304256645, "learning_rate": 0.00016211477802783103, "loss": 0.0339, "step": 478 }, { "epoch": 0.7181409295352323, "grad_norm": 0.044909007459069075, "learning_rate": 0.00016190939493098344, "loss": 0.0379, "step": 479 }, { "epoch": 0.719640179910045, "grad_norm": 0.062190968309072295, "learning_rate": 0.00016170358751407487, "loss": 0.0758, "step": 480 }, { "epoch": 0.7211394302848576, "grad_norm": 0.05377348813315943, "learning_rate": 0.00016149735718768644, "loss": 0.0329, "step": 481 }, { "epoch": 0.7226386806596702, "grad_norm": 0.053690857438836906, "learning_rate": 0.00016129070536529766, "loss": 0.037, "step": 482 }, { "epoch": 0.7241379310344828, "grad_norm": 0.053806227909165436, "learning_rate": 0.000161083633463277, "loss": 0.0429, "step": 483 }, { "epoch": 0.7256371814092953, "grad_norm": 0.058550946973171134, "learning_rate": 0.00016087614290087208, "loss": 0.0573, "step": 484 }, { "epoch": 0.7271364317841079, "grad_norm": 0.05208602315147625, "learning_rate": 0.00016066823510019998, "loss": 0.0322, "step": 485 }, { "epoch": 0.7286356821589205, "grad_norm": 0.048780008222607175, "learning_rate": 0.0001604599114862375, "loss": 0.0391, "step": 486 }, { "epoch": 0.7301349325337332, "grad_norm": 0.04257645362681531, "learning_rate": 0.00016025117348681132, "loss": 0.0326, "step": 487 }, { "epoch": 0.7316341829085458, "grad_norm": 0.049273729257806875, "learning_rate": 0.00016004202253258842, "loss": 0.0446, "step": 488 }, { "epoch": 0.7331334332833583, "grad_norm": 0.05564516630711474, "learning_rate": 0.00015983246005706593, "loss": 0.0567, "step": 489 }, { "epoch": 0.7346326836581709, "grad_norm": 0.05015891963593426, "learning_rate": 0.0001596224874965616, "loss": 0.0365, "step": 490 }, { "epoch": 0.7361319340329835, "grad_norm": 0.04227849702908071, "learning_rate": 0.00015941210629020388, "loss": 0.0278, "step": 491 }, { "epoch": 0.7376311844077961, "grad_norm": 0.09190608644561755, "learning_rate": 0.00015920131787992197, "loss": 0.1024, "step": 492 }, { "epoch": 0.7391304347826086, "grad_norm": 0.11304924052426712, "learning_rate": 0.00015899012371043604, "loss": 0.0675, "step": 493 }, { "epoch": 0.7406296851574213, "grad_norm": 0.05613518703388969, "learning_rate": 0.00015877852522924732, "loss": 0.0407, "step": 494 }, { "epoch": 0.7421289355322339, "grad_norm": 0.04754947301267445, "learning_rate": 0.00015856652388662808, "loss": 0.0305, "step": 495 }, { "epoch": 0.7436281859070465, "grad_norm": 0.0487917467215864, "learning_rate": 0.00015835412113561175, "loss": 0.0372, "step": 496 }, { "epoch": 0.7451274362818591, "grad_norm": 0.056051719216029486, "learning_rate": 0.00015814131843198308, "loss": 0.033, "step": 497 }, { "epoch": 0.7466266866566716, "grad_norm": 0.08355105349400162, "learning_rate": 0.0001579281172342679, "loss": 0.0618, "step": 498 }, { "epoch": 0.7481259370314842, "grad_norm": 0.09016322211305738, "learning_rate": 0.0001577145190037234, "loss": 0.047, "step": 499 }, { "epoch": 0.7496251874062968, "grad_norm": 0.04982867112956479, "learning_rate": 0.00015750052520432787, "loss": 0.0313, "step": 500 }, { "epoch": 0.7511244377811095, "grad_norm": 0.04381518020575622, "learning_rate": 0.0001572861373027709, "loss": 0.0348, "step": 501 }, { "epoch": 0.7526236881559221, "grad_norm": 0.05899789350925751, "learning_rate": 0.0001570713567684432, "loss": 0.0503, "step": 502 }, { "epoch": 0.7541229385307346, "grad_norm": 0.06316265954005644, "learning_rate": 0.0001568561850734264, "loss": 0.0291, "step": 503 }, { "epoch": 0.7556221889055472, "grad_norm": 0.08632078705291048, "learning_rate": 0.00015664062369248328, "loss": 0.0639, "step": 504 }, { "epoch": 0.7571214392803598, "grad_norm": 0.050182212371613015, "learning_rate": 0.00015642467410304742, "loss": 0.0305, "step": 505 }, { "epoch": 0.7586206896551724, "grad_norm": 0.07158241120198346, "learning_rate": 0.00015620833778521307, "loss": 0.0573, "step": 506 }, { "epoch": 0.760119940029985, "grad_norm": 0.08175078694638557, "learning_rate": 0.00015599161622172517, "loss": 0.0462, "step": 507 }, { "epoch": 0.7616191904047976, "grad_norm": 0.0475536897913194, "learning_rate": 0.00015577451089796905, "loss": 0.0501, "step": 508 }, { "epoch": 0.7631184407796102, "grad_norm": 0.05682015512029848, "learning_rate": 0.00015555702330196023, "loss": 0.0313, "step": 509 }, { "epoch": 0.7646176911544228, "grad_norm": 0.06137429357639015, "learning_rate": 0.00015533915492433443, "loss": 0.0519, "step": 510 }, { "epoch": 0.7661169415292354, "grad_norm": 0.038591640884982854, "learning_rate": 0.00015512090725833705, "loss": 0.0414, "step": 511 }, { "epoch": 0.767616191904048, "grad_norm": 0.055149664940849136, "learning_rate": 0.0001549022817998132, "loss": 0.0318, "step": 512 }, { "epoch": 0.7691154422788605, "grad_norm": 0.04445225405210936, "learning_rate": 0.00015468328004719726, "loss": 0.041, "step": 513 }, { "epoch": 0.7706146926536732, "grad_norm": 0.05691241429851254, "learning_rate": 0.00015446390350150273, "loss": 0.0379, "step": 514 }, { "epoch": 0.7721139430284858, "grad_norm": 0.05761804723554256, "learning_rate": 0.00015424415366631188, "loss": 0.0363, "step": 515 }, { "epoch": 0.7736131934032984, "grad_norm": 0.04929119952169534, "learning_rate": 0.0001540240320477655, "loss": 0.0426, "step": 516 }, { "epoch": 0.775112443778111, "grad_norm": 0.044305931962162155, "learning_rate": 0.0001538035401545525, "loss": 0.0422, "step": 517 }, { "epoch": 0.7766116941529235, "grad_norm": 0.06205510180982029, "learning_rate": 0.00015358267949789966, "loss": 0.0603, "step": 518 }, { "epoch": 0.7781109445277361, "grad_norm": 0.07453649068372155, "learning_rate": 0.00015336145159156115, "loss": 0.0543, "step": 519 }, { "epoch": 0.7796101949025487, "grad_norm": 0.0378467187255617, "learning_rate": 0.00015313985795180828, "loss": 0.04, "step": 520 }, { "epoch": 0.7811094452773614, "grad_norm": 0.057596547352352555, "learning_rate": 0.00015291790009741907, "loss": 0.0451, "step": 521 }, { "epoch": 0.782608695652174, "grad_norm": 0.04081926428494982, "learning_rate": 0.00015269557954966778, "loss": 0.0288, "step": 522 }, { "epoch": 0.7841079460269865, "grad_norm": 0.041887534030808915, "learning_rate": 0.00015247289783231453, "loss": 0.035, "step": 523 }, { "epoch": 0.7856071964017991, "grad_norm": 0.06860023926519257, "learning_rate": 0.0001522498564715949, "loss": 0.049, "step": 524 }, { "epoch": 0.7871064467766117, "grad_norm": 0.04260551619067463, "learning_rate": 0.0001520264569962094, "loss": 0.0427, "step": 525 }, { "epoch": 0.7886056971514243, "grad_norm": 0.03773306965490992, "learning_rate": 0.00015180270093731303, "loss": 0.0245, "step": 526 }, { "epoch": 0.7901049475262368, "grad_norm": 0.03431124298471051, "learning_rate": 0.00015157858982850475, "loss": 0.0289, "step": 527 }, { "epoch": 0.7916041979010495, "grad_norm": 0.07692552283154377, "learning_rate": 0.00015135412520581702, "loss": 0.0661, "step": 528 }, { "epoch": 0.7931034482758621, "grad_norm": 0.0347019843571733, "learning_rate": 0.0001511293086077052, "loss": 0.0267, "step": 529 }, { "epoch": 0.7946026986506747, "grad_norm": 0.05889548183130433, "learning_rate": 0.00015090414157503714, "loss": 0.0321, "step": 530 }, { "epoch": 0.7961019490254873, "grad_norm": 0.07066481807133856, "learning_rate": 0.0001506786256510824, "loss": 0.052, "step": 531 }, { "epoch": 0.7976011994002998, "grad_norm": 0.06186361286136524, "learning_rate": 0.00015045276238150192, "loss": 0.0479, "step": 532 }, { "epoch": 0.7991004497751124, "grad_norm": 0.061605402385295015, "learning_rate": 0.00015022655331433727, "loss": 0.0523, "step": 533 }, { "epoch": 0.800599700149925, "grad_norm": 0.04168389823340015, "learning_rate": 0.00015000000000000001, "loss": 0.0293, "step": 534 }, { "epoch": 0.8020989505247377, "grad_norm": 0.04409075971831587, "learning_rate": 0.00014977310399126123, "loss": 0.0294, "step": 535 }, { "epoch": 0.8035982008995503, "grad_norm": 0.03449538576446893, "learning_rate": 0.00014954586684324078, "loss": 0.0273, "step": 536 }, { "epoch": 0.8050974512743628, "grad_norm": 0.06487922260997829, "learning_rate": 0.00014931829011339659, "loss": 0.1013, "step": 537 }, { "epoch": 0.8065967016491754, "grad_norm": 0.04080309597694205, "learning_rate": 0.00014909037536151409, "loss": 0.0271, "step": 538 }, { "epoch": 0.808095952023988, "grad_norm": 0.039495017673297265, "learning_rate": 0.00014886212414969553, "loss": 0.0274, "step": 539 }, { "epoch": 0.8095952023988006, "grad_norm": 0.08062160395086124, "learning_rate": 0.00014863353804234905, "loss": 0.0417, "step": 540 }, { "epoch": 0.8110944527736131, "grad_norm": 0.05152809022416075, "learning_rate": 0.00014840461860617834, "loss": 0.0359, "step": 541 }, { "epoch": 0.8125937031484258, "grad_norm": 0.04922307205252768, "learning_rate": 0.00014817536741017152, "loss": 0.0308, "step": 542 }, { "epoch": 0.8140929535232384, "grad_norm": 0.07044648065491671, "learning_rate": 0.00014794578602559067, "loss": 0.0566, "step": 543 }, { "epoch": 0.815592203898051, "grad_norm": 0.0465206984979631, "learning_rate": 0.00014771587602596084, "loss": 0.0513, "step": 544 }, { "epoch": 0.8170914542728636, "grad_norm": 0.09746747148479479, "learning_rate": 0.00014748563898705946, "loss": 0.0377, "step": 545 }, { "epoch": 0.8185907046476761, "grad_norm": 0.050647534375834816, "learning_rate": 0.00014725507648690543, "loss": 0.048, "step": 546 }, { "epoch": 0.8200899550224887, "grad_norm": 0.05429153224198516, "learning_rate": 0.00014702419010574825, "loss": 0.0523, "step": 547 }, { "epoch": 0.8215892053973014, "grad_norm": 0.036339453217671985, "learning_rate": 0.00014679298142605734, "loss": 0.0233, "step": 548 }, { "epoch": 0.823088455772114, "grad_norm": 0.08317522768191385, "learning_rate": 0.00014656145203251114, "loss": 0.0619, "step": 549 }, { "epoch": 0.8245877061469266, "grad_norm": 0.04231199661380915, "learning_rate": 0.00014632960351198618, "loss": 0.0477, "step": 550 }, { "epoch": 0.8260869565217391, "grad_norm": 0.1779970750798073, "learning_rate": 0.00014609743745354624, "loss": 0.0978, "step": 551 }, { "epoch": 0.8275862068965517, "grad_norm": 0.06735162872665944, "learning_rate": 0.00014586495544843152, "loss": 0.0435, "step": 552 }, { "epoch": 0.8290854572713643, "grad_norm": 0.06794928340973848, "learning_rate": 0.0001456321590900476, "loss": 0.0548, "step": 553 }, { "epoch": 0.8305847076461769, "grad_norm": 0.040691650099011545, "learning_rate": 0.00014539904997395468, "loss": 0.0307, "step": 554 }, { "epoch": 0.8320839580209896, "grad_norm": 0.06310347101508698, "learning_rate": 0.00014516562969785657, "loss": 0.0518, "step": 555 }, { "epoch": 0.8335832083958021, "grad_norm": 0.06600280603320972, "learning_rate": 0.00014493189986158965, "loss": 0.0283, "step": 556 }, { "epoch": 0.8350824587706147, "grad_norm": 0.048741740097207054, "learning_rate": 0.00014469786206711214, "loss": 0.0369, "step": 557 }, { "epoch": 0.8365817091454273, "grad_norm": 0.0541340846516407, "learning_rate": 0.00014446351791849276, "loss": 0.0408, "step": 558 }, { "epoch": 0.8380809595202399, "grad_norm": 0.06569812050813671, "learning_rate": 0.00014422886902190014, "loss": 0.0559, "step": 559 }, { "epoch": 0.8395802098950524, "grad_norm": 0.07697796259986985, "learning_rate": 0.00014399391698559152, "loss": 0.0359, "step": 560 }, { "epoch": 0.841079460269865, "grad_norm": 0.043380007173197774, "learning_rate": 0.00014375866341990187, "loss": 0.0307, "step": 561 }, { "epoch": 0.8425787106446777, "grad_norm": 0.06012928930562709, "learning_rate": 0.00014352310993723277, "loss": 0.0359, "step": 562 }, { "epoch": 0.8440779610194903, "grad_norm": 0.05916041881920503, "learning_rate": 0.00014328725815204144, "loss": 0.0451, "step": 563 }, { "epoch": 0.8455772113943029, "grad_norm": 0.04584559728661184, "learning_rate": 0.00014305110968082952, "loss": 0.0295, "step": 564 }, { "epoch": 0.8470764617691154, "grad_norm": 0.048484148060844064, "learning_rate": 0.0001428146661421323, "loss": 0.0413, "step": 565 }, { "epoch": 0.848575712143928, "grad_norm": 0.051087086585204396, "learning_rate": 0.00014257792915650728, "loss": 0.0377, "step": 566 }, { "epoch": 0.8500749625187406, "grad_norm": 0.04873404112533848, "learning_rate": 0.00014234090034652324, "loss": 0.0426, "step": 567 }, { "epoch": 0.8515742128935532, "grad_norm": 0.05766048192964452, "learning_rate": 0.00014210358133674912, "loss": 0.033, "step": 568 }, { "epoch": 0.8530734632683659, "grad_norm": 0.054308917320710616, "learning_rate": 0.0001418659737537428, "loss": 0.0379, "step": 569 }, { "epoch": 0.8545727136431784, "grad_norm": 0.0460844103779593, "learning_rate": 0.00014162807922604012, "loss": 0.0289, "step": 570 }, { "epoch": 0.856071964017991, "grad_norm": 0.04438781665413021, "learning_rate": 0.00014138989938414348, "loss": 0.0265, "step": 571 }, { "epoch": 0.8575712143928036, "grad_norm": 0.1327712159880055, "learning_rate": 0.00014115143586051088, "loss": 0.0676, "step": 572 }, { "epoch": 0.8590704647676162, "grad_norm": 0.050894530499255325, "learning_rate": 0.00014091269028954458, "loss": 0.0438, "step": 573 }, { "epoch": 0.8605697151424287, "grad_norm": 0.05468020442615518, "learning_rate": 0.00014067366430758004, "loss": 0.042, "step": 574 }, { "epoch": 0.8620689655172413, "grad_norm": 0.09435379755279395, "learning_rate": 0.00014043435955287452, "loss": 0.0599, "step": 575 }, { "epoch": 0.863568215892054, "grad_norm": 0.046063192994359826, "learning_rate": 0.00014019477766559604, "loss": 0.0348, "step": 576 }, { "epoch": 0.8650674662668666, "grad_norm": 0.052806710216832004, "learning_rate": 0.00013995492028781202, "loss": 0.0432, "step": 577 }, { "epoch": 0.8665667166416792, "grad_norm": 0.05319984444801645, "learning_rate": 0.00013971478906347806, "loss": 0.0362, "step": 578 }, { "epoch": 0.8680659670164917, "grad_norm": 0.04452467143786472, "learning_rate": 0.0001394743856384267, "loss": 0.0314, "step": 579 }, { "epoch": 0.8695652173913043, "grad_norm": 0.05329898478614524, "learning_rate": 0.00013923371166035616, "loss": 0.0464, "step": 580 }, { "epoch": 0.8710644677661169, "grad_norm": 0.060886833574884516, "learning_rate": 0.00013899276877881884, "loss": 0.059, "step": 581 }, { "epoch": 0.8725637181409296, "grad_norm": 0.05294256278470512, "learning_rate": 0.0001387515586452103, "loss": 0.046, "step": 582 }, { "epoch": 0.8740629685157422, "grad_norm": 0.07449191398493356, "learning_rate": 0.00013851008291275782, "loss": 0.0343, "step": 583 }, { "epoch": 0.8755622188905547, "grad_norm": 0.06943351016979123, "learning_rate": 0.000138268343236509, "loss": 0.0558, "step": 584 }, { "epoch": 0.8770614692653673, "grad_norm": 0.06652499644935615, "learning_rate": 0.0001380263412733205, "loss": 0.0451, "step": 585 }, { "epoch": 0.8785607196401799, "grad_norm": 0.05833246316842718, "learning_rate": 0.00013778407868184672, "loss": 0.0457, "step": 586 }, { "epoch": 0.8800599700149925, "grad_norm": 0.04003719271959368, "learning_rate": 0.00013754155712252832, "loss": 0.0283, "step": 587 }, { "epoch": 0.881559220389805, "grad_norm": 0.05607671087323687, "learning_rate": 0.0001372987782575809, "loss": 0.04, "step": 588 }, { "epoch": 0.8830584707646177, "grad_norm": 0.040134500394464416, "learning_rate": 0.00013705574375098365, "loss": 0.0385, "step": 589 }, { "epoch": 0.8845577211394303, "grad_norm": 0.06555335932088245, "learning_rate": 0.00013681245526846783, "loss": 0.072, "step": 590 }, { "epoch": 0.8860569715142429, "grad_norm": 0.05066618534250665, "learning_rate": 0.00013656891447750544, "loss": 0.0428, "step": 591 }, { "epoch": 0.8875562218890555, "grad_norm": 0.04391780563799797, "learning_rate": 0.00013632512304729785, "loss": 0.0367, "step": 592 }, { "epoch": 0.889055472263868, "grad_norm": 0.06116296036386768, "learning_rate": 0.0001360810826487642, "loss": 0.0462, "step": 593 }, { "epoch": 0.8905547226386806, "grad_norm": 0.053934832910471985, "learning_rate": 0.00013583679495453, "loss": 0.046, "step": 594 }, { "epoch": 0.8920539730134932, "grad_norm": 0.045163829899199405, "learning_rate": 0.0001355922616389159, "loss": 0.0362, "step": 595 }, { "epoch": 0.8935532233883059, "grad_norm": 0.08655848208121547, "learning_rate": 0.00013534748437792573, "loss": 0.0406, "step": 596 }, { "epoch": 0.8950524737631185, "grad_norm": 0.05391530518298683, "learning_rate": 0.00013510246484923547, "loss": 0.0293, "step": 597 }, { "epoch": 0.896551724137931, "grad_norm": 0.04845707242796894, "learning_rate": 0.00013485720473218154, "loss": 0.0299, "step": 598 }, { "epoch": 0.8980509745127436, "grad_norm": 0.04916815685149742, "learning_rate": 0.0001346117057077493, "loss": 0.0386, "step": 599 }, { "epoch": 0.8995502248875562, "grad_norm": 0.06617662307537282, "learning_rate": 0.00013436596945856164, "loss": 0.0501, "step": 600 }, { "epoch": 0.9010494752623688, "grad_norm": 0.03991924636314799, "learning_rate": 0.0001341199976688672, "loss": 0.0282, "step": 601 }, { "epoch": 0.9025487256371814, "grad_norm": 0.050388728813716206, "learning_rate": 0.00013387379202452917, "loss": 0.0352, "step": 602 }, { "epoch": 0.904047976011994, "grad_norm": 0.03893791860248331, "learning_rate": 0.0001336273542130134, "loss": 0.021, "step": 603 }, { "epoch": 0.9055472263868066, "grad_norm": 0.059548986139220876, "learning_rate": 0.0001333806859233771, "loss": 0.0798, "step": 604 }, { "epoch": 0.9070464767616192, "grad_norm": 0.04599439337421781, "learning_rate": 0.0001331337888462571, "loss": 0.0351, "step": 605 }, { "epoch": 0.9085457271364318, "grad_norm": 0.05203471422944932, "learning_rate": 0.00013288666467385833, "loss": 0.0295, "step": 606 }, { "epoch": 0.9100449775112444, "grad_norm": 0.055165249926359204, "learning_rate": 0.0001326393150999422, "loss": 0.0345, "step": 607 }, { "epoch": 0.9115442278860569, "grad_norm": 0.13139481496590286, "learning_rate": 0.00013239174181981495, "loss": 0.0534, "step": 608 }, { "epoch": 0.9130434782608695, "grad_norm": 0.04500674501515753, "learning_rate": 0.00013214394653031616, "loss": 0.0203, "step": 609 }, { "epoch": 0.9145427286356822, "grad_norm": 0.05584447753073881, "learning_rate": 0.00013189593092980702, "loss": 0.0344, "step": 610 }, { "epoch": 0.9160419790104948, "grad_norm": 0.041591690138825445, "learning_rate": 0.00013164769671815862, "loss": 0.0201, "step": 611 }, { "epoch": 0.9175412293853074, "grad_norm": 0.05297004561152026, "learning_rate": 0.00013139924559674052, "loss": 0.0381, "step": 612 }, { "epoch": 0.9190404797601199, "grad_norm": 0.04848305915244054, "learning_rate": 0.00013115057926840884, "loss": 0.0349, "step": 613 }, { "epoch": 0.9205397301349325, "grad_norm": 0.03324583015768157, "learning_rate": 0.00013090169943749476, "loss": 0.028, "step": 614 }, { "epoch": 0.9220389805097451, "grad_norm": 0.06071140630728308, "learning_rate": 0.0001306526078097928, "loss": 0.0407, "step": 615 }, { "epoch": 0.9235382308845578, "grad_norm": 0.055677239200326835, "learning_rate": 0.00013040330609254903, "loss": 0.0362, "step": 616 }, { "epoch": 0.9250374812593704, "grad_norm": 0.05138799744779293, "learning_rate": 0.00013015379599444957, "loss": 0.0549, "step": 617 }, { "epoch": 0.9265367316341829, "grad_norm": 0.053161721439074654, "learning_rate": 0.00012990407922560868, "loss": 0.0468, "step": 618 }, { "epoch": 0.9280359820089955, "grad_norm": 0.0638192639644002, "learning_rate": 0.00012965415749755709, "loss": 0.0518, "step": 619 }, { "epoch": 0.9295352323838081, "grad_norm": 0.047780862073805754, "learning_rate": 0.0001294040325232304, "loss": 0.0462, "step": 620 }, { "epoch": 0.9310344827586207, "grad_norm": 0.04020955585417902, "learning_rate": 0.00012915370601695715, "loss": 0.0331, "step": 621 }, { "epoch": 0.9325337331334332, "grad_norm": 0.037639677657712055, "learning_rate": 0.00012890317969444716, "loss": 0.0248, "step": 622 }, { "epoch": 0.9340329835082459, "grad_norm": 0.05345077130247162, "learning_rate": 0.00012865245527277986, "loss": 0.0478, "step": 623 }, { "epoch": 0.9355322338830585, "grad_norm": 0.03711053033129508, "learning_rate": 0.00012840153447039228, "loss": 0.0199, "step": 624 }, { "epoch": 0.9370314842578711, "grad_norm": 0.04911139204212449, "learning_rate": 0.00012815041900706754, "loss": 0.0301, "step": 625 }, { "epoch": 0.9385307346326837, "grad_norm": 0.07752327109727443, "learning_rate": 0.00012789911060392294, "loss": 0.0692, "step": 626 }, { "epoch": 0.9400299850074962, "grad_norm": 0.13232474472513436, "learning_rate": 0.0001276476109833981, "loss": 0.0634, "step": 627 }, { "epoch": 0.9415292353823088, "grad_norm": 0.061926243742957965, "learning_rate": 0.00012739592186924328, "loss": 0.0397, "step": 628 }, { "epoch": 0.9430284857571214, "grad_norm": 0.04356242072040339, "learning_rate": 0.00012714404498650743, "loss": 0.0297, "step": 629 }, { "epoch": 0.9445277361319341, "grad_norm": 0.0376190098528388, "learning_rate": 0.00012689198206152657, "loss": 0.0271, "step": 630 }, { "epoch": 0.9460269865067467, "grad_norm": 0.05455511757321388, "learning_rate": 0.00012663973482191177, "loss": 0.0482, "step": 631 }, { "epoch": 0.9475262368815592, "grad_norm": 0.04442045716631641, "learning_rate": 0.0001263873049965373, "loss": 0.0291, "step": 632 }, { "epoch": 0.9490254872563718, "grad_norm": 0.054450506231905636, "learning_rate": 0.00012613469431552895, "loss": 0.0428, "step": 633 }, { "epoch": 0.9505247376311844, "grad_norm": 0.044327380794719165, "learning_rate": 0.00012588190451025207, "loss": 0.0286, "step": 634 }, { "epoch": 0.952023988005997, "grad_norm": 0.047204720536083225, "learning_rate": 0.00012562893731329967, "loss": 0.0406, "step": 635 }, { "epoch": 0.9535232383808095, "grad_norm": 0.19462933782113725, "learning_rate": 0.00012537579445848058, "loss": 0.1233, "step": 636 }, { "epoch": 0.9550224887556222, "grad_norm": 0.06994302712351169, "learning_rate": 0.00012512247768080756, "loss": 0.0403, "step": 637 }, { "epoch": 0.9565217391304348, "grad_norm": 0.06646995144905621, "learning_rate": 0.0001248689887164855, "loss": 0.0394, "step": 638 }, { "epoch": 0.9580209895052474, "grad_norm": 0.06551860726050548, "learning_rate": 0.00012461532930289933, "loss": 0.0613, "step": 639 }, { "epoch": 0.95952023988006, "grad_norm": 0.04577051878676127, "learning_rate": 0.00012436150117860225, "loss": 0.0267, "step": 640 }, { "epoch": 0.9610194902548725, "grad_norm": 0.0501593890528279, "learning_rate": 0.00012410750608330388, "loss": 0.0415, "step": 641 }, { "epoch": 0.9625187406296851, "grad_norm": 0.04213838081722136, "learning_rate": 0.0001238533457578581, "loss": 0.0363, "step": 642 }, { "epoch": 0.9640179910044977, "grad_norm": 0.04618274978280509, "learning_rate": 0.00012359902194425133, "loss": 0.0268, "step": 643 }, { "epoch": 0.9655172413793104, "grad_norm": 0.054945133222623746, "learning_rate": 0.00012334453638559057, "loss": 0.0411, "step": 644 }, { "epoch": 0.967016491754123, "grad_norm": 0.07629757994067972, "learning_rate": 0.00012308989082609122, "loss": 0.0363, "step": 645 }, { "epoch": 0.9685157421289355, "grad_norm": 0.040863705327469, "learning_rate": 0.00012283508701106557, "loss": 0.0302, "step": 646 }, { "epoch": 0.9700149925037481, "grad_norm": 0.07100180343789458, "learning_rate": 0.0001225801266869104, "loss": 0.044, "step": 647 }, { "epoch": 0.9715142428785607, "grad_norm": 0.06298768167998063, "learning_rate": 0.00012232501160109514, "loss": 0.0396, "step": 648 }, { "epoch": 0.9730134932533733, "grad_norm": 0.05510372575164882, "learning_rate": 0.00012206974350215015, "loss": 0.044, "step": 649 }, { "epoch": 0.974512743628186, "grad_norm": 0.0635198930062088, "learning_rate": 0.00012181432413965428, "loss": 0.0552, "step": 650 }, { "epoch": 0.9760119940029985, "grad_norm": 0.06043612042070538, "learning_rate": 0.00012155875526422332, "loss": 0.0617, "step": 651 }, { "epoch": 0.9775112443778111, "grad_norm": 0.040855526708167285, "learning_rate": 0.00012130303862749767, "loss": 0.0268, "step": 652 }, { "epoch": 0.9790104947526237, "grad_norm": 0.0496157050205137, "learning_rate": 0.00012104717598213056, "loss": 0.0225, "step": 653 }, { "epoch": 0.9805097451274363, "grad_norm": 0.04029704523356512, "learning_rate": 0.00012079116908177593, "loss": 0.0339, "step": 654 }, { "epoch": 0.9820089955022488, "grad_norm": 0.04963288277577189, "learning_rate": 0.00012053501968107641, "loss": 0.0346, "step": 655 }, { "epoch": 0.9835082458770614, "grad_norm": 0.06455026512135428, "learning_rate": 0.00012027872953565125, "loss": 0.0545, "step": 656 }, { "epoch": 0.9850074962518741, "grad_norm": 0.044681736834349894, "learning_rate": 0.00012002230040208447, "loss": 0.0377, "step": 657 }, { "epoch": 0.9865067466266867, "grad_norm": 0.06016026349629435, "learning_rate": 0.00011976573403791262, "loss": 0.0439, "step": 658 }, { "epoch": 0.9880059970014993, "grad_norm": 0.03916055819189204, "learning_rate": 0.00011950903220161285, "loss": 0.0319, "step": 659 }, { "epoch": 0.9895052473763118, "grad_norm": 0.0695326756638391, "learning_rate": 0.00011925219665259075, "loss": 0.0406, "step": 660 }, { "epoch": 0.9910044977511244, "grad_norm": 0.05484556659479051, "learning_rate": 0.00011899522915116848, "loss": 0.0413, "step": 661 }, { "epoch": 0.992503748125937, "grad_norm": 0.06761504502922643, "learning_rate": 0.00011873813145857249, "loss": 0.0502, "step": 662 }, { "epoch": 0.9940029985007496, "grad_norm": 0.03769561443583786, "learning_rate": 0.00011848090533692155, "loss": 0.0351, "step": 663 }, { "epoch": 0.9955022488755623, "grad_norm": 0.05884790570998015, "learning_rate": 0.00011822355254921478, "loss": 0.0499, "step": 664 }, { "epoch": 0.9970014992503748, "grad_norm": 0.05692462358820348, "learning_rate": 0.00011796607485931928, "loss": 0.0367, "step": 665 }, { "epoch": 0.9985007496251874, "grad_norm": 0.04183931910506595, "learning_rate": 0.00011770847403195834, "loss": 0.0362, "step": 666 }, { "epoch": 1.0, "grad_norm": 0.05719242886094664, "learning_rate": 0.0001174507518326992, "loss": 0.0373, "step": 667 }, { "epoch": 1.0, "eval_loss": 0.0589199960231781, "eval_runtime": 760.1358, "eval_samples_per_second": 13.65, "eval_steps_per_second": 0.854, "step": 667 }, { "epoch": 1.0014992503748126, "grad_norm": 0.03762475183995106, "learning_rate": 0.00011719291002794096, "loss": 0.0313, "step": 668 }, { "epoch": 1.0029985007496252, "grad_norm": 0.027248795473336453, "learning_rate": 0.00011693495038490245, "loss": 0.0202, "step": 669 }, { "epoch": 1.0044977511244377, "grad_norm": 0.041298897947596655, "learning_rate": 0.00011667687467161024, "loss": 0.0318, "step": 670 }, { "epoch": 1.0059970014992503, "grad_norm": 0.041340873666624624, "learning_rate": 0.0001164186846568863, "loss": 0.0192, "step": 671 }, { "epoch": 1.0074962518740629, "grad_norm": 0.03416279632561669, "learning_rate": 0.00011616038211033613, "loss": 0.0194, "step": 672 }, { "epoch": 1.0089955022488755, "grad_norm": 0.04269382987865334, "learning_rate": 0.00011590196880233644, "loss": 0.0306, "step": 673 }, { "epoch": 1.0104947526236883, "grad_norm": 0.04456460780740358, "learning_rate": 0.0001156434465040231, "loss": 0.0358, "step": 674 }, { "epoch": 1.0119940029985008, "grad_norm": 0.037486633011275576, "learning_rate": 0.00011538481698727904, "loss": 0.03, "step": 675 }, { "epoch": 1.0134932533733134, "grad_norm": 0.07944497636169431, "learning_rate": 0.00011512608202472194, "loss": 0.047, "step": 676 }, { "epoch": 1.014992503748126, "grad_norm": 0.07167365891576202, "learning_rate": 0.00011486724338969232, "loss": 0.0281, "step": 677 }, { "epoch": 1.0164917541229386, "grad_norm": 0.030600334090124463, "learning_rate": 0.00011460830285624118, "loss": 0.0185, "step": 678 }, { "epoch": 1.0179910044977512, "grad_norm": 0.0334860041071292, "learning_rate": 0.00011434926219911793, "loss": 0.0205, "step": 679 }, { "epoch": 1.0194902548725637, "grad_norm": 0.07452867174589614, "learning_rate": 0.00011409012319375827, "loss": 0.024, "step": 680 }, { "epoch": 1.0209895052473763, "grad_norm": 0.05352241311204091, "learning_rate": 0.00011383088761627192, "loss": 0.0153, "step": 681 }, { "epoch": 1.0224887556221889, "grad_norm": 0.05264974398072599, "learning_rate": 0.00011357155724343045, "loss": 0.0297, "step": 682 }, { "epoch": 1.0239880059970015, "grad_norm": 0.043453829569375796, "learning_rate": 0.00011331213385265524, "loss": 0.0285, "step": 683 }, { "epoch": 1.025487256371814, "grad_norm": 0.04694358278440352, "learning_rate": 0.00011305261922200519, "loss": 0.0308, "step": 684 }, { "epoch": 1.0269865067466266, "grad_norm": 0.052332835831569786, "learning_rate": 0.00011279301513016444, "loss": 0.0276, "step": 685 }, { "epoch": 1.0284857571214392, "grad_norm": 0.05731903299379123, "learning_rate": 0.00011253332335643043, "loss": 0.0368, "step": 686 }, { "epoch": 1.0299850074962518, "grad_norm": 0.03577714273632661, "learning_rate": 0.0001122735456807015, "loss": 0.0159, "step": 687 }, { "epoch": 1.0314842578710646, "grad_norm": 0.030987774585815354, "learning_rate": 0.00011201368388346471, "loss": 0.014, "step": 688 }, { "epoch": 1.0329835082458771, "grad_norm": 0.04126960363600113, "learning_rate": 0.00011175373974578378, "loss": 0.0224, "step": 689 }, { "epoch": 1.0344827586206897, "grad_norm": 0.03027314685021926, "learning_rate": 0.00011149371504928668, "loss": 0.0187, "step": 690 }, { "epoch": 1.0359820089955023, "grad_norm": 0.059016723534456986, "learning_rate": 0.00011123361157615355, "loss": 0.0336, "step": 691 }, { "epoch": 1.0374812593703149, "grad_norm": 0.03460241895483622, "learning_rate": 0.00011097343110910452, "loss": 0.0172, "step": 692 }, { "epoch": 1.0389805097451275, "grad_norm": 0.04905984564105954, "learning_rate": 0.00011071317543138738, "loss": 0.0202, "step": 693 }, { "epoch": 1.04047976011994, "grad_norm": 0.0477670312888383, "learning_rate": 0.00011045284632676536, "loss": 0.0257, "step": 694 }, { "epoch": 1.0419790104947526, "grad_norm": 0.0649026742045147, "learning_rate": 0.000110192445579505, "loss": 0.0528, "step": 695 }, { "epoch": 1.0434782608695652, "grad_norm": 0.03838342057763881, "learning_rate": 0.00010993197497436391, "loss": 0.0176, "step": 696 }, { "epoch": 1.0449775112443778, "grad_norm": 0.05081060598139383, "learning_rate": 0.00010967143629657842, "loss": 0.024, "step": 697 }, { "epoch": 1.0464767616191903, "grad_norm": 0.07793509916227749, "learning_rate": 0.00010941083133185146, "loss": 0.0475, "step": 698 }, { "epoch": 1.047976011994003, "grad_norm": 0.045160852102844276, "learning_rate": 0.00010915016186634026, "loss": 0.0169, "step": 699 }, { "epoch": 1.0494752623688155, "grad_norm": 0.08881360543272801, "learning_rate": 0.00010888942968664417, "loss": 0.023, "step": 700 }, { "epoch": 1.050974512743628, "grad_norm": 0.038412054850676614, "learning_rate": 0.00010862863657979237, "loss": 0.018, "step": 701 }, { "epoch": 1.0524737631184409, "grad_norm": 0.051809463503864034, "learning_rate": 0.00010836778433323158, "loss": 0.0266, "step": 702 }, { "epoch": 1.0539730134932535, "grad_norm": 0.059635918308077476, "learning_rate": 0.00010810687473481389, "loss": 0.0293, "step": 703 }, { "epoch": 1.055472263868066, "grad_norm": 0.05223829787695676, "learning_rate": 0.0001078459095727845, "loss": 0.0333, "step": 704 }, { "epoch": 1.0569715142428786, "grad_norm": 0.04788415350708223, "learning_rate": 0.00010758489063576939, "loss": 0.0225, "step": 705 }, { "epoch": 1.0584707646176912, "grad_norm": 0.07606032889772184, "learning_rate": 0.00010732381971276318, "loss": 0.0504, "step": 706 }, { "epoch": 1.0599700149925038, "grad_norm": 0.041830044788424676, "learning_rate": 0.00010706269859311669, "loss": 0.0339, "step": 707 }, { "epoch": 1.0614692653673163, "grad_norm": 0.056499034345126806, "learning_rate": 0.00010680152906652483, "loss": 0.0797, "step": 708 }, { "epoch": 1.062968515742129, "grad_norm": 0.050667538240860205, "learning_rate": 0.00010654031292301432, "loss": 0.0385, "step": 709 }, { "epoch": 1.0644677661169415, "grad_norm": 0.03606037951763869, "learning_rate": 0.00010627905195293135, "loss": 0.0249, "step": 710 }, { "epoch": 1.065967016491754, "grad_norm": 0.06482813154069911, "learning_rate": 0.00010601774794692935, "loss": 0.0322, "step": 711 }, { "epoch": 1.0674662668665666, "grad_norm": 0.031472797559762344, "learning_rate": 0.00010575640269595675, "loss": 0.0163, "step": 712 }, { "epoch": 1.0689655172413792, "grad_norm": 0.04865531174873738, "learning_rate": 0.0001054950179912446, "loss": 0.0265, "step": 713 }, { "epoch": 1.0704647676161918, "grad_norm": 0.035242178613106155, "learning_rate": 0.0001052335956242944, "loss": 0.017, "step": 714 }, { "epoch": 1.0719640179910046, "grad_norm": 0.10969495785314283, "learning_rate": 0.0001049721373868658, "loss": 0.051, "step": 715 }, { "epoch": 1.0734632683658172, "grad_norm": 0.055243920108870075, "learning_rate": 0.00010471064507096426, "loss": 0.0299, "step": 716 }, { "epoch": 1.0749625187406298, "grad_norm": 0.027766012133857647, "learning_rate": 0.00010444912046882888, "loss": 0.0121, "step": 717 }, { "epoch": 1.0764617691154423, "grad_norm": 0.05149363616690094, "learning_rate": 0.00010418756537291996, "loss": 0.0237, "step": 718 }, { "epoch": 1.077961019490255, "grad_norm": 0.0435854981748732, "learning_rate": 0.00010392598157590688, "loss": 0.0216, "step": 719 }, { "epoch": 1.0794602698650675, "grad_norm": 0.048516448094552064, "learning_rate": 0.00010366437087065564, "loss": 0.03, "step": 720 }, { "epoch": 1.08095952023988, "grad_norm": 0.05499780130973038, "learning_rate": 0.00010340273505021674, "loss": 0.0263, "step": 721 }, { "epoch": 1.0824587706146926, "grad_norm": 0.06511989459756257, "learning_rate": 0.00010314107590781284, "loss": 0.0374, "step": 722 }, { "epoch": 1.0839580209895052, "grad_norm": 0.05279506351690483, "learning_rate": 0.00010287939523682634, "loss": 0.026, "step": 723 }, { "epoch": 1.0854572713643178, "grad_norm": 0.049766335190166044, "learning_rate": 0.00010261769483078733, "loss": 0.0362, "step": 724 }, { "epoch": 1.0869565217391304, "grad_norm": 0.0616925740357659, "learning_rate": 0.00010235597648336104, "loss": 0.0272, "step": 725 }, { "epoch": 1.088455772113943, "grad_norm": 0.035580584914971236, "learning_rate": 0.0001020942419883357, "loss": 0.0246, "step": 726 }, { "epoch": 1.0899550224887555, "grad_norm": 0.05651314515303107, "learning_rate": 0.0001018324931396103, "loss": 0.0389, "step": 727 }, { "epoch": 1.0914542728635683, "grad_norm": 0.03403275870606777, "learning_rate": 0.00010157073173118208, "loss": 0.017, "step": 728 }, { "epoch": 1.092953523238381, "grad_norm": 0.04001508080874846, "learning_rate": 0.00010130895955713445, "loss": 0.0344, "step": 729 }, { "epoch": 1.0944527736131935, "grad_norm": 0.038898635048905054, "learning_rate": 0.00010104717841162458, "loss": 0.0139, "step": 730 }, { "epoch": 1.095952023988006, "grad_norm": 0.04852969796297334, "learning_rate": 0.00010078539008887114, "loss": 0.0303, "step": 731 }, { "epoch": 1.0974512743628186, "grad_norm": 0.030415262260471615, "learning_rate": 0.00010052359638314195, "loss": 0.0209, "step": 732 }, { "epoch": 1.0989505247376312, "grad_norm": 0.03296035338712638, "learning_rate": 0.00010026179908874183, "loss": 0.0176, "step": 733 }, { "epoch": 1.1004497751124438, "grad_norm": 0.06912112785698081, "learning_rate": 0.0001, "loss": 0.0911, "step": 734 }, { "epoch": 1.1019490254872564, "grad_norm": 0.04126417647100554, "learning_rate": 9.973820091125821e-05, "loss": 0.0278, "step": 735 }, { "epoch": 1.103448275862069, "grad_norm": 0.06337110286285316, "learning_rate": 9.947640361685804e-05, "loss": 0.0423, "step": 736 }, { "epoch": 1.1049475262368815, "grad_norm": 0.04785214604294027, "learning_rate": 9.921460991112891e-05, "loss": 0.019, "step": 737 }, { "epoch": 1.106446776611694, "grad_norm": 0.03813515686538096, "learning_rate": 9.895282158837545e-05, "loss": 0.0147, "step": 738 }, { "epoch": 1.1079460269865067, "grad_norm": 0.04010157831438066, "learning_rate": 9.869104044286558e-05, "loss": 0.0276, "step": 739 }, { "epoch": 1.1094452773613193, "grad_norm": 0.03406965983248011, "learning_rate": 9.842926826881796e-05, "loss": 0.0149, "step": 740 }, { "epoch": 1.1109445277361318, "grad_norm": 0.05122540759813054, "learning_rate": 9.816750686038973e-05, "loss": 0.0218, "step": 741 }, { "epoch": 1.1124437781109444, "grad_norm": 0.046947765546721, "learning_rate": 9.790575801166432e-05, "loss": 0.025, "step": 742 }, { "epoch": 1.1139430284857572, "grad_norm": 0.07498886528603434, "learning_rate": 9.764402351663901e-05, "loss": 0.0434, "step": 743 }, { "epoch": 1.1154422788605698, "grad_norm": 0.03187418225048272, "learning_rate": 9.73823051692127e-05, "loss": 0.0164, "step": 744 }, { "epoch": 1.1169415292353824, "grad_norm": 0.05124246170818971, "learning_rate": 9.712060476317367e-05, "loss": 0.027, "step": 745 }, { "epoch": 1.118440779610195, "grad_norm": 0.04064336939991715, "learning_rate": 9.685892409218717e-05, "loss": 0.0312, "step": 746 }, { "epoch": 1.1199400299850075, "grad_norm": 0.02940931109265308, "learning_rate": 9.659726494978325e-05, "loss": 0.0138, "step": 747 }, { "epoch": 1.12143928035982, "grad_norm": 0.07160694268820639, "learning_rate": 9.633562912934436e-05, "loss": 0.0638, "step": 748 }, { "epoch": 1.1229385307346327, "grad_norm": 0.03268947669426935, "learning_rate": 9.607401842409317e-05, "loss": 0.0152, "step": 749 }, { "epoch": 1.1244377811094453, "grad_norm": 0.04469947279909564, "learning_rate": 9.581243462708006e-05, "loss": 0.0519, "step": 750 }, { "epoch": 1.1259370314842578, "grad_norm": 0.03733568199384118, "learning_rate": 9.555087953117114e-05, "loss": 0.0191, "step": 751 }, { "epoch": 1.1274362818590704, "grad_norm": 0.04957459904827682, "learning_rate": 9.528935492903575e-05, "loss": 0.0258, "step": 752 }, { "epoch": 1.128935532233883, "grad_norm": 0.04291232620005789, "learning_rate": 9.502786261313423e-05, "loss": 0.0186, "step": 753 }, { "epoch": 1.1304347826086956, "grad_norm": 0.03645205093124093, "learning_rate": 9.476640437570562e-05, "loss": 0.0179, "step": 754 }, { "epoch": 1.1319340329835081, "grad_norm": 0.036883531854657765, "learning_rate": 9.450498200875546e-05, "loss": 0.0181, "step": 755 }, { "epoch": 1.133433283358321, "grad_norm": 0.03611209493776261, "learning_rate": 9.424359730404329e-05, "loss": 0.0149, "step": 756 }, { "epoch": 1.1349325337331335, "grad_norm": 0.03543885857842899, "learning_rate": 9.398225205307066e-05, "loss": 0.0179, "step": 757 }, { "epoch": 1.136431784107946, "grad_norm": 0.04885756178865602, "learning_rate": 9.372094804706867e-05, "loss": 0.0363, "step": 758 }, { "epoch": 1.1379310344827587, "grad_norm": 0.045201667917713226, "learning_rate": 9.345968707698569e-05, "loss": 0.015, "step": 759 }, { "epoch": 1.1394302848575713, "grad_norm": 0.045725248124270736, "learning_rate": 9.319847093347522e-05, "loss": 0.0242, "step": 760 }, { "epoch": 1.1409295352323838, "grad_norm": 0.03246760399537495, "learning_rate": 9.293730140688336e-05, "loss": 0.0136, "step": 761 }, { "epoch": 1.1424287856071964, "grad_norm": 0.04842555683642214, "learning_rate": 9.267618028723686e-05, "loss": 0.0178, "step": 762 }, { "epoch": 1.143928035982009, "grad_norm": 0.09178292202784064, "learning_rate": 9.241510936423062e-05, "loss": 0.0293, "step": 763 }, { "epoch": 1.1454272863568216, "grad_norm": 0.03732797810025069, "learning_rate": 9.215409042721552e-05, "loss": 0.032, "step": 764 }, { "epoch": 1.1469265367316341, "grad_norm": 0.06084284216603883, "learning_rate": 9.189312526518613e-05, "loss": 0.0298, "step": 765 }, { "epoch": 1.1484257871064467, "grad_norm": 0.05070843744675052, "learning_rate": 9.163221566676847e-05, "loss": 0.0115, "step": 766 }, { "epoch": 1.1499250374812593, "grad_norm": 0.040941353399746576, "learning_rate": 9.137136342020768e-05, "loss": 0.0219, "step": 767 }, { "epoch": 1.1514242878560719, "grad_norm": 0.056138839618842994, "learning_rate": 9.111057031335585e-05, "loss": 0.0209, "step": 768 }, { "epoch": 1.1529235382308847, "grad_norm": 0.053017337582656095, "learning_rate": 9.084983813365978e-05, "loss": 0.0139, "step": 769 }, { "epoch": 1.1544227886056972, "grad_norm": 0.0437160406734324, "learning_rate": 9.058916866814858e-05, "loss": 0.0292, "step": 770 }, { "epoch": 1.1559220389805098, "grad_norm": 0.06517357369678477, "learning_rate": 9.032856370342158e-05, "loss": 0.0314, "step": 771 }, { "epoch": 1.1574212893553224, "grad_norm": 0.05616399006235695, "learning_rate": 9.006802502563612e-05, "loss": 0.0251, "step": 772 }, { "epoch": 1.158920539730135, "grad_norm": 0.046115228142336546, "learning_rate": 8.980755442049502e-05, "loss": 0.0198, "step": 773 }, { "epoch": 1.1604197901049476, "grad_norm": 0.03835516395560399, "learning_rate": 8.954715367323468e-05, "loss": 0.0264, "step": 774 }, { "epoch": 1.1619190404797601, "grad_norm": 0.05643079022619911, "learning_rate": 8.928682456861265e-05, "loss": 0.0255, "step": 775 }, { "epoch": 1.1634182908545727, "grad_norm": 0.03736421317747863, "learning_rate": 8.902656889089548e-05, "loss": 0.0185, "step": 776 }, { "epoch": 1.1649175412293853, "grad_norm": 0.051100575541428585, "learning_rate": 8.876638842384644e-05, "loss": 0.0201, "step": 777 }, { "epoch": 1.1664167916041979, "grad_norm": 0.03430703468298361, "learning_rate": 8.850628495071336e-05, "loss": 0.0152, "step": 778 }, { "epoch": 1.1679160419790104, "grad_norm": 0.030283377580051823, "learning_rate": 8.824626025421626e-05, "loss": 0.0189, "step": 779 }, { "epoch": 1.169415292353823, "grad_norm": 0.04693707293564595, "learning_rate": 8.79863161165353e-05, "loss": 0.0265, "step": 780 }, { "epoch": 1.1709145427286356, "grad_norm": 0.0670635890351688, "learning_rate": 8.772645431929851e-05, "loss": 0.043, "step": 781 }, { "epoch": 1.1724137931034484, "grad_norm": 0.043848670372825894, "learning_rate": 8.746667664356956e-05, "loss": 0.0217, "step": 782 }, { "epoch": 1.1739130434782608, "grad_norm": 0.05412472149445996, "learning_rate": 8.720698486983557e-05, "loss": 0.0197, "step": 783 }, { "epoch": 1.1754122938530736, "grad_norm": 0.019649617759211542, "learning_rate": 8.694738077799488e-05, "loss": 0.0073, "step": 784 }, { "epoch": 1.1769115442278861, "grad_norm": 0.05225640536933059, "learning_rate": 8.668786614734478e-05, "loss": 0.0276, "step": 785 }, { "epoch": 1.1784107946026987, "grad_norm": 0.04315118247949963, "learning_rate": 8.642844275656957e-05, "loss": 0.0324, "step": 786 }, { "epoch": 1.1799100449775113, "grad_norm": 0.029412653837457406, "learning_rate": 8.616911238372812e-05, "loss": 0.014, "step": 787 }, { "epoch": 1.1814092953523239, "grad_norm": 0.04133482214444302, "learning_rate": 8.590987680624174e-05, "loss": 0.0202, "step": 788 }, { "epoch": 1.1829085457271364, "grad_norm": 0.06765748401829244, "learning_rate": 8.565073780088208e-05, "loss": 0.0441, "step": 789 }, { "epoch": 1.184407796101949, "grad_norm": 0.05271546437062786, "learning_rate": 8.539169714375885e-05, "loss": 0.0256, "step": 790 }, { "epoch": 1.1859070464767616, "grad_norm": 0.040199086711978846, "learning_rate": 8.51327566103077e-05, "loss": 0.0139, "step": 791 }, { "epoch": 1.1874062968515742, "grad_norm": 0.04400948652648687, "learning_rate": 8.487391797527808e-05, "loss": 0.0298, "step": 792 }, { "epoch": 1.1889055472263867, "grad_norm": 0.04023525299067876, "learning_rate": 8.461518301272099e-05, "loss": 0.025, "step": 793 }, { "epoch": 1.1904047976011993, "grad_norm": 0.03582585876670016, "learning_rate": 8.435655349597689e-05, "loss": 0.0206, "step": 794 }, { "epoch": 1.191904047976012, "grad_norm": 0.04826655494275668, "learning_rate": 8.409803119766361e-05, "loss": 0.0242, "step": 795 }, { "epoch": 1.1934032983508245, "grad_norm": 0.05078096150994414, "learning_rate": 8.383961788966391e-05, "loss": 0.0212, "step": 796 }, { "epoch": 1.1949025487256373, "grad_norm": 0.04974688478126328, "learning_rate": 8.358131534311372e-05, "loss": 0.0262, "step": 797 }, { "epoch": 1.1964017991004499, "grad_norm": 0.03592037995042853, "learning_rate": 8.332312532838978e-05, "loss": 0.0159, "step": 798 }, { "epoch": 1.1979010494752624, "grad_norm": 0.0491861501338674, "learning_rate": 8.306504961509754e-05, "loss": 0.0234, "step": 799 }, { "epoch": 1.199400299850075, "grad_norm": 0.027759792015070672, "learning_rate": 8.280708997205904e-05, "loss": 0.0136, "step": 800 }, { "epoch": 1.2008995502248876, "grad_norm": 0.0548378934219744, "learning_rate": 8.254924816730082e-05, "loss": 0.0251, "step": 801 }, { "epoch": 1.2023988005997002, "grad_norm": 0.031042646875865575, "learning_rate": 8.229152596804168e-05, "loss": 0.0114, "step": 802 }, { "epoch": 1.2038980509745127, "grad_norm": 0.04185558020402479, "learning_rate": 8.203392514068074e-05, "loss": 0.0196, "step": 803 }, { "epoch": 1.2053973013493253, "grad_norm": 0.03573774524393669, "learning_rate": 8.177644745078526e-05, "loss": 0.029, "step": 804 }, { "epoch": 1.206896551724138, "grad_norm": 0.04500572749598161, "learning_rate": 8.151909466307845e-05, "loss": 0.0283, "step": 805 }, { "epoch": 1.2083958020989505, "grad_norm": 0.055472853240349465, "learning_rate": 8.126186854142752e-05, "loss": 0.0278, "step": 806 }, { "epoch": 1.209895052473763, "grad_norm": 0.04737096525754299, "learning_rate": 8.100477084883156e-05, "loss": 0.0212, "step": 807 }, { "epoch": 1.2113943028485756, "grad_norm": 0.06377978614115022, "learning_rate": 8.074780334740928e-05, "loss": 0.0264, "step": 808 }, { "epoch": 1.2128935532233882, "grad_norm": 0.1037734762277636, "learning_rate": 8.049096779838719e-05, "loss": 0.0653, "step": 809 }, { "epoch": 1.214392803598201, "grad_norm": 0.04938456491428775, "learning_rate": 8.023426596208739e-05, "loss": 0.0359, "step": 810 }, { "epoch": 1.2158920539730136, "grad_norm": 0.03871245605295673, "learning_rate": 7.997769959791554e-05, "loss": 0.0189, "step": 811 }, { "epoch": 1.2173913043478262, "grad_norm": 0.04836114902803376, "learning_rate": 7.972127046434878e-05, "loss": 0.0216, "step": 812 }, { "epoch": 1.2188905547226387, "grad_norm": 0.06566867333027651, "learning_rate": 7.946498031892364e-05, "loss": 0.0289, "step": 813 }, { "epoch": 1.2203898050974513, "grad_norm": 0.07837421636117799, "learning_rate": 7.920883091822408e-05, "loss": 0.0312, "step": 814 }, { "epoch": 1.221889055472264, "grad_norm": 0.05650900430884517, "learning_rate": 7.895282401786945e-05, "loss": 0.025, "step": 815 }, { "epoch": 1.2233883058470765, "grad_norm": 0.04878788564349848, "learning_rate": 7.869696137250235e-05, "loss": 0.0268, "step": 816 }, { "epoch": 1.224887556221889, "grad_norm": 0.048966519858149045, "learning_rate": 7.844124473577672e-05, "loss": 0.0315, "step": 817 }, { "epoch": 1.2263868065967016, "grad_norm": 0.03824407231228468, "learning_rate": 7.818567586034577e-05, "loss": 0.0198, "step": 818 }, { "epoch": 1.2278860569715142, "grad_norm": 0.0529744116915096, "learning_rate": 7.79302564978499e-05, "loss": 0.0381, "step": 819 }, { "epoch": 1.2293853073463268, "grad_norm": 0.05426023606207575, "learning_rate": 7.767498839890488e-05, "loss": 0.0295, "step": 820 }, { "epoch": 1.2308845577211394, "grad_norm": 0.03534659066797283, "learning_rate": 7.741987331308964e-05, "loss": 0.0197, "step": 821 }, { "epoch": 1.232383808095952, "grad_norm": 0.15173262879838179, "learning_rate": 7.716491298893442e-05, "loss": 0.0299, "step": 822 }, { "epoch": 1.2338830584707647, "grad_norm": 0.04753989562431895, "learning_rate": 7.691010917390874e-05, "loss": 0.021, "step": 823 }, { "epoch": 1.235382308845577, "grad_norm": 0.045421468563024, "learning_rate": 7.66554636144095e-05, "loss": 0.0315, "step": 824 }, { "epoch": 1.23688155922039, "grad_norm": 0.047073846935776716, "learning_rate": 7.64009780557487e-05, "loss": 0.0214, "step": 825 }, { "epoch": 1.2383808095952025, "grad_norm": 0.05224603067353402, "learning_rate": 7.614665424214193e-05, "loss": 0.0217, "step": 826 }, { "epoch": 1.239880059970015, "grad_norm": 0.04520300461414657, "learning_rate": 7.589249391669616e-05, "loss": 0.0213, "step": 827 }, { "epoch": 1.2413793103448276, "grad_norm": 0.04736894258409455, "learning_rate": 7.563849882139776e-05, "loss": 0.0237, "step": 828 }, { "epoch": 1.2428785607196402, "grad_norm": 0.053766180923070186, "learning_rate": 7.53846706971007e-05, "loss": 0.0323, "step": 829 }, { "epoch": 1.2443778110944528, "grad_norm": 0.05528412145749755, "learning_rate": 7.513101128351454e-05, "loss": 0.0487, "step": 830 }, { "epoch": 1.2458770614692654, "grad_norm": 0.05566014947606353, "learning_rate": 7.487752231919245e-05, "loss": 0.0267, "step": 831 }, { "epoch": 1.247376311844078, "grad_norm": 0.04083106652022665, "learning_rate": 7.462420554151944e-05, "loss": 0.0278, "step": 832 }, { "epoch": 1.2488755622188905, "grad_norm": 0.06891562106271396, "learning_rate": 7.437106268670034e-05, "loss": 0.0294, "step": 833 }, { "epoch": 1.250374812593703, "grad_norm": 0.07342638900103611, "learning_rate": 7.411809548974792e-05, "loss": 0.0845, "step": 834 }, { "epoch": 1.2518740629685157, "grad_norm": 0.050466963864324595, "learning_rate": 7.386530568447107e-05, "loss": 0.0298, "step": 835 }, { "epoch": 1.2533733133433285, "grad_norm": 0.06686910207648407, "learning_rate": 7.361269500346274e-05, "loss": 0.0275, "step": 836 }, { "epoch": 1.2548725637181408, "grad_norm": 0.046075881059396175, "learning_rate": 7.336026517808827e-05, "loss": 0.0214, "step": 837 }, { "epoch": 1.2563718140929536, "grad_norm": 0.049483313231029974, "learning_rate": 7.310801793847344e-05, "loss": 0.0295, "step": 838 }, { "epoch": 1.2578710644677662, "grad_norm": 0.06374973164823877, "learning_rate": 7.285595501349258e-05, "loss": 0.0544, "step": 839 }, { "epoch": 1.2593703148425788, "grad_norm": 0.2122029055345762, "learning_rate": 7.260407813075676e-05, "loss": 0.099, "step": 840 }, { "epoch": 1.2608695652173914, "grad_norm": 0.02852293268003914, "learning_rate": 7.235238901660195e-05, "loss": 0.0166, "step": 841 }, { "epoch": 1.262368815592204, "grad_norm": 0.04115201623040188, "learning_rate": 7.210088939607708e-05, "loss": 0.0254, "step": 842 }, { "epoch": 1.2638680659670165, "grad_norm": 0.05628533766124335, "learning_rate": 7.184958099293246e-05, "loss": 0.0365, "step": 843 }, { "epoch": 1.265367316341829, "grad_norm": 0.08094057068973667, "learning_rate": 7.159846552960774e-05, "loss": 0.0222, "step": 844 }, { "epoch": 1.2668665667166417, "grad_norm": 0.03799623111533529, "learning_rate": 7.134754472722017e-05, "loss": 0.0224, "step": 845 }, { "epoch": 1.2683658170914542, "grad_norm": 0.04700845879809844, "learning_rate": 7.109682030555283e-05, "loss": 0.0224, "step": 846 }, { "epoch": 1.2698650674662668, "grad_norm": 0.048949756388667436, "learning_rate": 7.08462939830429e-05, "loss": 0.018, "step": 847 }, { "epoch": 1.2713643178410794, "grad_norm": 0.047029812926556, "learning_rate": 7.059596747676962e-05, "loss": 0.0266, "step": 848 }, { "epoch": 1.272863568215892, "grad_norm": 0.030351552812581083, "learning_rate": 7.034584250244291e-05, "loss": 0.0151, "step": 849 }, { "epoch": 1.2743628185907045, "grad_norm": 0.052275221441315155, "learning_rate": 7.009592077439134e-05, "loss": 0.0198, "step": 850 }, { "epoch": 1.2758620689655173, "grad_norm": 0.04376783013783442, "learning_rate": 6.984620400555044e-05, "loss": 0.0167, "step": 851 }, { "epoch": 1.2773613193403297, "grad_norm": 0.054785397673240276, "learning_rate": 6.959669390745097e-05, "loss": 0.0209, "step": 852 }, { "epoch": 1.2788605697151425, "grad_norm": 0.04062686089747498, "learning_rate": 6.934739219020726e-05, "loss": 0.015, "step": 853 }, { "epoch": 1.280359820089955, "grad_norm": 0.06936614542121676, "learning_rate": 6.909830056250527e-05, "loss": 0.0272, "step": 854 }, { "epoch": 1.2818590704647677, "grad_norm": 0.06026615690676764, "learning_rate": 6.88494207315912e-05, "loss": 0.0271, "step": 855 }, { "epoch": 1.2833583208395802, "grad_norm": 0.07021407853748118, "learning_rate": 6.860075440325951e-05, "loss": 0.0308, "step": 856 }, { "epoch": 1.2848575712143928, "grad_norm": 0.052574527846576825, "learning_rate": 6.835230328184138e-05, "loss": 0.0175, "step": 857 }, { "epoch": 1.2863568215892054, "grad_norm": 0.05298669412741438, "learning_rate": 6.8104069070193e-05, "loss": 0.0176, "step": 858 }, { "epoch": 1.287856071964018, "grad_norm": 0.04213736228746085, "learning_rate": 6.785605346968386e-05, "loss": 0.0175, "step": 859 }, { "epoch": 1.2893553223388305, "grad_norm": 0.03297170359235013, "learning_rate": 6.760825818018508e-05, "loss": 0.0293, "step": 860 }, { "epoch": 1.2908545727136431, "grad_norm": 0.07532993960457951, "learning_rate": 6.736068490005783e-05, "loss": 0.0361, "step": 861 }, { "epoch": 1.2923538230884557, "grad_norm": 0.06949245335822227, "learning_rate": 6.711333532614168e-05, "loss": 0.0428, "step": 862 }, { "epoch": 1.2938530734632683, "grad_norm": 0.033794502875320105, "learning_rate": 6.68662111537429e-05, "loss": 0.019, "step": 863 }, { "epoch": 1.295352323838081, "grad_norm": 0.042862124096238796, "learning_rate": 6.661931407662292e-05, "loss": 0.0165, "step": 864 }, { "epoch": 1.2968515742128934, "grad_norm": 0.03762469114322251, "learning_rate": 6.637264578698664e-05, "loss": 0.0177, "step": 865 }, { "epoch": 1.2983508245877062, "grad_norm": 0.06681907271371797, "learning_rate": 6.612620797547087e-05, "loss": 0.0425, "step": 866 }, { "epoch": 1.2998500749625188, "grad_norm": 0.06461028945930743, "learning_rate": 6.588000233113282e-05, "loss": 0.033, "step": 867 }, { "epoch": 1.3013493253373314, "grad_norm": 0.03602981021835219, "learning_rate": 6.56340305414384e-05, "loss": 0.0185, "step": 868 }, { "epoch": 1.302848575712144, "grad_norm": 0.07041331964732868, "learning_rate": 6.538829429225069e-05, "loss": 0.0375, "step": 869 }, { "epoch": 1.3043478260869565, "grad_norm": 0.053845021312289056, "learning_rate": 6.51427952678185e-05, "loss": 0.0264, "step": 870 }, { "epoch": 1.3058470764617691, "grad_norm": 0.03459818578421578, "learning_rate": 6.489753515076456e-05, "loss": 0.0145, "step": 871 }, { "epoch": 1.3073463268365817, "grad_norm": 0.04351524337603825, "learning_rate": 6.465251562207431e-05, "loss": 0.0207, "step": 872 }, { "epoch": 1.3088455772113943, "grad_norm": 0.04456815741914844, "learning_rate": 6.440773836108413e-05, "loss": 0.0278, "step": 873 }, { "epoch": 1.3103448275862069, "grad_norm": 0.061834383611745806, "learning_rate": 6.416320504546997e-05, "loss": 0.0332, "step": 874 }, { "epoch": 1.3118440779610194, "grad_norm": 0.045688059219150534, "learning_rate": 6.391891735123582e-05, "loss": 0.0435, "step": 875 }, { "epoch": 1.313343328335832, "grad_norm": 0.0499097200922593, "learning_rate": 6.367487695270217e-05, "loss": 0.0225, "step": 876 }, { "epoch": 1.3148425787106448, "grad_norm": 0.03930515307322255, "learning_rate": 6.343108552249457e-05, "loss": 0.0213, "step": 877 }, { "epoch": 1.3163418290854572, "grad_norm": 0.04750578901510908, "learning_rate": 6.318754473153221e-05, "loss": 0.0344, "step": 878 }, { "epoch": 1.31784107946027, "grad_norm": 0.05610290200169092, "learning_rate": 6.294425624901638e-05, "loss": 0.0322, "step": 879 }, { "epoch": 1.3193403298350825, "grad_norm": 0.029473494282681995, "learning_rate": 6.27012217424191e-05, "loss": 0.0134, "step": 880 }, { "epoch": 1.3208395802098951, "grad_norm": 0.05781224980544649, "learning_rate": 6.245844287747168e-05, "loss": 0.0354, "step": 881 }, { "epoch": 1.3223388305847077, "grad_norm": 0.06267338674733103, "learning_rate": 6.22159213181533e-05, "loss": 0.0361, "step": 882 }, { "epoch": 1.3238380809595203, "grad_norm": 0.038066306071301584, "learning_rate": 6.197365872667952e-05, "loss": 0.0181, "step": 883 }, { "epoch": 1.3253373313343328, "grad_norm": 0.07878387260295196, "learning_rate": 6.173165676349103e-05, "loss": 0.0682, "step": 884 }, { "epoch": 1.3268365817091454, "grad_norm": 0.04524683438782359, "learning_rate": 6.14899170872422e-05, "loss": 0.0295, "step": 885 }, { "epoch": 1.328335832083958, "grad_norm": 0.034393149243530124, "learning_rate": 6.12484413547897e-05, "loss": 0.0129, "step": 886 }, { "epoch": 1.3298350824587706, "grad_norm": 0.04589569278929691, "learning_rate": 6.1007231221181206e-05, "loss": 0.0214, "step": 887 }, { "epoch": 1.3313343328335832, "grad_norm": 0.0999532474797914, "learning_rate": 6.076628833964388e-05, "loss": 0.0241, "step": 888 }, { "epoch": 1.3328335832083957, "grad_norm": 0.06084071636014458, "learning_rate": 6.052561436157329e-05, "loss": 0.0242, "step": 889 }, { "epoch": 1.3343328335832085, "grad_norm": 0.04917934868115785, "learning_rate": 6.0285210936521955e-05, "loss": 0.0197, "step": 890 }, { "epoch": 1.3358320839580209, "grad_norm": 0.069354361338318, "learning_rate": 6.0045079712188014e-05, "loss": 0.0275, "step": 891 }, { "epoch": 1.3373313343328337, "grad_norm": 0.048448533551978, "learning_rate": 5.9805222334404e-05, "loss": 0.0167, "step": 892 }, { "epoch": 1.338830584707646, "grad_norm": 0.052410701670247924, "learning_rate": 5.956564044712551e-05, "loss": 0.0207, "step": 893 }, { "epoch": 1.3403298350824588, "grad_norm": 0.042448711771357485, "learning_rate": 5.9326335692419995e-05, "loss": 0.0192, "step": 894 }, { "epoch": 1.3418290854572714, "grad_norm": 0.05234893125921864, "learning_rate": 5.908730971045542e-05, "loss": 0.0317, "step": 895 }, { "epoch": 1.343328335832084, "grad_norm": 0.04331314810299154, "learning_rate": 5.884856413948913e-05, "loss": 0.0249, "step": 896 }, { "epoch": 1.3448275862068966, "grad_norm": 0.05916456115329105, "learning_rate": 5.8610100615856524e-05, "loss": 0.0462, "step": 897 }, { "epoch": 1.3463268365817092, "grad_norm": 0.062115100364669586, "learning_rate": 5.83719207739599e-05, "loss": 0.0256, "step": 898 }, { "epoch": 1.3478260869565217, "grad_norm": 0.035880225141386626, "learning_rate": 5.8134026246257225e-05, "loss": 0.0191, "step": 899 }, { "epoch": 1.3493253373313343, "grad_norm": 0.0573517254033885, "learning_rate": 5.789641866325091e-05, "loss": 0.0262, "step": 900 }, { "epoch": 1.3508245877061469, "grad_norm": 0.060303599881880644, "learning_rate": 5.76590996534768e-05, "loss": 0.0278, "step": 901 }, { "epoch": 1.3523238380809595, "grad_norm": 0.09628521288179971, "learning_rate": 5.7422070843492734e-05, "loss": 0.0356, "step": 902 }, { "epoch": 1.353823088455772, "grad_norm": 0.03617057375661126, "learning_rate": 5.718533385786772e-05, "loss": 0.02, "step": 903 }, { "epoch": 1.3553223388305846, "grad_norm": 0.044174128599893586, "learning_rate": 5.694889031917047e-05, "loss": 0.022, "step": 904 }, { "epoch": 1.3568215892053974, "grad_norm": 0.06221884291802812, "learning_rate": 5.671274184795865e-05, "loss": 0.0233, "step": 905 }, { "epoch": 1.3583208395802098, "grad_norm": 0.0711661250860426, "learning_rate": 5.647689006276726e-05, "loss": 0.0449, "step": 906 }, { "epoch": 1.3598200899550226, "grad_norm": 0.03490331687529149, "learning_rate": 5.624133658009817e-05, "loss": 0.0192, "step": 907 }, { "epoch": 1.3613193403298351, "grad_norm": 0.06280166629494677, "learning_rate": 5.6006083014408484e-05, "loss": 0.0257, "step": 908 }, { "epoch": 1.3628185907046477, "grad_norm": 0.049297420747513795, "learning_rate": 5.577113097809989e-05, "loss": 0.0287, "step": 909 }, { "epoch": 1.3643178410794603, "grad_norm": 0.04383360423444908, "learning_rate": 5.553648208150728e-05, "loss": 0.0159, "step": 910 }, { "epoch": 1.3658170914542729, "grad_norm": 0.06660129279871392, "learning_rate": 5.53021379328879e-05, "loss": 0.0316, "step": 911 }, { "epoch": 1.3673163418290855, "grad_norm": 0.030935221991709352, "learning_rate": 5.506810013841036e-05, "loss": 0.0141, "step": 912 }, { "epoch": 1.368815592203898, "grad_norm": 0.06598665791404248, "learning_rate": 5.483437030214343e-05, "loss": 0.0238, "step": 913 }, { "epoch": 1.3703148425787106, "grad_norm": 0.05185709390744075, "learning_rate": 5.4600950026045326e-05, "loss": 0.0268, "step": 914 }, { "epoch": 1.3718140929535232, "grad_norm": 0.06112957554123978, "learning_rate": 5.436784090995239e-05, "loss": 0.0384, "step": 915 }, { "epoch": 1.3733133433283358, "grad_norm": 0.02981664047266952, "learning_rate": 5.4135044551568546e-05, "loss": 0.0179, "step": 916 }, { "epoch": 1.3748125937031483, "grad_norm": 0.03404310873892391, "learning_rate": 5.390256254645378e-05, "loss": 0.0133, "step": 917 }, { "epoch": 1.3763118440779611, "grad_norm": 0.02748593423758078, "learning_rate": 5.3670396488013854e-05, "loss": 0.01, "step": 918 }, { "epoch": 1.3778110944527735, "grad_norm": 0.044108908050431356, "learning_rate": 5.343854796748886e-05, "loss": 0.016, "step": 919 }, { "epoch": 1.3793103448275863, "grad_norm": 0.04555240266490368, "learning_rate": 5.320701857394268e-05, "loss": 0.0242, "step": 920 }, { "epoch": 1.3808095952023989, "grad_norm": 0.025073591589012855, "learning_rate": 5.297580989425177e-05, "loss": 0.0102, "step": 921 }, { "epoch": 1.3823088455772115, "grad_norm": 0.03799373887428538, "learning_rate": 5.274492351309461e-05, "loss": 0.0199, "step": 922 }, { "epoch": 1.383808095952024, "grad_norm": 0.0634147924268712, "learning_rate": 5.251436101294056e-05, "loss": 0.0261, "step": 923 }, { "epoch": 1.3853073463268366, "grad_norm": 0.05771328219018494, "learning_rate": 5.2284123974039154e-05, "loss": 0.0283, "step": 924 }, { "epoch": 1.3868065967016492, "grad_norm": 0.04619876933398743, "learning_rate": 5.205421397440935e-05, "loss": 0.0218, "step": 925 }, { "epoch": 1.3883058470764618, "grad_norm": 0.05211266829503708, "learning_rate": 5.182463258982846e-05, "loss": 0.0273, "step": 926 }, { "epoch": 1.3898050974512743, "grad_norm": 0.0405536781310804, "learning_rate": 5.159538139382167e-05, "loss": 0.0163, "step": 927 }, { "epoch": 1.391304347826087, "grad_norm": 0.05228157708952931, "learning_rate": 5.1366461957650954e-05, "loss": 0.0148, "step": 928 }, { "epoch": 1.3928035982008995, "grad_norm": 0.051865023170566474, "learning_rate": 5.113787585030454e-05, "loss": 0.0253, "step": 929 }, { "epoch": 1.394302848575712, "grad_norm": 0.08414586524418576, "learning_rate": 5.090962463848592e-05, "loss": 0.0256, "step": 930 }, { "epoch": 1.3958020989505249, "grad_norm": 0.04333393205488659, "learning_rate": 5.068170988660346e-05, "loss": 0.0155, "step": 931 }, { "epoch": 1.3973013493253372, "grad_norm": 0.033783013394866904, "learning_rate": 5.045413315675924e-05, "loss": 0.0178, "step": 932 }, { "epoch": 1.39880059970015, "grad_norm": 0.041794554720093746, "learning_rate": 5.0226896008738786e-05, "loss": 0.0134, "step": 933 }, { "epoch": 1.4002998500749624, "grad_norm": 0.04766610890656114, "learning_rate": 5.000000000000002e-05, "loss": 0.0275, "step": 934 }, { "epoch": 1.4017991004497752, "grad_norm": 0.03981235051153443, "learning_rate": 4.977344668566275e-05, "loss": 0.0171, "step": 935 }, { "epoch": 1.4032983508245878, "grad_norm": 0.050230959070964275, "learning_rate": 4.9547237618498085e-05, "loss": 0.0247, "step": 936 }, { "epoch": 1.4047976011994003, "grad_norm": 0.04796385344740444, "learning_rate": 4.9321374348917585e-05, "loss": 0.0199, "step": 937 }, { "epoch": 1.406296851574213, "grad_norm": 0.03650112855229006, "learning_rate": 4.909585842496287e-05, "loss": 0.0197, "step": 938 }, { "epoch": 1.4077961019490255, "grad_norm": 0.058389964584798164, "learning_rate": 4.887069139229481e-05, "loss": 0.0305, "step": 939 }, { "epoch": 1.409295352323838, "grad_norm": 0.03657890885550993, "learning_rate": 4.864587479418302e-05, "loss": 0.0172, "step": 940 }, { "epoch": 1.4107946026986506, "grad_norm": 0.039034241468565725, "learning_rate": 4.842141017149526e-05, "loss": 0.0289, "step": 941 }, { "epoch": 1.4122938530734632, "grad_norm": 0.05779842475140324, "learning_rate": 4.8197299062686995e-05, "loss": 0.0194, "step": 942 }, { "epoch": 1.4137931034482758, "grad_norm": 0.04423489804284638, "learning_rate": 4.79735430037906e-05, "loss": 0.0167, "step": 943 }, { "epoch": 1.4152923538230884, "grad_norm": 0.06732550480304075, "learning_rate": 4.7750143528405126e-05, "loss": 0.0234, "step": 944 }, { "epoch": 1.416791604197901, "grad_norm": 0.048358500900887474, "learning_rate": 4.752710216768551e-05, "loss": 0.0126, "step": 945 }, { "epoch": 1.4182908545727138, "grad_norm": 0.0618555856514782, "learning_rate": 4.7304420450332244e-05, "loss": 0.0531, "step": 946 }, { "epoch": 1.419790104947526, "grad_norm": 0.021844937288201197, "learning_rate": 4.708209990258095e-05, "loss": 0.0092, "step": 947 }, { "epoch": 1.421289355322339, "grad_norm": 0.043762847247643975, "learning_rate": 4.68601420481917e-05, "loss": 0.0186, "step": 948 }, { "epoch": 1.4227886056971515, "grad_norm": 0.08941262921253182, "learning_rate": 4.6638548408438856e-05, "loss": 0.033, "step": 949 }, { "epoch": 1.424287856071964, "grad_norm": 0.07137680432326722, "learning_rate": 4.6417320502100316e-05, "loss": 0.019, "step": 950 }, { "epoch": 1.4257871064467766, "grad_norm": 0.05581448919899184, "learning_rate": 4.619645984544752e-05, "loss": 0.0175, "step": 951 }, { "epoch": 1.4272863568215892, "grad_norm": 0.029522071224618904, "learning_rate": 4.59759679522345e-05, "loss": 0.0191, "step": 952 }, { "epoch": 1.4287856071964018, "grad_norm": 0.03786913455549227, "learning_rate": 4.575584633368815e-05, "loss": 0.0197, "step": 953 }, { "epoch": 1.4302848575712144, "grad_norm": 0.059428163125673714, "learning_rate": 4.5536096498497295e-05, "loss": 0.0296, "step": 954 }, { "epoch": 1.431784107946027, "grad_norm": 0.03276712389103159, "learning_rate": 4.5316719952802776e-05, "loss": 0.0206, "step": 955 }, { "epoch": 1.4332833583208395, "grad_norm": 0.0411950835656051, "learning_rate": 4.5097718200186814e-05, "loss": 0.0147, "step": 956 }, { "epoch": 1.434782608695652, "grad_norm": 0.03726680229656808, "learning_rate": 4.4879092741663e-05, "loss": 0.0135, "step": 957 }, { "epoch": 1.4362818590704647, "grad_norm": 0.051339763257998224, "learning_rate": 4.46608450756656e-05, "loss": 0.0294, "step": 958 }, { "epoch": 1.4377811094452775, "grad_norm": 0.06516988257924636, "learning_rate": 4.444297669803981e-05, "loss": 0.0377, "step": 959 }, { "epoch": 1.4392803598200898, "grad_norm": 0.04522777428706767, "learning_rate": 4.4225489102030995e-05, "loss": 0.0164, "step": 960 }, { "epoch": 1.4407796101949026, "grad_norm": 0.057957991417301825, "learning_rate": 4.400838377827483e-05, "loss": 0.0292, "step": 961 }, { "epoch": 1.4422788605697152, "grad_norm": 0.04382325188747058, "learning_rate": 4.379166221478697e-05, "loss": 0.013, "step": 962 }, { "epoch": 1.4437781109445278, "grad_norm": 0.045850857607972945, "learning_rate": 4.35753258969526e-05, "loss": 0.0292, "step": 963 }, { "epoch": 1.4452773613193404, "grad_norm": 0.052319076851113866, "learning_rate": 4.335937630751674e-05, "loss": 0.0203, "step": 964 }, { "epoch": 1.446776611694153, "grad_norm": 0.04430763147098959, "learning_rate": 4.31438149265736e-05, "loss": 0.0174, "step": 965 }, { "epoch": 1.4482758620689655, "grad_norm": 0.037336734388071224, "learning_rate": 4.2928643231556844e-05, "loss": 0.0187, "step": 966 }, { "epoch": 1.449775112443778, "grad_norm": 0.03488717623172854, "learning_rate": 4.271386269722909e-05, "loss": 0.0215, "step": 967 }, { "epoch": 1.4512743628185907, "grad_norm": 0.11146756348958964, "learning_rate": 4.249947479567218e-05, "loss": 0.0632, "step": 968 }, { "epoch": 1.4527736131934033, "grad_norm": 0.0723346597125285, "learning_rate": 4.228548099627665e-05, "loss": 0.022, "step": 969 }, { "epoch": 1.4542728635682158, "grad_norm": 0.05077819533323603, "learning_rate": 4.207188276573214e-05, "loss": 0.0262, "step": 970 }, { "epoch": 1.4557721139430284, "grad_norm": 0.04220137883748374, "learning_rate": 4.185868156801694e-05, "loss": 0.022, "step": 971 }, { "epoch": 1.4572713643178412, "grad_norm": 0.04257073910089388, "learning_rate": 4.1645878864388266e-05, "loss": 0.0188, "step": 972 }, { "epoch": 1.4587706146926536, "grad_norm": 0.07051244032099359, "learning_rate": 4.143347611337194e-05, "loss": 0.0229, "step": 973 }, { "epoch": 1.4602698650674664, "grad_norm": 0.04924278467634398, "learning_rate": 4.12214747707527e-05, "loss": 0.02, "step": 974 }, { "epoch": 1.461769115442279, "grad_norm": 0.04465214846691316, "learning_rate": 4.1009876289563976e-05, "loss": 0.0204, "step": 975 }, { "epoch": 1.4632683658170915, "grad_norm": 0.04262641492883479, "learning_rate": 4.0798682120078044e-05, "loss": 0.0192, "step": 976 }, { "epoch": 1.464767616191904, "grad_norm": 0.03647273601340537, "learning_rate": 4.058789370979615e-05, "loss": 0.0184, "step": 977 }, { "epoch": 1.4662668665667167, "grad_norm": 0.04885135157810087, "learning_rate": 4.037751250343841e-05, "loss": 0.0303, "step": 978 }, { "epoch": 1.4677661169415293, "grad_norm": 0.08387069269883196, "learning_rate": 4.01675399429341e-05, "loss": 0.0234, "step": 979 }, { "epoch": 1.4692653673163418, "grad_norm": 0.038204055040209454, "learning_rate": 3.9957977467411615e-05, "loss": 0.0179, "step": 980 }, { "epoch": 1.4707646176911544, "grad_norm": 0.03786765301339409, "learning_rate": 3.9748826513188686e-05, "loss": 0.0165, "step": 981 }, { "epoch": 1.472263868065967, "grad_norm": 0.05231654154725003, "learning_rate": 3.954008851376252e-05, "loss": 0.0172, "step": 982 }, { "epoch": 1.4737631184407796, "grad_norm": 0.03335030350491901, "learning_rate": 3.933176489980005e-05, "loss": 0.0106, "step": 983 }, { "epoch": 1.4752623688155921, "grad_norm": 0.03979273183143895, "learning_rate": 3.9123857099127936e-05, "loss": 0.0212, "step": 984 }, { "epoch": 1.4767616191904047, "grad_norm": 0.05391309807180965, "learning_rate": 3.891636653672302e-05, "loss": 0.0157, "step": 985 }, { "epoch": 1.4782608695652173, "grad_norm": 0.03361291341690922, "learning_rate": 3.8709294634702376e-05, "loss": 0.0139, "step": 986 }, { "epoch": 1.47976011994003, "grad_norm": 0.035585947784928226, "learning_rate": 3.850264281231357e-05, "loss": 0.0219, "step": 987 }, { "epoch": 1.4812593703148424, "grad_norm": 0.03333439340678102, "learning_rate": 3.829641248592515e-05, "loss": 0.0129, "step": 988 }, { "epoch": 1.4827586206896552, "grad_norm": 0.040345561974199404, "learning_rate": 3.8090605069016595e-05, "loss": 0.0234, "step": 989 }, { "epoch": 1.4842578710644678, "grad_norm": 0.07403742521938174, "learning_rate": 3.788522197216897e-05, "loss": 0.02, "step": 990 }, { "epoch": 1.4857571214392804, "grad_norm": 0.05711176669065304, "learning_rate": 3.7680264603054994e-05, "loss": 0.0397, "step": 991 }, { "epoch": 1.487256371814093, "grad_norm": 0.06610246075931671, "learning_rate": 3.747573436642951e-05, "loss": 0.0277, "step": 992 }, { "epoch": 1.4887556221889056, "grad_norm": 0.03232609404738584, "learning_rate": 3.7271632664119846e-05, "loss": 0.0091, "step": 993 }, { "epoch": 1.4902548725637181, "grad_norm": 0.03787459790735476, "learning_rate": 3.7067960895016275e-05, "loss": 0.0209, "step": 994 }, { "epoch": 1.4917541229385307, "grad_norm": 0.04786121404829169, "learning_rate": 3.686472045506223e-05, "loss": 0.013, "step": 995 }, { "epoch": 1.4932533733133433, "grad_norm": 0.04350531988383918, "learning_rate": 3.6661912737245e-05, "loss": 0.0292, "step": 996 }, { "epoch": 1.4947526236881559, "grad_norm": 0.03931742385537951, "learning_rate": 3.645953913158593e-05, "loss": 0.027, "step": 997 }, { "epoch": 1.4962518740629684, "grad_norm": 0.041467982689201134, "learning_rate": 3.6257601025131026e-05, "loss": 0.0238, "step": 998 }, { "epoch": 1.497751124437781, "grad_norm": 0.04379142918036115, "learning_rate": 3.6056099801941534e-05, "loss": 0.0192, "step": 999 }, { "epoch": 1.4992503748125938, "grad_norm": 0.03952572570808585, "learning_rate": 3.585503684308421e-05, "loss": 0.0185, "step": 1000 }, { "epoch": 1.5007496251874062, "grad_norm": 0.04682486078750102, "learning_rate": 3.565441352662211e-05, "loss": 0.0172, "step": 1001 }, { "epoch": 1.502248875562219, "grad_norm": 0.05532999211305211, "learning_rate": 3.545423122760493e-05, "loss": 0.0277, "step": 1002 }, { "epoch": 1.5037481259370313, "grad_norm": 0.06235019222714452, "learning_rate": 3.525449131805982e-05, "loss": 0.0319, "step": 1003 }, { "epoch": 1.5052473763118441, "grad_norm": 0.06024901707215258, "learning_rate": 3.5055195166981645e-05, "loss": 0.0375, "step": 1004 }, { "epoch": 1.5067466266866567, "grad_norm": 0.0678387358265547, "learning_rate": 3.4856344140323985e-05, "loss": 0.0281, "step": 1005 }, { "epoch": 1.5082458770614693, "grad_norm": 0.054204263066234194, "learning_rate": 3.465793960098945e-05, "loss": 0.0434, "step": 1006 }, { "epoch": 1.5097451274362819, "grad_norm": 0.045371529886557356, "learning_rate": 3.445998290882062e-05, "loss": 0.0225, "step": 1007 }, { "epoch": 1.5112443778110944, "grad_norm": 0.12719949916957105, "learning_rate": 3.426247542059041e-05, "loss": 0.0482, "step": 1008 }, { "epoch": 1.512743628185907, "grad_norm": 0.039729045535986905, "learning_rate": 3.406541848999312e-05, "loss": 0.0135, "step": 1009 }, { "epoch": 1.5142428785607196, "grad_norm": 0.07776064383164068, "learning_rate": 3.386881346763483e-05, "loss": 0.0344, "step": 1010 }, { "epoch": 1.5157421289355324, "grad_norm": 0.031920259092681355, "learning_rate": 3.367266170102432e-05, "loss": 0.0123, "step": 1011 }, { "epoch": 1.5172413793103448, "grad_norm": 0.03993466785133325, "learning_rate": 3.347696453456393e-05, "loss": 0.0195, "step": 1012 }, { "epoch": 1.5187406296851576, "grad_norm": 0.03649095428535925, "learning_rate": 3.328172330954001e-05, "loss": 0.0127, "step": 1013 }, { "epoch": 1.52023988005997, "grad_norm": 0.05184896776912534, "learning_rate": 3.308693936411421e-05, "loss": 0.0389, "step": 1014 }, { "epoch": 1.5217391304347827, "grad_norm": 0.041105616161891526, "learning_rate": 3.289261403331372e-05, "loss": 0.0273, "step": 1015 }, { "epoch": 1.523238380809595, "grad_norm": 0.05332829161038912, "learning_rate": 3.269874864902269e-05, "loss": 0.0312, "step": 1016 }, { "epoch": 1.5247376311844079, "grad_norm": 0.029829522522220115, "learning_rate": 3.2505344539972705e-05, "loss": 0.0122, "step": 1017 }, { "epoch": 1.5262368815592204, "grad_norm": 0.05136530388164031, "learning_rate": 3.231240303173394e-05, "loss": 0.0128, "step": 1018 }, { "epoch": 1.527736131934033, "grad_norm": 0.026747951417845865, "learning_rate": 3.211992544670582e-05, "loss": 0.0139, "step": 1019 }, { "epoch": 1.5292353823088456, "grad_norm": 0.043634723111218424, "learning_rate": 3.192791310410822e-05, "loss": 0.0145, "step": 1020 }, { "epoch": 1.5307346326836582, "grad_norm": 0.032266616085346804, "learning_rate": 3.173636731997222e-05, "loss": 0.0136, "step": 1021 }, { "epoch": 1.5322338830584707, "grad_norm": 0.05249669591805282, "learning_rate": 3.154528940713113e-05, "loss": 0.0216, "step": 1022 }, { "epoch": 1.5337331334332833, "grad_norm": 0.04769500104005904, "learning_rate": 3.1354680675211635e-05, "loss": 0.0165, "step": 1023 }, { "epoch": 1.535232383808096, "grad_norm": 0.035229159668666236, "learning_rate": 3.116454243062459e-05, "loss": 0.0194, "step": 1024 }, { "epoch": 1.5367316341829085, "grad_norm": 0.05407112124934083, "learning_rate": 3.0974875976556284e-05, "loss": 0.0236, "step": 1025 }, { "epoch": 1.5382308845577213, "grad_norm": 0.025074397773416118, "learning_rate": 3.078568261295933e-05, "loss": 0.0141, "step": 1026 }, { "epoch": 1.5397301349325336, "grad_norm": 0.0445437291021861, "learning_rate": 3.059696363654386e-05, "loss": 0.0186, "step": 1027 }, { "epoch": 1.5412293853073464, "grad_norm": 0.0441384879409754, "learning_rate": 3.0408720340768572e-05, "loss": 0.0139, "step": 1028 }, { "epoch": 1.5427286356821588, "grad_norm": 0.0479088329800761, "learning_rate": 3.0220954015832003e-05, "loss": 0.0302, "step": 1029 }, { "epoch": 1.5442278860569716, "grad_norm": 0.034357781660932715, "learning_rate": 3.0033665948663448e-05, "loss": 0.0179, "step": 1030 }, { "epoch": 1.545727136431784, "grad_norm": 0.04351013918478569, "learning_rate": 2.9846857422914433e-05, "loss": 0.0237, "step": 1031 }, { "epoch": 1.5472263868065967, "grad_norm": 0.052463537989510735, "learning_rate": 2.9660529718949627e-05, "loss": 0.0264, "step": 1032 }, { "epoch": 1.5487256371814093, "grad_norm": 0.03561598179282845, "learning_rate": 2.9474684113838257e-05, "loss": 0.0199, "step": 1033 }, { "epoch": 1.550224887556222, "grad_norm": 0.03365987985086509, "learning_rate": 2.9289321881345254e-05, "loss": 0.0122, "step": 1034 }, { "epoch": 1.5517241379310345, "grad_norm": 0.043573488595870076, "learning_rate": 2.910444429192265e-05, "loss": 0.0203, "step": 1035 }, { "epoch": 1.553223388305847, "grad_norm": 0.03592975923317004, "learning_rate": 2.8920052612700754e-05, "loss": 0.0148, "step": 1036 }, { "epoch": 1.5547226386806596, "grad_norm": 0.03740369119589898, "learning_rate": 2.8736148107479467e-05, "loss": 0.0122, "step": 1037 }, { "epoch": 1.5562218890554722, "grad_norm": 0.05866623225512895, "learning_rate": 2.8552732036719687e-05, "loss": 0.026, "step": 1038 }, { "epoch": 1.557721139430285, "grad_norm": 0.037514551907470446, "learning_rate": 2.8369805657534575e-05, "loss": 0.0101, "step": 1039 }, { "epoch": 1.5592203898050974, "grad_norm": 0.06919743559839049, "learning_rate": 2.8187370223681132e-05, "loss": 0.0186, "step": 1040 }, { "epoch": 1.5607196401799102, "grad_norm": 0.10860031837919593, "learning_rate": 2.800542698555132e-05, "loss": 0.0331, "step": 1041 }, { "epoch": 1.5622188905547225, "grad_norm": 0.04053934454593709, "learning_rate": 2.7823977190163786e-05, "loss": 0.0304, "step": 1042 }, { "epoch": 1.5637181409295353, "grad_norm": 0.04589812353371874, "learning_rate": 2.764302208115508e-05, "loss": 0.0288, "step": 1043 }, { "epoch": 1.5652173913043477, "grad_norm": 0.051759071563514204, "learning_rate": 2.746256289877126e-05, "loss": 0.0221, "step": 1044 }, { "epoch": 1.5667166416791605, "grad_norm": 0.04419463652932288, "learning_rate": 2.728260087985933e-05, "loss": 0.021, "step": 1045 }, { "epoch": 1.568215892053973, "grad_norm": 0.044966668706100936, "learning_rate": 2.7103137257858868e-05, "loss": 0.0337, "step": 1046 }, { "epoch": 1.5697151424287856, "grad_norm": 0.04302495760541486, "learning_rate": 2.6924173262793383e-05, "loss": 0.0137, "step": 1047 }, { "epoch": 1.5712143928035982, "grad_norm": 0.08218579196147292, "learning_rate": 2.6745710121262136e-05, "loss": 0.0365, "step": 1048 }, { "epoch": 1.5727136431784108, "grad_norm": 0.061167493252746986, "learning_rate": 2.6567749056431467e-05, "loss": 0.0258, "step": 1049 }, { "epoch": 1.5742128935532234, "grad_norm": 0.13364423709623616, "learning_rate": 2.639029128802657e-05, "loss": 0.0626, "step": 1050 }, { "epoch": 1.575712143928036, "grad_norm": 0.03214675852865434, "learning_rate": 2.6213338032323176e-05, "loss": 0.0133, "step": 1051 }, { "epoch": 1.5772113943028487, "grad_norm": 0.06328373661244126, "learning_rate": 2.603689050213902e-05, "loss": 0.0243, "step": 1052 }, { "epoch": 1.578710644677661, "grad_norm": 0.04742194356052198, "learning_rate": 2.5860949906825795e-05, "loss": 0.0285, "step": 1053 }, { "epoch": 1.580209895052474, "grad_norm": 0.048588116559285674, "learning_rate": 2.5685517452260567e-05, "loss": 0.0307, "step": 1054 }, { "epoch": 1.5817091454272862, "grad_norm": 0.04993527437674772, "learning_rate": 2.5510594340837824e-05, "loss": 0.0232, "step": 1055 }, { "epoch": 1.583208395802099, "grad_norm": 0.043690403745079286, "learning_rate": 2.5336181771460876e-05, "loss": 0.0205, "step": 1056 }, { "epoch": 1.5847076461769114, "grad_norm": 0.07438950423299034, "learning_rate": 2.516228093953402e-05, "loss": 0.0306, "step": 1057 }, { "epoch": 1.5862068965517242, "grad_norm": 0.03036817361902045, "learning_rate": 2.4988893036954043e-05, "loss": 0.0148, "step": 1058 }, { "epoch": 1.5877061469265368, "grad_norm": 0.0330965964673103, "learning_rate": 2.4816019252102273e-05, "loss": 0.0195, "step": 1059 }, { "epoch": 1.5892053973013494, "grad_norm": 0.035050676413594786, "learning_rate": 2.464366076983623e-05, "loss": 0.0136, "step": 1060 }, { "epoch": 1.590704647676162, "grad_norm": 0.04101060544019866, "learning_rate": 2.4471818771481648e-05, "loss": 0.0141, "step": 1061 }, { "epoch": 1.5922038980509745, "grad_norm": 0.06280027292019925, "learning_rate": 2.4300494434824373e-05, "loss": 0.0224, "step": 1062 }, { "epoch": 1.593703148425787, "grad_norm": 0.04131360603164168, "learning_rate": 2.412968893410218e-05, "loss": 0.017, "step": 1063 }, { "epoch": 1.5952023988005997, "grad_norm": 0.03827142257211483, "learning_rate": 2.3959403439996907e-05, "loss": 0.0221, "step": 1064 }, { "epoch": 1.5967016491754122, "grad_norm": 0.06504367581713787, "learning_rate": 2.3789639119626216e-05, "loss": 0.0267, "step": 1065 }, { "epoch": 1.5982008995502248, "grad_norm": 0.06818969255684648, "learning_rate": 2.362039713653581e-05, "loss": 0.0228, "step": 1066 }, { "epoch": 1.5997001499250376, "grad_norm": 0.038392151336867984, "learning_rate": 2.34516786506912e-05, "loss": 0.0154, "step": 1067 }, { "epoch": 1.60119940029985, "grad_norm": 0.06573629069388562, "learning_rate": 2.328348481847006e-05, "loss": 0.0251, "step": 1068 }, { "epoch": 1.6026986506746628, "grad_norm": 0.06148804102636816, "learning_rate": 2.3115816792654056e-05, "loss": 0.0311, "step": 1069 }, { "epoch": 1.6041979010494751, "grad_norm": 0.04471759363710799, "learning_rate": 2.2948675722421086e-05, "loss": 0.0191, "step": 1070 }, { "epoch": 1.605697151424288, "grad_norm": 0.0415533636509639, "learning_rate": 2.2782062753337306e-05, "loss": 0.0184, "step": 1071 }, { "epoch": 1.6071964017991005, "grad_norm": 0.05794733183745205, "learning_rate": 2.2615979027349387e-05, "loss": 0.0255, "step": 1072 }, { "epoch": 1.608695652173913, "grad_norm": 0.0397678166261267, "learning_rate": 2.2450425682776565e-05, "loss": 0.0218, "step": 1073 }, { "epoch": 1.6101949025487257, "grad_norm": 0.04761970893285829, "learning_rate": 2.2285403854302912e-05, "loss": 0.0195, "step": 1074 }, { "epoch": 1.6116941529235382, "grad_norm": 0.054708750805224075, "learning_rate": 2.2120914672969606e-05, "loss": 0.0207, "step": 1075 }, { "epoch": 1.6131934032983508, "grad_norm": 0.04536721690981569, "learning_rate": 2.195695926616702e-05, "loss": 0.0115, "step": 1076 }, { "epoch": 1.6146926536731634, "grad_norm": 0.03693822651785582, "learning_rate": 2.1793538757627218e-05, "loss": 0.0269, "step": 1077 }, { "epoch": 1.616191904047976, "grad_norm": 0.05430171107827179, "learning_rate": 2.163065426741603e-05, "loss": 0.0234, "step": 1078 }, { "epoch": 1.6176911544227885, "grad_norm": 0.03499504155734339, "learning_rate": 2.146830691192553e-05, "loss": 0.0215, "step": 1079 }, { "epoch": 1.6191904047976013, "grad_norm": 0.041779929626990946, "learning_rate": 2.1306497803866277e-05, "loss": 0.0219, "step": 1080 }, { "epoch": 1.6206896551724137, "grad_norm": 0.05662420578938302, "learning_rate": 2.1145228052259824e-05, "loss": 0.0266, "step": 1081 }, { "epoch": 1.6221889055472265, "grad_norm": 0.0632358405041289, "learning_rate": 2.098449876243096e-05, "loss": 0.0513, "step": 1082 }, { "epoch": 1.6236881559220389, "grad_norm": 0.03370533424123817, "learning_rate": 2.0824311036000276e-05, "loss": 0.0164, "step": 1083 }, { "epoch": 1.6251874062968517, "grad_norm": 0.03757323666893171, "learning_rate": 2.0664665970876496e-05, "loss": 0.0158, "step": 1084 }, { "epoch": 1.626686656671664, "grad_norm": 0.03313883661650599, "learning_rate": 2.0505564661249e-05, "loss": 0.0164, "step": 1085 }, { "epoch": 1.6281859070464768, "grad_norm": 0.04129568741796651, "learning_rate": 2.0347008197580374e-05, "loss": 0.0262, "step": 1086 }, { "epoch": 1.6296851574212894, "grad_norm": 0.050633574006442895, "learning_rate": 2.0188997666598854e-05, "loss": 0.0263, "step": 1087 }, { "epoch": 1.631184407796102, "grad_norm": 0.062148357793015474, "learning_rate": 2.0031534151290943e-05, "loss": 0.0322, "step": 1088 }, { "epoch": 1.6326836581709145, "grad_norm": 0.045969526657587925, "learning_rate": 1.9874618730893946e-05, "loss": 0.0168, "step": 1089 }, { "epoch": 1.6341829085457271, "grad_norm": 0.04612858174233097, "learning_rate": 1.9718252480888566e-05, "loss": 0.0286, "step": 1090 }, { "epoch": 1.6356821589205397, "grad_norm": 0.040575925623268116, "learning_rate": 1.9562436472991552e-05, "loss": 0.019, "step": 1091 }, { "epoch": 1.6371814092953523, "grad_norm": 0.04417146500100262, "learning_rate": 1.9407171775148436e-05, "loss": 0.0202, "step": 1092 }, { "epoch": 1.638680659670165, "grad_norm": 0.036770078126215765, "learning_rate": 1.9252459451526007e-05, "loss": 0.0129, "step": 1093 }, { "epoch": 1.6401799100449774, "grad_norm": 0.0364801904957448, "learning_rate": 1.9098300562505266e-05, "loss": 0.015, "step": 1094 }, { "epoch": 1.6416791604197902, "grad_norm": 0.0254972319896911, "learning_rate": 1.8944696164673946e-05, "loss": 0.0139, "step": 1095 }, { "epoch": 1.6431784107946026, "grad_norm": 0.047093621027718184, "learning_rate": 1.879164731081937e-05, "loss": 0.0252, "step": 1096 }, { "epoch": 1.6446776611694154, "grad_norm": 0.04485052170740798, "learning_rate": 1.863915504992131e-05, "loss": 0.0202, "step": 1097 }, { "epoch": 1.6461769115442277, "grad_norm": 0.0387698802952675, "learning_rate": 1.848722042714457e-05, "loss": 0.0117, "step": 1098 }, { "epoch": 1.6476761619190405, "grad_norm": 0.0690313351636082, "learning_rate": 1.833584448383211e-05, "loss": 0.0225, "step": 1099 }, { "epoch": 1.6491754122938531, "grad_norm": 0.041762472042913064, "learning_rate": 1.818502825749764e-05, "loss": 0.0189, "step": 1100 }, { "epoch": 1.6506746626686657, "grad_norm": 0.043282701400431646, "learning_rate": 1.8034772781818775e-05, "loss": 0.0352, "step": 1101 }, { "epoch": 1.6521739130434783, "grad_norm": 0.05165034058427283, "learning_rate": 1.78850790866296e-05, "loss": 0.0225, "step": 1102 }, { "epoch": 1.6536731634182908, "grad_norm": 0.02717101883001689, "learning_rate": 1.773594819791403e-05, "loss": 0.0193, "step": 1103 }, { "epoch": 1.6551724137931034, "grad_norm": 0.04662587755796276, "learning_rate": 1.7587381137798432e-05, "loss": 0.0178, "step": 1104 }, { "epoch": 1.656671664167916, "grad_norm": 0.031822487513949387, "learning_rate": 1.7439378924544835e-05, "loss": 0.0189, "step": 1105 }, { "epoch": 1.6581709145427288, "grad_norm": 0.03280026362696025, "learning_rate": 1.7291942572543807e-05, "loss": 0.0123, "step": 1106 }, { "epoch": 1.6596701649175412, "grad_norm": 0.030180458183959057, "learning_rate": 1.714507309230766e-05, "loss": 0.0141, "step": 1107 }, { "epoch": 1.661169415292354, "grad_norm": 0.05658463863566536, "learning_rate": 1.6998771490463262e-05, "loss": 0.0647, "step": 1108 }, { "epoch": 1.6626686656671663, "grad_norm": 0.040781014716140526, "learning_rate": 1.6853038769745467e-05, "loss": 0.016, "step": 1109 }, { "epoch": 1.6641679160419791, "grad_norm": 0.04738429854592745, "learning_rate": 1.6707875928990058e-05, "loss": 0.03, "step": 1110 }, { "epoch": 1.6656671664167915, "grad_norm": 0.025248986437211247, "learning_rate": 1.656328396312682e-05, "loss": 0.0119, "step": 1111 }, { "epoch": 1.6671664167916043, "grad_norm": 0.08248141981476365, "learning_rate": 1.6419263863172997e-05, "loss": 0.0401, "step": 1112 }, { "epoch": 1.6686656671664168, "grad_norm": 0.040376633264695476, "learning_rate": 1.6275816616226113e-05, "loss": 0.0162, "step": 1113 }, { "epoch": 1.6701649175412294, "grad_norm": 0.04326524362791187, "learning_rate": 1.6132943205457606e-05, "loss": 0.0174, "step": 1114 }, { "epoch": 1.671664167916042, "grad_norm": 0.04807125621491688, "learning_rate": 1.599064461010581e-05, "loss": 0.0324, "step": 1115 }, { "epoch": 1.6731634182908546, "grad_norm": 0.03729442007384064, "learning_rate": 1.5848921805469397e-05, "loss": 0.0211, "step": 1116 }, { "epoch": 1.6746626686656672, "grad_norm": 0.029001614577088476, "learning_rate": 1.5707775762900547e-05, "loss": 0.0129, "step": 1117 }, { "epoch": 1.6761619190404797, "grad_norm": 0.030669347972443403, "learning_rate": 1.5567207449798515e-05, "loss": 0.01, "step": 1118 }, { "epoch": 1.6776611694152923, "grad_norm": 0.04829682712396594, "learning_rate": 1.542721782960268e-05, "loss": 0.0232, "step": 1119 }, { "epoch": 1.6791604197901049, "grad_norm": 0.04806923915457913, "learning_rate": 1.528780786178631e-05, "loss": 0.0202, "step": 1120 }, { "epoch": 1.6806596701649177, "grad_norm": 0.02440255176031142, "learning_rate": 1.5148978501849642e-05, "loss": 0.0128, "step": 1121 }, { "epoch": 1.68215892053973, "grad_norm": 0.04063406158834547, "learning_rate": 1.5010730701313625e-05, "loss": 0.0171, "step": 1122 }, { "epoch": 1.6836581709145428, "grad_norm": 0.04035147945576393, "learning_rate": 1.4873065407713149e-05, "loss": 0.031, "step": 1123 }, { "epoch": 1.6851574212893552, "grad_norm": 0.04027060084795129, "learning_rate": 1.4735983564590783e-05, "loss": 0.0182, "step": 1124 }, { "epoch": 1.686656671664168, "grad_norm": 0.045280521178547195, "learning_rate": 1.4599486111490102e-05, "loss": 0.0198, "step": 1125 }, { "epoch": 1.6881559220389803, "grad_norm": 0.04881923000165193, "learning_rate": 1.4463573983949341e-05, "loss": 0.0167, "step": 1126 }, { "epoch": 1.6896551724137931, "grad_norm": 0.03529851016922019, "learning_rate": 1.4328248113495047e-05, "loss": 0.0165, "step": 1127 }, { "epoch": 1.6911544227886057, "grad_norm": 0.04376517979604228, "learning_rate": 1.4193509427635543e-05, "loss": 0.0266, "step": 1128 }, { "epoch": 1.6926536731634183, "grad_norm": 0.03839629594972581, "learning_rate": 1.405935884985473e-05, "loss": 0.0109, "step": 1129 }, { "epoch": 1.6941529235382309, "grad_norm": 0.05394784900443594, "learning_rate": 1.3925797299605647e-05, "loss": 0.0289, "step": 1130 }, { "epoch": 1.6956521739130435, "grad_norm": 0.04473750186048263, "learning_rate": 1.379282569230419e-05, "loss": 0.0183, "step": 1131 }, { "epoch": 1.697151424287856, "grad_norm": 0.03971094977709959, "learning_rate": 1.3660444939322836e-05, "loss": 0.0131, "step": 1132 }, { "epoch": 1.6986506746626686, "grad_norm": 0.044781528569248615, "learning_rate": 1.3528655947984504e-05, "loss": 0.0606, "step": 1133 }, { "epoch": 1.7001499250374814, "grad_norm": 0.05687647115833445, "learning_rate": 1.339745962155613e-05, "loss": 0.0261, "step": 1134 }, { "epoch": 1.7016491754122938, "grad_norm": 0.051596153458091924, "learning_rate": 1.3266856859242705e-05, "loss": 0.0217, "step": 1135 }, { "epoch": 1.7031484257871066, "grad_norm": 0.05150733833063215, "learning_rate": 1.3136848556180892e-05, "loss": 0.0306, "step": 1136 }, { "epoch": 1.704647676161919, "grad_norm": 0.042545970063710185, "learning_rate": 1.3007435603433039e-05, "loss": 0.0195, "step": 1137 }, { "epoch": 1.7061469265367317, "grad_norm": 0.044603161701574794, "learning_rate": 1.2878618887981064e-05, "loss": 0.0241, "step": 1138 }, { "epoch": 1.707646176911544, "grad_norm": 0.05764350301977242, "learning_rate": 1.2750399292720283e-05, "loss": 0.0216, "step": 1139 }, { "epoch": 1.7091454272863569, "grad_norm": 0.05335928285880417, "learning_rate": 1.262277769645348e-05, "loss": 0.0375, "step": 1140 }, { "epoch": 1.7106446776611695, "grad_norm": 0.036542466613040704, "learning_rate": 1.2495754973884766e-05, "loss": 0.0217, "step": 1141 }, { "epoch": 1.712143928035982, "grad_norm": 0.0317543900475218, "learning_rate": 1.2369331995613665e-05, "loss": 0.0089, "step": 1142 }, { "epoch": 1.7136431784107946, "grad_norm": 0.043953143850443466, "learning_rate": 1.224350962812908e-05, "loss": 0.0221, "step": 1143 }, { "epoch": 1.7151424287856072, "grad_norm": 0.04475451692804134, "learning_rate": 1.2118288733803473e-05, "loss": 0.0269, "step": 1144 }, { "epoch": 1.7166416791604198, "grad_norm": 0.0468528264899152, "learning_rate": 1.1993670170886806e-05, "loss": 0.0251, "step": 1145 }, { "epoch": 1.7181409295352323, "grad_norm": 0.04567111932909807, "learning_rate": 1.1869654793500784e-05, "loss": 0.0256, "step": 1146 }, { "epoch": 1.7196401799100451, "grad_norm": 0.04665397165143054, "learning_rate": 1.17462434516329e-05, "loss": 0.0287, "step": 1147 }, { "epoch": 1.7211394302848575, "grad_norm": 0.043566933804619214, "learning_rate": 1.1623436991130654e-05, "loss": 0.0193, "step": 1148 }, { "epoch": 1.7226386806596703, "grad_norm": 0.03910118483915222, "learning_rate": 1.1501236253695823e-05, "loss": 0.019, "step": 1149 }, { "epoch": 1.7241379310344827, "grad_norm": 0.042811788583233146, "learning_rate": 1.1379642076878527e-05, "loss": 0.0192, "step": 1150 }, { "epoch": 1.7256371814092955, "grad_norm": 0.03259518496802405, "learning_rate": 1.1258655294071685e-05, "loss": 0.0114, "step": 1151 }, { "epoch": 1.7271364317841078, "grad_norm": 0.03610690907310535, "learning_rate": 1.1138276734505104e-05, "loss": 0.0088, "step": 1152 }, { "epoch": 1.7286356821589206, "grad_norm": 0.058981199182703795, "learning_rate": 1.1018507223240038e-05, "loss": 0.0303, "step": 1153 }, { "epoch": 1.7301349325337332, "grad_norm": 0.04288242904489575, "learning_rate": 1.0899347581163221e-05, "loss": 0.0316, "step": 1154 }, { "epoch": 1.7316341829085458, "grad_norm": 0.0322212652007745, "learning_rate": 1.0780798624981547e-05, "loss": 0.0127, "step": 1155 }, { "epoch": 1.7331334332833583, "grad_norm": 0.04036547160797265, "learning_rate": 1.0662861167216243e-05, "loss": 0.0268, "step": 1156 }, { "epoch": 1.734632683658171, "grad_norm": 0.027977601513989023, "learning_rate": 1.0545536016197499e-05, "loss": 0.0106, "step": 1157 }, { "epoch": 1.7361319340329835, "grad_norm": 0.04130410258519725, "learning_rate": 1.042882397605871e-05, "loss": 0.0164, "step": 1158 }, { "epoch": 1.737631184407796, "grad_norm": 0.09110404984904212, "learning_rate": 1.0312725846731175e-05, "loss": 0.032, "step": 1159 }, { "epoch": 1.7391304347826086, "grad_norm": 0.04675587064403581, "learning_rate": 1.0197242423938446e-05, "loss": 0.0112, "step": 1160 }, { "epoch": 1.7406296851574212, "grad_norm": 0.05310767217545907, "learning_rate": 1.008237449919096e-05, "loss": 0.0208, "step": 1161 }, { "epoch": 1.742128935532234, "grad_norm": 0.040653637020288395, "learning_rate": 9.968122859780648e-06, "loss": 0.0206, "step": 1162 }, { "epoch": 1.7436281859070464, "grad_norm": 0.04403126184179531, "learning_rate": 9.854488288775422e-06, "loss": 0.0151, "step": 1163 }, { "epoch": 1.7451274362818592, "grad_norm": 0.030076096749230896, "learning_rate": 9.74147156501396e-06, "loss": 0.0113, "step": 1164 }, { "epoch": 1.7466266866566715, "grad_norm": 0.029651097446938043, "learning_rate": 9.629073463100146e-06, "loss": 0.0082, "step": 1165 }, { "epoch": 1.7481259370314843, "grad_norm": 0.03931051532613931, "learning_rate": 9.517294753398064e-06, "loss": 0.0206, "step": 1166 }, { "epoch": 1.7496251874062967, "grad_norm": 0.04890773527042702, "learning_rate": 9.406136202026416e-06, "loss": 0.0186, "step": 1167 }, { "epoch": 1.7511244377811095, "grad_norm": 0.05681064457174664, "learning_rate": 9.295598570853514e-06, "loss": 0.0241, "step": 1168 }, { "epoch": 1.752623688155922, "grad_norm": 0.033925497614483074, "learning_rate": 9.185682617491863e-06, "loss": 0.0169, "step": 1169 }, { "epoch": 1.7541229385307346, "grad_norm": 0.050500607910895993, "learning_rate": 9.076389095293148e-06, "loss": 0.03, "step": 1170 }, { "epoch": 1.7556221889055472, "grad_norm": 0.03528769870947365, "learning_rate": 8.967718753342902e-06, "loss": 0.015, "step": 1171 }, { "epoch": 1.7571214392803598, "grad_norm": 0.03966138334448402, "learning_rate": 8.85967233645547e-06, "loss": 0.0151, "step": 1172 }, { "epoch": 1.7586206896551724, "grad_norm": 0.05472974585600949, "learning_rate": 8.752250585168931e-06, "loss": 0.0272, "step": 1173 }, { "epoch": 1.760119940029985, "grad_norm": 0.04658097856564392, "learning_rate": 8.645454235739903e-06, "loss": 0.0273, "step": 1174 }, { "epoch": 1.7616191904047978, "grad_norm": 0.04039569121015692, "learning_rate": 8.539284020138638e-06, "loss": 0.0137, "step": 1175 }, { "epoch": 1.76311844077961, "grad_norm": 0.04821730129601109, "learning_rate": 8.433740666043898e-06, "loss": 0.034, "step": 1176 }, { "epoch": 1.764617691154423, "grad_norm": 0.04511525491973174, "learning_rate": 8.328824896838006e-06, "loss": 0.0105, "step": 1177 }, { "epoch": 1.7661169415292353, "grad_norm": 0.04234182612968715, "learning_rate": 8.224537431601886e-06, "loss": 0.0159, "step": 1178 }, { "epoch": 1.767616191904048, "grad_norm": 0.0321447391016331, "learning_rate": 8.12087898511018e-06, "loss": 0.0133, "step": 1179 }, { "epoch": 1.7691154422788604, "grad_norm": 0.07341997220230499, "learning_rate": 8.017850267826232e-06, "loss": 0.0343, "step": 1180 }, { "epoch": 1.7706146926536732, "grad_norm": 0.05203785895813134, "learning_rate": 7.915451985897382e-06, "loss": 0.0321, "step": 1181 }, { "epoch": 1.7721139430284858, "grad_norm": 0.041084567895159414, "learning_rate": 7.81368484114996e-06, "loss": 0.0147, "step": 1182 }, { "epoch": 1.7736131934032984, "grad_norm": 0.031087393200696114, "learning_rate": 7.71254953108458e-06, "loss": 0.0095, "step": 1183 }, { "epoch": 1.775112443778111, "grad_norm": 0.058786196696954986, "learning_rate": 7.612046748871327e-06, "loss": 0.0343, "step": 1184 }, { "epoch": 1.7766116941529235, "grad_norm": 0.04113739414695195, "learning_rate": 7.512177183345026e-06, "loss": 0.0181, "step": 1185 }, { "epoch": 1.778110944527736, "grad_norm": 0.03883727700834806, "learning_rate": 7.412941519000527e-06, "loss": 0.0145, "step": 1186 }, { "epoch": 1.7796101949025487, "grad_norm": 0.041402364797904934, "learning_rate": 7.314340435987921e-06, "loss": 0.0147, "step": 1187 }, { "epoch": 1.7811094452773615, "grad_norm": 0.05299280675746249, "learning_rate": 7.216374610108012e-06, "loss": 0.0199, "step": 1188 }, { "epoch": 1.7826086956521738, "grad_norm": 0.02456671956395044, "learning_rate": 7.119044712807577e-06, "loss": 0.0107, "step": 1189 }, { "epoch": 1.7841079460269866, "grad_norm": 0.03171980923691591, "learning_rate": 7.022351411174866e-06, "loss": 0.0142, "step": 1190 }, { "epoch": 1.785607196401799, "grad_norm": 0.05325976441184379, "learning_rate": 6.92629536793491e-06, "loss": 0.0212, "step": 1191 }, { "epoch": 1.7871064467766118, "grad_norm": 0.04370623276607864, "learning_rate": 6.830877241445111e-06, "loss": 0.0246, "step": 1192 }, { "epoch": 1.7886056971514241, "grad_norm": 0.06432358936521287, "learning_rate": 6.736097685690601e-06, "loss": 0.0208, "step": 1193 }, { "epoch": 1.790104947526237, "grad_norm": 0.03404343307449647, "learning_rate": 6.6419573502798374e-06, "loss": 0.0125, "step": 1194 }, { "epoch": 1.7916041979010495, "grad_norm": 0.056376443801887534, "learning_rate": 6.548456880440135e-06, "loss": 0.0234, "step": 1195 }, { "epoch": 1.793103448275862, "grad_norm": 0.04235044723804527, "learning_rate": 6.455596917013273e-06, "loss": 0.0231, "step": 1196 }, { "epoch": 1.7946026986506747, "grad_norm": 0.03653428510070535, "learning_rate": 6.363378096451011e-06, "loss": 0.0245, "step": 1197 }, { "epoch": 1.7961019490254873, "grad_norm": 0.03721457535492326, "learning_rate": 6.2718010508108545e-06, "loss": 0.0133, "step": 1198 }, { "epoch": 1.7976011994002998, "grad_norm": 0.038261575822808695, "learning_rate": 6.180866407751595e-06, "loss": 0.0169, "step": 1199 }, { "epoch": 1.7991004497751124, "grad_norm": 0.044065953447000386, "learning_rate": 6.090574790529091e-06, "loss": 0.0167, "step": 1200 }, { "epoch": 1.800599700149925, "grad_norm": 0.04771815395917711, "learning_rate": 6.000926817991992e-06, "loss": 0.0181, "step": 1201 }, { "epoch": 1.8020989505247376, "grad_norm": 0.03612809486970171, "learning_rate": 5.911923104577455e-06, "loss": 0.0143, "step": 1202 }, { "epoch": 1.8035982008995504, "grad_norm": 0.06286516442714955, "learning_rate": 5.823564260306968e-06, "loss": 0.0287, "step": 1203 }, { "epoch": 1.8050974512743627, "grad_norm": 0.040689385196554954, "learning_rate": 5.735850890782157e-06, "loss": 0.0154, "step": 1204 }, { "epoch": 1.8065967016491755, "grad_norm": 0.05935355542909139, "learning_rate": 5.6487835971806515e-06, "loss": 0.0249, "step": 1205 }, { "epoch": 1.8080959520239879, "grad_norm": 0.026213533109652836, "learning_rate": 5.562362976251901e-06, "loss": 0.0102, "step": 1206 }, { "epoch": 1.8095952023988007, "grad_norm": 0.0235654511150688, "learning_rate": 5.476589620313188e-06, "loss": 0.0105, "step": 1207 }, { "epoch": 1.811094452773613, "grad_norm": 0.0635366912137273, "learning_rate": 5.39146411724547e-06, "loss": 0.018, "step": 1208 }, { "epoch": 1.8125937031484258, "grad_norm": 0.041487867509385566, "learning_rate": 5.306987050489442e-06, "loss": 0.0366, "step": 1209 }, { "epoch": 1.8140929535232384, "grad_norm": 0.04638926153042607, "learning_rate": 5.223158999041444e-06, "loss": 0.0252, "step": 1210 }, { "epoch": 1.815592203898051, "grad_norm": 0.060453405404883885, "learning_rate": 5.13998053744954e-06, "loss": 0.033, "step": 1211 }, { "epoch": 1.8170914542728636, "grad_norm": 0.026690496589016864, "learning_rate": 5.057452235809624e-06, "loss": 0.0136, "step": 1212 }, { "epoch": 1.8185907046476761, "grad_norm": 0.04363010283975346, "learning_rate": 4.975574659761406e-06, "loss": 0.0228, "step": 1213 }, { "epoch": 1.8200899550224887, "grad_norm": 0.046092550493469546, "learning_rate": 4.8943483704846475e-06, "loss": 0.0261, "step": 1214 }, { "epoch": 1.8215892053973013, "grad_norm": 0.05531628192606062, "learning_rate": 4.8137739246952195e-06, "loss": 0.0145, "step": 1215 }, { "epoch": 1.823088455772114, "grad_norm": 0.08123174441052526, "learning_rate": 4.733851874641382e-06, "loss": 0.0317, "step": 1216 }, { "epoch": 1.8245877061469264, "grad_norm": 0.04288348187685808, "learning_rate": 4.65458276809988e-06, "loss": 0.0203, "step": 1217 }, { "epoch": 1.8260869565217392, "grad_norm": 0.04334355640597478, "learning_rate": 4.575967148372317e-06, "loss": 0.0158, "step": 1218 }, { "epoch": 1.8275862068965516, "grad_norm": 0.037010861171650276, "learning_rate": 4.498005554281337e-06, "loss": 0.0173, "step": 1219 }, { "epoch": 1.8290854572713644, "grad_norm": 0.0389453296408419, "learning_rate": 4.420698520166988e-06, "loss": 0.0125, "step": 1220 }, { "epoch": 1.8305847076461768, "grad_norm": 0.04714273729670721, "learning_rate": 4.34404657588302e-06, "loss": 0.0214, "step": 1221 }, { "epoch": 1.8320839580209896, "grad_norm": 0.07183803253489271, "learning_rate": 4.268050246793276e-06, "loss": 0.0633, "step": 1222 }, { "epoch": 1.8335832083958021, "grad_norm": 0.03354509726743248, "learning_rate": 4.19271005376809e-06, "loss": 0.0209, "step": 1223 }, { "epoch": 1.8350824587706147, "grad_norm": 0.09728906022869646, "learning_rate": 4.118026513180695e-06, "loss": 0.0475, "step": 1224 }, { "epoch": 1.8365817091454273, "grad_norm": 0.03431271996396712, "learning_rate": 4.044000136903736e-06, "loss": 0.013, "step": 1225 }, { "epoch": 1.8380809595202399, "grad_norm": 0.041831243188060474, "learning_rate": 3.970631432305694e-06, "loss": 0.0266, "step": 1226 }, { "epoch": 1.8395802098950524, "grad_norm": 0.057870391772892306, "learning_rate": 3.897920902247465e-06, "loss": 0.0248, "step": 1227 }, { "epoch": 1.841079460269865, "grad_norm": 0.04691162453728428, "learning_rate": 3.825869045078867e-06, "loss": 0.0132, "step": 1228 }, { "epoch": 1.8425787106446778, "grad_norm": 0.03562764414364893, "learning_rate": 3.7544763546352834e-06, "loss": 0.0235, "step": 1229 }, { "epoch": 1.8440779610194902, "grad_norm": 0.06764557500741161, "learning_rate": 3.68374332023419e-06, "loss": 0.0416, "step": 1230 }, { "epoch": 1.845577211394303, "grad_norm": 0.042910585008751875, "learning_rate": 3.6136704266719113e-06, "loss": 0.0221, "step": 1231 }, { "epoch": 1.8470764617691153, "grad_norm": 0.038280889879914855, "learning_rate": 3.5442581542201923e-06, "loss": 0.0162, "step": 1232 }, { "epoch": 1.8485757121439281, "grad_norm": 0.05556984768768352, "learning_rate": 3.475506978623e-06, "loss": 0.023, "step": 1233 }, { "epoch": 1.8500749625187405, "grad_norm": 0.0503673855715429, "learning_rate": 3.40741737109318e-06, "loss": 0.016, "step": 1234 }, { "epoch": 1.8515742128935533, "grad_norm": 0.03274390981831604, "learning_rate": 3.339989798309273e-06, "loss": 0.0115, "step": 1235 }, { "epoch": 1.8530734632683659, "grad_norm": 0.036087042920603704, "learning_rate": 3.273224722412327e-06, "loss": 0.0136, "step": 1236 }, { "epoch": 1.8545727136431784, "grad_norm": 0.06008954234162343, "learning_rate": 3.2071226010026765e-06, "loss": 0.0199, "step": 1237 }, { "epoch": 1.856071964017991, "grad_norm": 0.04454845739153237, "learning_rate": 3.1416838871368924e-06, "loss": 0.0258, "step": 1238 }, { "epoch": 1.8575712143928036, "grad_norm": 0.04443343456113566, "learning_rate": 3.0769090293245705e-06, "loss": 0.0199, "step": 1239 }, { "epoch": 1.8590704647676162, "grad_norm": 0.06673692405990893, "learning_rate": 3.012798471525324e-06, "loss": 0.0633, "step": 1240 }, { "epoch": 1.8605697151424287, "grad_norm": 0.049212673297310765, "learning_rate": 2.949352653145754e-06, "loss": 0.0262, "step": 1241 }, { "epoch": 1.8620689655172413, "grad_norm": 0.06719806910277995, "learning_rate": 2.8865720090364034e-06, "loss": 0.0232, "step": 1242 }, { "epoch": 1.863568215892054, "grad_norm": 0.05588717349611258, "learning_rate": 2.8244569694887426e-06, "loss": 0.0136, "step": 1243 }, { "epoch": 1.8650674662668667, "grad_norm": 0.0404242627373259, "learning_rate": 2.7630079602323442e-06, "loss": 0.012, "step": 1244 }, { "epoch": 1.866566716641679, "grad_norm": 0.03967545521601158, "learning_rate": 2.7022254024318015e-06, "loss": 0.0175, "step": 1245 }, { "epoch": 1.8680659670164919, "grad_norm": 0.026424239288222608, "learning_rate": 2.6421097126839712e-06, "loss": 0.0143, "step": 1246 }, { "epoch": 1.8695652173913042, "grad_norm": 0.02958120459046709, "learning_rate": 2.582661303015066e-06, "loss": 0.0089, "step": 1247 }, { "epoch": 1.871064467766117, "grad_norm": 0.029911673211796365, "learning_rate": 2.5238805808778242e-06, "loss": 0.0098, "step": 1248 }, { "epoch": 1.8725637181409296, "grad_norm": 0.056981287060817974, "learning_rate": 2.465767949148734e-06, "loss": 0.0227, "step": 1249 }, { "epoch": 1.8740629685157422, "grad_norm": 0.0376261021688721, "learning_rate": 2.4083238061252567e-06, "loss": 0.0174, "step": 1250 }, { "epoch": 1.8755622188905547, "grad_norm": 0.034644589104120896, "learning_rate": 2.351548545523141e-06, "loss": 0.0131, "step": 1251 }, { "epoch": 1.8770614692653673, "grad_norm": 0.039399746567982966, "learning_rate": 2.295442556473637e-06, "loss": 0.0162, "step": 1252 }, { "epoch": 1.87856071964018, "grad_norm": 0.03207964462958592, "learning_rate": 2.240006223520941e-06, "loss": 0.0078, "step": 1253 }, { "epoch": 1.8800599700149925, "grad_norm": 0.05260614234208255, "learning_rate": 2.1852399266194314e-06, "loss": 0.0185, "step": 1254 }, { "epoch": 1.881559220389805, "grad_norm": 0.05106484042867738, "learning_rate": 2.1311440411312166e-06, "loss": 0.0226, "step": 1255 }, { "epoch": 1.8830584707646176, "grad_norm": 0.04263872143801973, "learning_rate": 2.0777189378234143e-06, "loss": 0.0117, "step": 1256 }, { "epoch": 1.8845577211394304, "grad_norm": 0.07846972233061164, "learning_rate": 2.024964982865729e-06, "loss": 0.0339, "step": 1257 }, { "epoch": 1.8860569715142428, "grad_norm": 0.029117608897231563, "learning_rate": 1.9728825378278246e-06, "loss": 0.0071, "step": 1258 }, { "epoch": 1.8875562218890556, "grad_norm": 0.0388167533672085, "learning_rate": 1.921471959676957e-06, "loss": 0.0123, "step": 1259 }, { "epoch": 1.889055472263868, "grad_norm": 0.04539482718528663, "learning_rate": 1.8707336007754873e-06, "loss": 0.0179, "step": 1260 }, { "epoch": 1.8905547226386807, "grad_norm": 0.045445265904142454, "learning_rate": 1.820667808878429e-06, "loss": 0.0219, "step": 1261 }, { "epoch": 1.892053973013493, "grad_norm": 0.028894890651339714, "learning_rate": 1.771274927131139e-06, "loss": 0.0145, "step": 1262 }, { "epoch": 1.893553223388306, "grad_norm": 0.042388630611865694, "learning_rate": 1.7225552940668632e-06, "loss": 0.0157, "step": 1263 }, { "epoch": 1.8950524737631185, "grad_norm": 0.06876839316091303, "learning_rate": 1.6745092436045494e-06, "loss": 0.0351, "step": 1264 }, { "epoch": 1.896551724137931, "grad_norm": 0.03573171364688593, "learning_rate": 1.6271371050464056e-06, "loss": 0.0237, "step": 1265 }, { "epoch": 1.8980509745127436, "grad_norm": 0.038482317398168285, "learning_rate": 1.580439203075812e-06, "loss": 0.0124, "step": 1266 }, { "epoch": 1.8995502248875562, "grad_norm": 0.05737954330837195, "learning_rate": 1.5344158577549229e-06, "loss": 0.0364, "step": 1267 }, { "epoch": 1.9010494752623688, "grad_norm": 0.04099948506323669, "learning_rate": 1.4890673845226133e-06, "loss": 0.0197, "step": 1268 }, { "epoch": 1.9025487256371814, "grad_norm": 0.03313456374887663, "learning_rate": 1.444394094192225e-06, "loss": 0.02, "step": 1269 }, { "epoch": 1.9040479760119942, "grad_norm": 0.043427692868520065, "learning_rate": 1.400396292949513e-06, "loss": 0.0202, "step": 1270 }, { "epoch": 1.9055472263868065, "grad_norm": 0.07327623247350938, "learning_rate": 1.3570742823504567e-06, "loss": 0.0245, "step": 1271 }, { "epoch": 1.9070464767616193, "grad_norm": 0.05525179025128676, "learning_rate": 1.3144283593192752e-06, "loss": 0.0246, "step": 1272 }, { "epoch": 1.9085457271364317, "grad_norm": 0.06295368734028475, "learning_rate": 1.272458816146349e-06, "loss": 0.0246, "step": 1273 }, { "epoch": 1.9100449775112445, "grad_norm": 0.034583926367543656, "learning_rate": 1.231165940486234e-06, "loss": 0.0155, "step": 1274 }, { "epoch": 1.9115442278860568, "grad_norm": 0.0437209002683007, "learning_rate": 1.1905500153556625e-06, "loss": 0.0162, "step": 1275 }, { "epoch": 1.9130434782608696, "grad_norm": 0.03904804270714857, "learning_rate": 1.1506113191316447e-06, "loss": 0.0154, "step": 1276 }, { "epoch": 1.9145427286356822, "grad_norm": 0.05420766292701498, "learning_rate": 1.1113501255495485e-06, "loss": 0.0195, "step": 1277 }, { "epoch": 1.9160419790104948, "grad_norm": 0.04566901515231431, "learning_rate": 1.0727667037011668e-06, "loss": 0.0125, "step": 1278 }, { "epoch": 1.9175412293853074, "grad_norm": 0.07198157977225833, "learning_rate": 1.0348613180329757e-06, "loss": 0.0219, "step": 1279 }, { "epoch": 1.91904047976012, "grad_norm": 0.032622544247452506, "learning_rate": 9.976342283442463e-07, "loss": 0.0085, "step": 1280 }, { "epoch": 1.9205397301349325, "grad_norm": 0.03912032612952811, "learning_rate": 9.610856897852682e-07, "loss": 0.0097, "step": 1281 }, { "epoch": 1.922038980509745, "grad_norm": 0.05117531076928059, "learning_rate": 9.252159528556403e-07, "loss": 0.0219, "step": 1282 }, { "epoch": 1.9235382308845579, "grad_norm": 0.048504658124269, "learning_rate": 8.900252634025274e-07, "loss": 0.0203, "step": 1283 }, { "epoch": 1.9250374812593702, "grad_norm": 0.06377251552840921, "learning_rate": 8.555138626189618e-07, "loss": 0.0235, "step": 1284 }, { "epoch": 1.926536731634183, "grad_norm": 0.0353549864848296, "learning_rate": 8.216819870422332e-07, "loss": 0.0141, "step": 1285 }, { "epoch": 1.9280359820089954, "grad_norm": 0.039798781616309425, "learning_rate": 7.885298685522235e-07, "loss": 0.0147, "step": 1286 }, { "epoch": 1.9295352323838082, "grad_norm": 0.03541629405787621, "learning_rate": 7.560577343698305e-07, "loss": 0.0102, "step": 1287 }, { "epoch": 1.9310344827586206, "grad_norm": 0.05561642747147643, "learning_rate": 7.242658070554464e-07, "loss": 0.0206, "step": 1288 }, { "epoch": 1.9325337331334334, "grad_norm": 0.03534178862057278, "learning_rate": 6.931543045073708e-07, "loss": 0.0095, "step": 1289 }, { "epoch": 1.934032983508246, "grad_norm": 0.04209596571556163, "learning_rate": 6.627234399603555e-07, "loss": 0.0245, "step": 1290 }, { "epoch": 1.9355322338830585, "grad_norm": 0.06513898157808579, "learning_rate": 6.32973421984151e-07, "loss": 0.026, "step": 1291 }, { "epoch": 1.937031484257871, "grad_norm": 0.04921099872743036, "learning_rate": 6.039044544820404e-07, "loss": 0.0194, "step": 1292 }, { "epoch": 1.9385307346326837, "grad_norm": 0.031094335152381827, "learning_rate": 5.755167366894632e-07, "loss": 0.0121, "step": 1293 }, { "epoch": 1.9400299850074962, "grad_norm": 0.050223040034524, "learning_rate": 5.478104631726711e-07, "loss": 0.0173, "step": 1294 }, { "epoch": 1.9415292353823088, "grad_norm": 0.03520172367952888, "learning_rate": 5.207858238273522e-07, "loss": 0.019, "step": 1295 }, { "epoch": 1.9430284857571214, "grad_norm": 0.054375401258943634, "learning_rate": 4.944430038773762e-07, "loss": 0.0335, "step": 1296 }, { "epoch": 1.944527736131934, "grad_norm": 0.04736400384851895, "learning_rate": 4.6878218387346183e-07, "loss": 0.0245, "step": 1297 }, { "epoch": 1.9460269865067468, "grad_norm": 0.05631107325781845, "learning_rate": 4.438035396920004e-07, "loss": 0.0173, "step": 1298 }, { "epoch": 1.9475262368815591, "grad_norm": 0.02312088515804648, "learning_rate": 4.1950724253383423e-07, "loss": 0.0098, "step": 1299 }, { "epoch": 1.949025487256372, "grad_norm": 0.06415814349625676, "learning_rate": 3.958934589230467e-07, "loss": 0.0137, "step": 1300 }, { "epoch": 1.9505247376311843, "grad_norm": 0.042944785560430666, "learning_rate": 3.7296235070587435e-07, "loss": 0.0129, "step": 1301 }, { "epoch": 1.952023988005997, "grad_norm": 0.03691120345897048, "learning_rate": 3.50714075049563e-07, "loss": 0.0165, "step": 1302 }, { "epoch": 1.9535232383808094, "grad_norm": 0.047643378112865786, "learning_rate": 3.2914878444131326e-07, "loss": 0.0191, "step": 1303 }, { "epoch": 1.9550224887556222, "grad_norm": 0.04866045710270131, "learning_rate": 3.0826662668720364e-07, "loss": 0.0211, "step": 1304 }, { "epoch": 1.9565217391304348, "grad_norm": 0.035743634358242155, "learning_rate": 2.8806774491120234e-07, "loss": 0.0188, "step": 1305 }, { "epoch": 1.9580209895052474, "grad_norm": 0.0499970530068033, "learning_rate": 2.685522775541904e-07, "loss": 0.0275, "step": 1306 }, { "epoch": 1.95952023988006, "grad_norm": 0.043717836582837635, "learning_rate": 2.497203583729957e-07, "loss": 0.0223, "step": 1307 }, { "epoch": 1.9610194902548725, "grad_norm": 0.06115844054156324, "learning_rate": 2.315721164394713e-07, "loss": 0.0259, "step": 1308 }, { "epoch": 1.9625187406296851, "grad_norm": 0.04674640262799417, "learning_rate": 2.141076761396521e-07, "loss": 0.0296, "step": 1309 }, { "epoch": 1.9640179910044977, "grad_norm": 0.04102373844741974, "learning_rate": 1.973271571728441e-07, "loss": 0.0215, "step": 1310 }, { "epoch": 1.9655172413793105, "grad_norm": 0.05296227342955121, "learning_rate": 1.8123067455084741e-07, "loss": 0.0193, "step": 1311 }, { "epoch": 1.9670164917541229, "grad_norm": 0.04607714250992321, "learning_rate": 1.6581833859716788e-07, "loss": 0.0233, "step": 1312 }, { "epoch": 1.9685157421289357, "grad_norm": 0.05305402469014974, "learning_rate": 1.5109025494620677e-07, "loss": 0.0186, "step": 1313 }, { "epoch": 1.970014992503748, "grad_norm": 0.027178810058074054, "learning_rate": 1.3704652454261668e-07, "loss": 0.008, "step": 1314 }, { "epoch": 1.9715142428785608, "grad_norm": 0.037694527414950885, "learning_rate": 1.2368724364053563e-07, "loss": 0.0095, "step": 1315 }, { "epoch": 1.9730134932533732, "grad_norm": 0.05169984692035686, "learning_rate": 1.1101250380300965e-07, "loss": 0.022, "step": 1316 }, { "epoch": 1.974512743628186, "grad_norm": 0.05203649072494641, "learning_rate": 9.902239190124896e-08, "loss": 0.0205, "step": 1317 }, { "epoch": 1.9760119940029985, "grad_norm": 0.05317970284385352, "learning_rate": 8.771699011416168e-08, "loss": 0.0305, "step": 1318 }, { "epoch": 1.9775112443778111, "grad_norm": 0.046052084092380735, "learning_rate": 7.709637592770991e-08, "loss": 0.02, "step": 1319 }, { "epoch": 1.9790104947526237, "grad_norm": 0.020664658138814455, "learning_rate": 6.71606221343768e-08, "loss": 0.0056, "step": 1320 }, { "epoch": 1.9805097451274363, "grad_norm": 0.04220144244776692, "learning_rate": 5.790979683271136e-08, "loss": 0.0151, "step": 1321 }, { "epoch": 1.9820089955022488, "grad_norm": 0.04084288034944265, "learning_rate": 4.934396342684e-08, "loss": 0.0142, "step": 1322 }, { "epoch": 1.9835082458770614, "grad_norm": 0.05769798222724036, "learning_rate": 4.146318062603349e-08, "loss": 0.0226, "step": 1323 }, { "epoch": 1.9850074962518742, "grad_norm": 0.1100081838811883, "learning_rate": 3.4267502444274015e-08, "loss": 0.0365, "step": 1324 }, { "epoch": 1.9865067466266866, "grad_norm": 0.08427099553965144, "learning_rate": 2.7756978199944273e-08, "loss": 0.0217, "step": 1325 }, { "epoch": 1.9880059970014994, "grad_norm": 0.040887716983706905, "learning_rate": 2.193165251545004e-08, "loss": 0.0154, "step": 1326 }, { "epoch": 1.9895052473763117, "grad_norm": 0.057352690068278346, "learning_rate": 1.6791565316920388e-08, "loss": 0.0193, "step": 1327 }, { "epoch": 1.9910044977511245, "grad_norm": 0.03068600118328415, "learning_rate": 1.2336751833941229e-08, "loss": 0.0175, "step": 1328 }, { "epoch": 1.992503748125937, "grad_norm": 0.04984601232484485, "learning_rate": 8.567242599299973e-09, "loss": 0.0183, "step": 1329 }, { "epoch": 1.9940029985007497, "grad_norm": 0.04524382063881194, "learning_rate": 5.483063448785686e-09, "loss": 0.0199, "step": 1330 }, { "epoch": 1.9955022488755623, "grad_norm": 0.047304184345328626, "learning_rate": 3.0842355210336515e-09, "loss": 0.0228, "step": 1331 }, { "epoch": 1.9970014992503748, "grad_norm": 0.057086267630760426, "learning_rate": 1.3707752573255405e-09, "loss": 0.0232, "step": 1332 }, { "epoch": 1.9985007496251874, "grad_norm": 0.037788074538081265, "learning_rate": 3.426944015227918e-10, "loss": 0.0204, "step": 1333 }, { "epoch": 2.0, "grad_norm": 0.060579723830781834, "learning_rate": 0.0, "loss": 0.0344, "step": 1334 }, { "epoch": 2.0, "eval_loss": 0.05502813309431076, "eval_runtime": 759.9841, "eval_samples_per_second": 13.653, "eval_steps_per_second": 0.854, "step": 1334 }, { "epoch": 2.001499250374813, "grad_norm": 0.030526633679231297, "learning_rate": 6.0285210936521955e-05, "loss": 0.0084, "step": 1335 }, { "epoch": 2.002998500749625, "grad_norm": 0.03260815507881283, "learning_rate": 6.012509310747538e-05, "loss": 0.017, "step": 1336 }, { "epoch": 2.004497751124438, "grad_norm": 0.034484132015163385, "learning_rate": 5.9965096744310526e-05, "loss": 0.0097, "step": 1337 }, { "epoch": 2.0059970014992503, "grad_norm": 0.05283220855422709, "learning_rate": 5.9805222334404e-05, "loss": 0.0311, "step": 1338 }, { "epoch": 2.007496251874063, "grad_norm": 0.026253460939661438, "learning_rate": 5.964547036476099e-05, "loss": 0.0133, "step": 1339 }, { "epoch": 2.0089955022488755, "grad_norm": 0.03822168266257684, "learning_rate": 5.9485841322013755e-05, "loss": 0.03, "step": 1340 }, { "epoch": 2.0104947526236883, "grad_norm": 0.029895481476596094, "learning_rate": 5.9326335692419995e-05, "loss": 0.0073, "step": 1341 }, { "epoch": 2.0119940029985006, "grad_norm": 0.04614213751413987, "learning_rate": 5.916695396186154e-05, "loss": 0.0225, "step": 1342 }, { "epoch": 2.0134932533733134, "grad_norm": 0.07837138979926729, "learning_rate": 5.900769661584272e-05, "loss": 0.0387, "step": 1343 }, { "epoch": 2.0149925037481258, "grad_norm": 0.03684782192112253, "learning_rate": 5.884856413948913e-05, "loss": 0.0164, "step": 1344 }, { "epoch": 2.0164917541229386, "grad_norm": 0.024790672249180405, "learning_rate": 5.868955701754584e-05, "loss": 0.0085, "step": 1345 }, { "epoch": 2.017991004497751, "grad_norm": 0.07922540501835143, "learning_rate": 5.853067573437612e-05, "loss": 0.0193, "step": 1346 }, { "epoch": 2.0194902548725637, "grad_norm": 0.07279842456933099, "learning_rate": 5.83719207739599e-05, "loss": 0.0263, "step": 1347 }, { "epoch": 2.0209895052473765, "grad_norm": 0.03991862147578883, "learning_rate": 5.82132926198923e-05, "loss": 0.0105, "step": 1348 }, { "epoch": 2.022488755622189, "grad_norm": 0.042136413125324325, "learning_rate": 5.805479175538229e-05, "loss": 0.0103, "step": 1349 }, { "epoch": 2.0239880059970017, "grad_norm": 0.04650910883470616, "learning_rate": 5.789641866325091e-05, "loss": 0.0103, "step": 1350 }, { "epoch": 2.025487256371814, "grad_norm": 0.031640051013897205, "learning_rate": 5.773817382593008e-05, "loss": 0.0095, "step": 1351 }, { "epoch": 2.026986506746627, "grad_norm": 0.04644126210808561, "learning_rate": 5.758005772546097e-05, "loss": 0.0142, "step": 1352 }, { "epoch": 2.028485757121439, "grad_norm": 0.07139958336994845, "learning_rate": 5.7422070843492734e-05, "loss": 0.0155, "step": 1353 }, { "epoch": 2.029985007496252, "grad_norm": 0.05667965744148484, "learning_rate": 5.7264213661280765e-05, "loss": 0.0148, "step": 1354 }, { "epoch": 2.0314842578710643, "grad_norm": 0.0483332794145295, "learning_rate": 5.710648665968543e-05, "loss": 0.009, "step": 1355 }, { "epoch": 2.032983508245877, "grad_norm": 0.07707735758620766, "learning_rate": 5.694889031917047e-05, "loss": 0.0139, "step": 1356 }, { "epoch": 2.0344827586206895, "grad_norm": 0.05296978345646472, "learning_rate": 5.679142511980175e-05, "loss": 0.0064, "step": 1357 }, { "epoch": 2.0359820089955023, "grad_norm": 0.05077834707823501, "learning_rate": 5.663409154124557e-05, "loss": 0.0176, "step": 1358 }, { "epoch": 2.0374812593703147, "grad_norm": 0.044143826749535583, "learning_rate": 5.647689006276726e-05, "loss": 0.0084, "step": 1359 }, { "epoch": 2.0389805097451275, "grad_norm": 0.05422495708518017, "learning_rate": 5.63198211632298e-05, "loss": 0.0169, "step": 1360 }, { "epoch": 2.04047976011994, "grad_norm": 0.04874716450172107, "learning_rate": 5.616288532109225e-05, "loss": 0.0124, "step": 1361 }, { "epoch": 2.0419790104947526, "grad_norm": 0.06939726463790843, "learning_rate": 5.6006083014408484e-05, "loss": 0.0149, "step": 1362 }, { "epoch": 2.0434782608695654, "grad_norm": 0.08249839354082386, "learning_rate": 5.584941472082549e-05, "loss": 0.0149, "step": 1363 }, { "epoch": 2.0449775112443778, "grad_norm": 0.037299631634492514, "learning_rate": 5.5692880917582046e-05, "loss": 0.0102, "step": 1364 }, { "epoch": 2.0464767616191906, "grad_norm": 0.07116111066221473, "learning_rate": 5.553648208150728e-05, "loss": 0.0108, "step": 1365 }, { "epoch": 2.047976011994003, "grad_norm": 0.07882506023683916, "learning_rate": 5.5380218689019125e-05, "loss": 0.0178, "step": 1366 }, { "epoch": 2.0494752623688157, "grad_norm": 0.038643229336508024, "learning_rate": 5.522409121612304e-05, "loss": 0.017, "step": 1367 }, { "epoch": 2.050974512743628, "grad_norm": 0.044983719123297856, "learning_rate": 5.506810013841036e-05, "loss": 0.0109, "step": 1368 }, { "epoch": 2.052473763118441, "grad_norm": 0.02127872332720705, "learning_rate": 5.491224593105695e-05, "loss": 0.006, "step": 1369 }, { "epoch": 2.0539730134932532, "grad_norm": 0.04531549401992356, "learning_rate": 5.475652906882173e-05, "loss": 0.0241, "step": 1370 }, { "epoch": 2.055472263868066, "grad_norm": 0.05353781377678766, "learning_rate": 5.4600950026045326e-05, "loss": 0.0183, "step": 1371 }, { "epoch": 2.0569715142428784, "grad_norm": 0.03164654699862933, "learning_rate": 5.4445509276648466e-05, "loss": 0.0046, "step": 1372 }, { "epoch": 2.058470764617691, "grad_norm": 0.032906457495002546, "learning_rate": 5.4290207294130615e-05, "loss": 0.0059, "step": 1373 }, { "epoch": 2.0599700149925035, "grad_norm": 0.03354019985816602, "learning_rate": 5.4135044551568546e-05, "loss": 0.0092, "step": 1374 }, { "epoch": 2.0614692653673163, "grad_norm": 0.03491971395787508, "learning_rate": 5.398002152161484e-05, "loss": 0.0076, "step": 1375 }, { "epoch": 2.062968515742129, "grad_norm": 0.032959044651153006, "learning_rate": 5.382513867649663e-05, "loss": 0.0097, "step": 1376 }, { "epoch": 2.0644677661169415, "grad_norm": 0.052261940463670636, "learning_rate": 5.3670396488013854e-05, "loss": 0.0205, "step": 1377 }, { "epoch": 2.0659670164917543, "grad_norm": 0.04470879681498284, "learning_rate": 5.351579542753807e-05, "loss": 0.014, "step": 1378 }, { "epoch": 2.0674662668665666, "grad_norm": 0.03567561984300728, "learning_rate": 5.33613359660109e-05, "loss": 0.0115, "step": 1379 }, { "epoch": 2.0689655172413794, "grad_norm": 0.051219231023299464, "learning_rate": 5.320701857394268e-05, "loss": 0.0186, "step": 1380 }, { "epoch": 2.070464767616192, "grad_norm": 0.03593213649901054, "learning_rate": 5.305284372141095e-05, "loss": 0.0079, "step": 1381 }, { "epoch": 2.0719640179910046, "grad_norm": 0.04593157344092489, "learning_rate": 5.2898811878059e-05, "loss": 0.007, "step": 1382 }, { "epoch": 2.073463268365817, "grad_norm": 0.026998482178819172, "learning_rate": 5.274492351309461e-05, "loss": 0.0076, "step": 1383 }, { "epoch": 2.0749625187406298, "grad_norm": 0.06694277013267871, "learning_rate": 5.259117909528839e-05, "loss": 0.0287, "step": 1384 }, { "epoch": 2.076461769115442, "grad_norm": 0.04784933560448736, "learning_rate": 5.243757909297247e-05, "loss": 0.0208, "step": 1385 }, { "epoch": 2.077961019490255, "grad_norm": 0.04315966968570286, "learning_rate": 5.2284123974039154e-05, "loss": 0.0126, "step": 1386 }, { "epoch": 2.0794602698650673, "grad_norm": 0.037267520402257566, "learning_rate": 5.213081420593933e-05, "loss": 0.0108, "step": 1387 }, { "epoch": 2.08095952023988, "grad_norm": 0.056583335801616, "learning_rate": 5.197765025568109e-05, "loss": 0.0152, "step": 1388 }, { "epoch": 2.082458770614693, "grad_norm": 0.052305309697487534, "learning_rate": 5.182463258982846e-05, "loss": 0.013, "step": 1389 }, { "epoch": 2.0839580209895052, "grad_norm": 0.05404489918344409, "learning_rate": 5.1671761674499765e-05, "loss": 0.0151, "step": 1390 }, { "epoch": 2.085457271364318, "grad_norm": 0.039207847861306155, "learning_rate": 5.15190379753663e-05, "loss": 0.0161, "step": 1391 }, { "epoch": 2.0869565217391304, "grad_norm": 0.056056691851198825, "learning_rate": 5.1366461957650954e-05, "loss": 0.0113, "step": 1392 }, { "epoch": 2.088455772113943, "grad_norm": 0.07556511845267178, "learning_rate": 5.121403408612672e-05, "loss": 0.0165, "step": 1393 }, { "epoch": 2.0899550224887555, "grad_norm": 0.0446161948430827, "learning_rate": 5.1061754825115374e-05, "loss": 0.0152, "step": 1394 }, { "epoch": 2.0914542728635683, "grad_norm": 0.05616467657378434, "learning_rate": 5.090962463848592e-05, "loss": 0.0182, "step": 1395 }, { "epoch": 2.0929535232383807, "grad_norm": 0.07070249064708027, "learning_rate": 5.07576439896533e-05, "loss": 0.0198, "step": 1396 }, { "epoch": 2.0944527736131935, "grad_norm": 0.060127913574885505, "learning_rate": 5.0605813341576924e-05, "loss": 0.0143, "step": 1397 }, { "epoch": 2.095952023988006, "grad_norm": 0.06067598140840922, "learning_rate": 5.045413315675924e-05, "loss": 0.0226, "step": 1398 }, { "epoch": 2.0974512743628186, "grad_norm": 0.041993761062937916, "learning_rate": 5.0302603897244474e-05, "loss": 0.0096, "step": 1399 }, { "epoch": 2.098950524737631, "grad_norm": 0.04365553922484038, "learning_rate": 5.015122602461698e-05, "loss": 0.0091, "step": 1400 }, { "epoch": 2.100449775112444, "grad_norm": 0.02977769936940776, "learning_rate": 5.000000000000002e-05, "loss": 0.0071, "step": 1401 }, { "epoch": 2.101949025487256, "grad_norm": 0.029553763597095526, "learning_rate": 4.9848926284054255e-05, "loss": 0.0071, "step": 1402 }, { "epoch": 2.103448275862069, "grad_norm": 0.04402728282071913, "learning_rate": 4.969800533697649e-05, "loss": 0.0104, "step": 1403 }, { "epoch": 2.1049475262368817, "grad_norm": 0.04811877540233524, "learning_rate": 4.9547237618498085e-05, "loss": 0.0086, "step": 1404 }, { "epoch": 2.106446776611694, "grad_norm": 0.09241803834890136, "learning_rate": 4.939662358788364e-05, "loss": 0.0145, "step": 1405 }, { "epoch": 2.107946026986507, "grad_norm": 0.056831916289827006, "learning_rate": 4.924616370392961e-05, "loss": 0.0126, "step": 1406 }, { "epoch": 2.1094452773613193, "grad_norm": 0.04431417830054459, "learning_rate": 4.909585842496287e-05, "loss": 0.0108, "step": 1407 }, { "epoch": 2.110944527736132, "grad_norm": 0.04425585885215059, "learning_rate": 4.894570820883944e-05, "loss": 0.0126, "step": 1408 }, { "epoch": 2.1124437781109444, "grad_norm": 0.04004557406045061, "learning_rate": 4.8795713512942865e-05, "loss": 0.0076, "step": 1409 }, { "epoch": 2.113943028485757, "grad_norm": 0.053630460067237214, "learning_rate": 4.864587479418302e-05, "loss": 0.0105, "step": 1410 }, { "epoch": 2.1154422788605696, "grad_norm": 0.04789810127257941, "learning_rate": 4.8496192508994576e-05, "loss": 0.0153, "step": 1411 }, { "epoch": 2.1169415292353824, "grad_norm": 0.045260113348342214, "learning_rate": 4.834666711333582e-05, "loss": 0.009, "step": 1412 }, { "epoch": 2.1184407796101947, "grad_norm": 0.03897446058669094, "learning_rate": 4.8197299062686995e-05, "loss": 0.0137, "step": 1413 }, { "epoch": 2.1199400299850075, "grad_norm": 0.04760025206120078, "learning_rate": 4.8048088812049096e-05, "loss": 0.0145, "step": 1414 }, { "epoch": 2.12143928035982, "grad_norm": 0.041240639670586346, "learning_rate": 4.78990368159424e-05, "loss": 0.0158, "step": 1415 }, { "epoch": 2.1229385307346327, "grad_norm": 0.0767178982606092, "learning_rate": 4.7750143528405126e-05, "loss": 0.0317, "step": 1416 }, { "epoch": 2.1244377811094455, "grad_norm": 0.06951320687165209, "learning_rate": 4.7601409402992106e-05, "loss": 0.0215, "step": 1417 }, { "epoch": 2.125937031484258, "grad_norm": 0.0376654940983486, "learning_rate": 4.745283489277325e-05, "loss": 0.0064, "step": 1418 }, { "epoch": 2.1274362818590706, "grad_norm": 0.0406115405255025, "learning_rate": 4.7304420450332244e-05, "loss": 0.0092, "step": 1419 }, { "epoch": 2.128935532233883, "grad_norm": 0.04263626570177118, "learning_rate": 4.71561665277653e-05, "loss": 0.0102, "step": 1420 }, { "epoch": 2.130434782608696, "grad_norm": 0.043526034150673305, "learning_rate": 4.700807357667952e-05, "loss": 0.0127, "step": 1421 }, { "epoch": 2.131934032983508, "grad_norm": 0.061694508628830164, "learning_rate": 4.68601420481917e-05, "loss": 0.0154, "step": 1422 }, { "epoch": 2.133433283358321, "grad_norm": 0.02736493766034135, "learning_rate": 4.6712372392927e-05, "loss": 0.0031, "step": 1423 }, { "epoch": 2.1349325337331333, "grad_norm": 0.044103353881944696, "learning_rate": 4.656476506101737e-05, "loss": 0.0068, "step": 1424 }, { "epoch": 2.136431784107946, "grad_norm": 0.04642610291093845, "learning_rate": 4.6417320502100316e-05, "loss": 0.0122, "step": 1425 }, { "epoch": 2.1379310344827585, "grad_norm": 0.046082875327096894, "learning_rate": 4.6270039165317605e-05, "loss": 0.0087, "step": 1426 }, { "epoch": 2.1394302848575713, "grad_norm": 0.0708451752034666, "learning_rate": 4.612292149931369e-05, "loss": 0.0239, "step": 1427 }, { "epoch": 2.1409295352323836, "grad_norm": 0.048742881372561137, "learning_rate": 4.59759679522345e-05, "loss": 0.0131, "step": 1428 }, { "epoch": 2.1424287856071964, "grad_norm": 0.037825120700997646, "learning_rate": 4.582917897172603e-05, "loss": 0.008, "step": 1429 }, { "epoch": 2.143928035982009, "grad_norm": 0.039576887673642355, "learning_rate": 4.568255500493292e-05, "loss": 0.0077, "step": 1430 }, { "epoch": 2.1454272863568216, "grad_norm": 0.04615903896006217, "learning_rate": 4.5536096498497295e-05, "loss": 0.0128, "step": 1431 }, { "epoch": 2.1469265367316344, "grad_norm": 0.05184700373275429, "learning_rate": 4.5389803898557106e-05, "loss": 0.0135, "step": 1432 }, { "epoch": 2.1484257871064467, "grad_norm": 0.05512409656768961, "learning_rate": 4.524367765074499e-05, "loss": 0.0191, "step": 1433 }, { "epoch": 2.1499250374812595, "grad_norm": 0.037605843781708934, "learning_rate": 4.5097718200186814e-05, "loss": 0.0085, "step": 1434 }, { "epoch": 2.151424287856072, "grad_norm": 0.0377762954626737, "learning_rate": 4.495192599150044e-05, "loss": 0.008, "step": 1435 }, { "epoch": 2.1529235382308847, "grad_norm": 0.09983197406271346, "learning_rate": 4.480630146879419e-05, "loss": 0.0395, "step": 1436 }, { "epoch": 2.154422788605697, "grad_norm": 0.0462254950663306, "learning_rate": 4.46608450756656e-05, "loss": 0.0084, "step": 1437 }, { "epoch": 2.15592203898051, "grad_norm": 0.044564591662625865, "learning_rate": 4.451555725520009e-05, "loss": 0.0081, "step": 1438 }, { "epoch": 2.157421289355322, "grad_norm": 0.04963717588235369, "learning_rate": 4.437043844996952e-05, "loss": 0.016, "step": 1439 }, { "epoch": 2.158920539730135, "grad_norm": 0.04383443652884915, "learning_rate": 4.4225489102030995e-05, "loss": 0.0143, "step": 1440 }, { "epoch": 2.1604197901049473, "grad_norm": 0.05901275047312385, "learning_rate": 4.4080709652925336e-05, "loss": 0.0329, "step": 1441 }, { "epoch": 2.16191904047976, "grad_norm": 0.057701053237886664, "learning_rate": 4.393610054367585e-05, "loss": 0.0158, "step": 1442 }, { "epoch": 2.1634182908545725, "grad_norm": 0.032964363694990895, "learning_rate": 4.379166221478697e-05, "loss": 0.0066, "step": 1443 }, { "epoch": 2.1649175412293853, "grad_norm": 0.04389828162413142, "learning_rate": 4.364739510624286e-05, "loss": 0.0164, "step": 1444 }, { "epoch": 2.166416791604198, "grad_norm": 0.0444285469562732, "learning_rate": 4.350329965750621e-05, "loss": 0.0137, "step": 1445 }, { "epoch": 2.1679160419790104, "grad_norm": 0.04033184033008043, "learning_rate": 4.335937630751674e-05, "loss": 0.0082, "step": 1446 }, { "epoch": 2.1694152923538232, "grad_norm": 0.04133013162262341, "learning_rate": 4.32156254946899e-05, "loss": 0.0148, "step": 1447 }, { "epoch": 2.1709145427286356, "grad_norm": 0.04479829808766385, "learning_rate": 4.307204765691558e-05, "loss": 0.0135, "step": 1448 }, { "epoch": 2.1724137931034484, "grad_norm": 0.038005140860989536, "learning_rate": 4.2928643231556844e-05, "loss": 0.0088, "step": 1449 }, { "epoch": 2.1739130434782608, "grad_norm": 0.04944720263306113, "learning_rate": 4.27854126554484e-05, "loss": 0.0141, "step": 1450 }, { "epoch": 2.1754122938530736, "grad_norm": 0.036944980214473636, "learning_rate": 4.264235636489542e-05, "loss": 0.0119, "step": 1451 }, { "epoch": 2.176911544227886, "grad_norm": 0.03355142923993928, "learning_rate": 4.249947479567218e-05, "loss": 0.0086, "step": 1452 }, { "epoch": 2.1784107946026987, "grad_norm": 0.03593832486948932, "learning_rate": 4.235676838302068e-05, "loss": 0.0077, "step": 1453 }, { "epoch": 2.179910044977511, "grad_norm": 0.028107461232522123, "learning_rate": 4.221423756164948e-05, "loss": 0.0063, "step": 1454 }, { "epoch": 2.181409295352324, "grad_norm": 0.06824963682160574, "learning_rate": 4.207188276573214e-05, "loss": 0.0201, "step": 1455 }, { "epoch": 2.1829085457271367, "grad_norm": 0.06144807651242615, "learning_rate": 4.1929704428906026e-05, "loss": 0.0086, "step": 1456 }, { "epoch": 2.184407796101949, "grad_norm": 0.033546077444945685, "learning_rate": 4.1787702984271074e-05, "loss": 0.0165, "step": 1457 }, { "epoch": 2.185907046476762, "grad_norm": 0.05347378206377532, "learning_rate": 4.1645878864388266e-05, "loss": 0.0208, "step": 1458 }, { "epoch": 2.187406296851574, "grad_norm": 0.04962962076984121, "learning_rate": 4.150423250127845e-05, "loss": 0.0091, "step": 1459 }, { "epoch": 2.188905547226387, "grad_norm": 0.05529687338394011, "learning_rate": 4.1362764326421064e-05, "loss": 0.0109, "step": 1460 }, { "epoch": 2.1904047976011993, "grad_norm": 0.04613412146675067, "learning_rate": 4.12214747707527e-05, "loss": 0.0173, "step": 1461 }, { "epoch": 2.191904047976012, "grad_norm": 0.07673891829272307, "learning_rate": 4.1080364264665774e-05, "loss": 0.0214, "step": 1462 }, { "epoch": 2.1934032983508245, "grad_norm": 0.06300515486789615, "learning_rate": 4.093943323800745e-05, "loss": 0.0234, "step": 1463 }, { "epoch": 2.1949025487256373, "grad_norm": 0.033506273076591, "learning_rate": 4.0798682120078044e-05, "loss": 0.0053, "step": 1464 }, { "epoch": 2.1964017991004496, "grad_norm": 0.056817557856699966, "learning_rate": 4.065811133962987e-05, "loss": 0.0177, "step": 1465 }, { "epoch": 2.1979010494752624, "grad_norm": 0.05023022865741867, "learning_rate": 4.0517721324865884e-05, "loss": 0.0048, "step": 1466 }, { "epoch": 2.199400299850075, "grad_norm": 0.05491462712075261, "learning_rate": 4.037751250343841e-05, "loss": 0.0113, "step": 1467 }, { "epoch": 2.2008995502248876, "grad_norm": 0.04175729132601896, "learning_rate": 4.023748530244789e-05, "loss": 0.0056, "step": 1468 }, { "epoch": 2.2023988005997, "grad_norm": 0.05675930835600266, "learning_rate": 4.009764014844143e-05, "loss": 0.0083, "step": 1469 }, { "epoch": 2.2038980509745127, "grad_norm": 0.04479109003044637, "learning_rate": 3.9957977467411615e-05, "loss": 0.0067, "step": 1470 }, { "epoch": 2.2053973013493255, "grad_norm": 0.07147407160092135, "learning_rate": 3.981849768479517e-05, "loss": 0.0084, "step": 1471 }, { "epoch": 2.206896551724138, "grad_norm": 0.043091491533339646, "learning_rate": 3.967920122547175e-05, "loss": 0.0153, "step": 1472 }, { "epoch": 2.2083958020989507, "grad_norm": 0.03128488029866975, "learning_rate": 3.954008851376252e-05, "loss": 0.0112, "step": 1473 }, { "epoch": 2.209895052473763, "grad_norm": 0.04158959465981512, "learning_rate": 3.940115997342891e-05, "loss": 0.0103, "step": 1474 }, { "epoch": 2.211394302848576, "grad_norm": 0.13221307776902036, "learning_rate": 3.9262416027671356e-05, "loss": 0.0453, "step": 1475 }, { "epoch": 2.212893553223388, "grad_norm": 0.05892082943897017, "learning_rate": 3.9123857099127936e-05, "loss": 0.0181, "step": 1476 }, { "epoch": 2.214392803598201, "grad_norm": 0.05630268194117581, "learning_rate": 3.8985483609873244e-05, "loss": 0.0108, "step": 1477 }, { "epoch": 2.2158920539730134, "grad_norm": 0.05183425927723012, "learning_rate": 3.8847295981416896e-05, "loss": 0.0171, "step": 1478 }, { "epoch": 2.217391304347826, "grad_norm": 0.0657409520932719, "learning_rate": 3.8709294634702376e-05, "loss": 0.0167, "step": 1479 }, { "epoch": 2.2188905547226385, "grad_norm": 0.038619103418217976, "learning_rate": 3.857147999010567e-05, "loss": 0.0162, "step": 1480 }, { "epoch": 2.2203898050974513, "grad_norm": 0.04096807453440185, "learning_rate": 3.843385246743417e-05, "loss": 0.0107, "step": 1481 }, { "epoch": 2.2218890554722637, "grad_norm": 0.05019401146627568, "learning_rate": 3.829641248592515e-05, "loss": 0.0225, "step": 1482 }, { "epoch": 2.2233883058470765, "grad_norm": 0.03672403442708744, "learning_rate": 3.8159160464244606e-05, "loss": 0.01, "step": 1483 }, { "epoch": 2.224887556221889, "grad_norm": 0.024794475931188113, "learning_rate": 3.802209682048602e-05, "loss": 0.0054, "step": 1484 }, { "epoch": 2.2263868065967016, "grad_norm": 0.04410192288961744, "learning_rate": 3.788522197216897e-05, "loss": 0.0076, "step": 1485 }, { "epoch": 2.2278860569715144, "grad_norm": 0.05107537883022716, "learning_rate": 3.774853633623806e-05, "loss": 0.0088, "step": 1486 }, { "epoch": 2.229385307346327, "grad_norm": 0.05577006855214796, "learning_rate": 3.7612040329061405e-05, "loss": 0.0217, "step": 1487 }, { "epoch": 2.2308845577211396, "grad_norm": 0.039857059388175874, "learning_rate": 3.747573436642951e-05, "loss": 0.0081, "step": 1488 }, { "epoch": 2.232383808095952, "grad_norm": 0.0384663647554164, "learning_rate": 3.733961886355398e-05, "loss": 0.009, "step": 1489 }, { "epoch": 2.2338830584707647, "grad_norm": 0.049273609114417555, "learning_rate": 3.720369423506622e-05, "loss": 0.0155, "step": 1490 }, { "epoch": 2.235382308845577, "grad_norm": 0.033532191368091685, "learning_rate": 3.7067960895016275e-05, "loss": 0.0072, "step": 1491 }, { "epoch": 2.23688155922039, "grad_norm": 0.04503556398704985, "learning_rate": 3.6932419256871406e-05, "loss": 0.0141, "step": 1492 }, { "epoch": 2.2383808095952022, "grad_norm": 0.05142000964521401, "learning_rate": 3.679706973351491e-05, "loss": 0.0108, "step": 1493 }, { "epoch": 2.239880059970015, "grad_norm": 0.0876457403280232, "learning_rate": 3.6661912737245e-05, "loss": 0.0291, "step": 1494 }, { "epoch": 2.2413793103448274, "grad_norm": 0.044558339670786255, "learning_rate": 3.6526948679773257e-05, "loss": 0.0104, "step": 1495 }, { "epoch": 2.24287856071964, "grad_norm": 0.044561154097932286, "learning_rate": 3.6392177972223594e-05, "loss": 0.0082, "step": 1496 }, { "epoch": 2.244377811094453, "grad_norm": 0.04868383395902538, "learning_rate": 3.6257601025131026e-05, "loss": 0.0113, "step": 1497 }, { "epoch": 2.2458770614692654, "grad_norm": 0.04172072540339245, "learning_rate": 3.612321824844024e-05, "loss": 0.0176, "step": 1498 }, { "epoch": 2.247376311844078, "grad_norm": 0.03236642412170006, "learning_rate": 3.5989030051504434e-05, "loss": 0.0115, "step": 1499 }, { "epoch": 2.2488755622188905, "grad_norm": 0.039260085337301305, "learning_rate": 3.585503684308421e-05, "loss": 0.0102, "step": 1500 }, { "epoch": 2.2503748125937033, "grad_norm": 0.037582210428259974, "learning_rate": 3.5721239031346066e-05, "loss": 0.0109, "step": 1501 }, { "epoch": 2.2518740629685157, "grad_norm": 0.041090729210075276, "learning_rate": 3.558763702386135e-05, "loss": 0.0074, "step": 1502 }, { "epoch": 2.2533733133433285, "grad_norm": 0.032337449915865756, "learning_rate": 3.545423122760493e-05, "loss": 0.0083, "step": 1503 }, { "epoch": 2.254872563718141, "grad_norm": 0.039350512713843, "learning_rate": 3.5321022048954035e-05, "loss": 0.0131, "step": 1504 }, { "epoch": 2.2563718140929536, "grad_norm": 0.04339315612545275, "learning_rate": 3.518800989368691e-05, "loss": 0.0099, "step": 1505 }, { "epoch": 2.257871064467766, "grad_norm": 0.029058192562887976, "learning_rate": 3.5055195166981645e-05, "loss": 0.0043, "step": 1506 }, { "epoch": 2.2593703148425788, "grad_norm": 0.07125305496752797, "learning_rate": 3.492257827341492e-05, "loss": 0.0135, "step": 1507 }, { "epoch": 2.260869565217391, "grad_norm": 0.045867837619974795, "learning_rate": 3.479015961696077e-05, "loss": 0.0102, "step": 1508 }, { "epoch": 2.262368815592204, "grad_norm": 0.035267429016013714, "learning_rate": 3.465793960098945e-05, "loss": 0.0071, "step": 1509 }, { "epoch": 2.2638680659670163, "grad_norm": 0.03260091263652589, "learning_rate": 3.452591862826603e-05, "loss": 0.011, "step": 1510 }, { "epoch": 2.265367316341829, "grad_norm": 0.04058198108830388, "learning_rate": 3.439409710094929e-05, "loss": 0.0068, "step": 1511 }, { "epoch": 2.266866566716642, "grad_norm": 0.04952419185234942, "learning_rate": 3.426247542059041e-05, "loss": 0.0107, "step": 1512 }, { "epoch": 2.2683658170914542, "grad_norm": 0.046925097632493966, "learning_rate": 3.413105398813195e-05, "loss": 0.0089, "step": 1513 }, { "epoch": 2.269865067466267, "grad_norm": 0.033475324121524125, "learning_rate": 3.3999833203906326e-05, "loss": 0.0113, "step": 1514 }, { "epoch": 2.2713643178410794, "grad_norm": 0.047232578170488856, "learning_rate": 3.386881346763483e-05, "loss": 0.018, "step": 1515 }, { "epoch": 2.272863568215892, "grad_norm": 0.06318786942516515, "learning_rate": 3.373799517842627e-05, "loss": 0.021, "step": 1516 }, { "epoch": 2.2743628185907045, "grad_norm": 0.05874046267941404, "learning_rate": 3.360737873477584e-05, "loss": 0.0163, "step": 1517 }, { "epoch": 2.2758620689655173, "grad_norm": 0.053281566861240165, "learning_rate": 3.347696453456393e-05, "loss": 0.0253, "step": 1518 }, { "epoch": 2.2773613193403297, "grad_norm": 0.04864313942217458, "learning_rate": 3.334675297505476e-05, "loss": 0.011, "step": 1519 }, { "epoch": 2.2788605697151425, "grad_norm": 0.03932790276423907, "learning_rate": 3.3216744452895354e-05, "loss": 0.0066, "step": 1520 }, { "epoch": 2.280359820089955, "grad_norm": 0.05388383603716058, "learning_rate": 3.308693936411421e-05, "loss": 0.0164, "step": 1521 }, { "epoch": 2.2818590704647677, "grad_norm": 0.059210537711188005, "learning_rate": 3.2957338104120096e-05, "loss": 0.0211, "step": 1522 }, { "epoch": 2.2833583208395805, "grad_norm": 0.03391640055470483, "learning_rate": 3.2827941067700996e-05, "loss": 0.0095, "step": 1523 }, { "epoch": 2.284857571214393, "grad_norm": 0.046108466735742315, "learning_rate": 3.269874864902269e-05, "loss": 0.0121, "step": 1524 }, { "epoch": 2.286356821589205, "grad_norm": 0.029553171354202632, "learning_rate": 3.2569761241627696e-05, "loss": 0.0068, "step": 1525 }, { "epoch": 2.287856071964018, "grad_norm": 0.05917914395956801, "learning_rate": 3.244097923843398e-05, "loss": 0.0174, "step": 1526 }, { "epoch": 2.2893553223388308, "grad_norm": 0.0501098278181535, "learning_rate": 3.231240303173394e-05, "loss": 0.0124, "step": 1527 }, { "epoch": 2.290854572713643, "grad_norm": 0.05486928431377507, "learning_rate": 3.218403301319296e-05, "loss": 0.0103, "step": 1528 }, { "epoch": 2.292353823088456, "grad_norm": 0.03881235745568428, "learning_rate": 3.205586957384838e-05, "loss": 0.0097, "step": 1529 }, { "epoch": 2.2938530734632683, "grad_norm": 0.051425672269962086, "learning_rate": 3.192791310410822e-05, "loss": 0.0072, "step": 1530 }, { "epoch": 2.295352323838081, "grad_norm": 0.03434030113669453, "learning_rate": 3.1800163993750166e-05, "loss": 0.006, "step": 1531 }, { "epoch": 2.2968515742128934, "grad_norm": 0.06088574673287806, "learning_rate": 3.16726226319201e-05, "loss": 0.0309, "step": 1532 }, { "epoch": 2.2983508245877062, "grad_norm": 0.050410995201691795, "learning_rate": 3.154528940713113e-05, "loss": 0.0119, "step": 1533 }, { "epoch": 2.2998500749625186, "grad_norm": 0.03733199016104634, "learning_rate": 3.141816470726238e-05, "loss": 0.0131, "step": 1534 }, { "epoch": 2.3013493253373314, "grad_norm": 0.04130140056610659, "learning_rate": 3.129124891955771e-05, "loss": 0.0105, "step": 1535 }, { "epoch": 2.3028485757121437, "grad_norm": 0.033174106752094965, "learning_rate": 3.116454243062459e-05, "loss": 0.0168, "step": 1536 }, { "epoch": 2.3043478260869565, "grad_norm": 0.1438371200725692, "learning_rate": 3.103804562643302e-05, "loss": 0.1241, "step": 1537 }, { "epoch": 2.3058470764617693, "grad_norm": 0.04625158470637066, "learning_rate": 3.0911758892314166e-05, "loss": 0.014, "step": 1538 }, { "epoch": 2.3073463268365817, "grad_norm": 0.07587323672779746, "learning_rate": 3.078568261295933e-05, "loss": 0.041, "step": 1539 }, { "epoch": 2.3088455772113945, "grad_norm": 0.04717314691473391, "learning_rate": 3.0659817172418693e-05, "loss": 0.0168, "step": 1540 }, { "epoch": 2.310344827586207, "grad_norm": 0.056904306887284116, "learning_rate": 3.053416295410026e-05, "loss": 0.0271, "step": 1541 }, { "epoch": 2.3118440779610197, "grad_norm": 0.05225748974710264, "learning_rate": 3.0408720340768572e-05, "loss": 0.0199, "step": 1542 }, { "epoch": 2.313343328335832, "grad_norm": 0.027519904567484686, "learning_rate": 3.0283489714543556e-05, "loss": 0.0043, "step": 1543 }, { "epoch": 2.314842578710645, "grad_norm": 0.044522929974943656, "learning_rate": 3.0158471456899428e-05, "loss": 0.0095, "step": 1544 }, { "epoch": 2.316341829085457, "grad_norm": 0.03422845505314717, "learning_rate": 3.0033665948663448e-05, "loss": 0.0044, "step": 1545 }, { "epoch": 2.31784107946027, "grad_norm": 0.04409770557592122, "learning_rate": 2.9909073570014912e-05, "loss": 0.0132, "step": 1546 }, { "epoch": 2.3193403298350823, "grad_norm": 0.05235014133031246, "learning_rate": 2.9784694700483762e-05, "loss": 0.0142, "step": 1547 }, { "epoch": 2.320839580209895, "grad_norm": 0.054797224418192, "learning_rate": 2.9660529718949627e-05, "loss": 0.0213, "step": 1548 }, { "epoch": 2.3223388305847075, "grad_norm": 0.05562034477470897, "learning_rate": 2.953657900364053e-05, "loss": 0.0219, "step": 1549 }, { "epoch": 2.3238380809595203, "grad_norm": 0.04251356727567171, "learning_rate": 2.94128429321319e-05, "loss": 0.013, "step": 1550 }, { "epoch": 2.3253373313343326, "grad_norm": 0.06524250288280184, "learning_rate": 2.9289321881345254e-05, "loss": 0.0469, "step": 1551 }, { "epoch": 2.3268365817091454, "grad_norm": 0.02779154545520239, "learning_rate": 2.9166016227547133e-05, "loss": 0.0063, "step": 1552 }, { "epoch": 2.3283358320839582, "grad_norm": 0.05956826227184479, "learning_rate": 2.904292634634793e-05, "loss": 0.0129, "step": 1553 }, { "epoch": 2.3298350824587706, "grad_norm": 0.04251497268994643, "learning_rate": 2.8920052612700754e-05, "loss": 0.0118, "step": 1554 }, { "epoch": 2.3313343328335834, "grad_norm": 0.03270063785303954, "learning_rate": 2.879739540090036e-05, "loss": 0.0118, "step": 1555 }, { "epoch": 2.3328335832083957, "grad_norm": 0.04275952830653872, "learning_rate": 2.8674955084581857e-05, "loss": 0.0129, "step": 1556 }, { "epoch": 2.3343328335832085, "grad_norm": 0.03808980692148684, "learning_rate": 2.8552732036719687e-05, "loss": 0.0115, "step": 1557 }, { "epoch": 2.335832083958021, "grad_norm": 0.04211694412181197, "learning_rate": 2.8430726629626413e-05, "loss": 0.0096, "step": 1558 }, { "epoch": 2.3373313343328337, "grad_norm": 0.08377128491024728, "learning_rate": 2.8308939234951726e-05, "loss": 0.0204, "step": 1559 }, { "epoch": 2.338830584707646, "grad_norm": 0.04886421927904402, "learning_rate": 2.8187370223681132e-05, "loss": 0.0129, "step": 1560 }, { "epoch": 2.340329835082459, "grad_norm": 0.03249443016326716, "learning_rate": 2.8066019966134904e-05, "loss": 0.0134, "step": 1561 }, { "epoch": 2.341829085457271, "grad_norm": 0.0682646983117486, "learning_rate": 2.7944888831966987e-05, "loss": 0.015, "step": 1562 }, { "epoch": 2.343328335832084, "grad_norm": 0.062153885256310916, "learning_rate": 2.7823977190163786e-05, "loss": 0.0146, "step": 1563 }, { "epoch": 2.344827586206897, "grad_norm": 0.06875104640865384, "learning_rate": 2.770328540904319e-05, "loss": 0.0207, "step": 1564 }, { "epoch": 2.346326836581709, "grad_norm": 0.04559235897489107, "learning_rate": 2.7582813856253275e-05, "loss": 0.0123, "step": 1565 }, { "epoch": 2.3478260869565215, "grad_norm": 0.06364832744778683, "learning_rate": 2.746256289877126e-05, "loss": 0.0174, "step": 1566 }, { "epoch": 2.3493253373313343, "grad_norm": 0.08985733143769112, "learning_rate": 2.734253290290242e-05, "loss": 0.0274, "step": 1567 }, { "epoch": 2.350824587706147, "grad_norm": 0.048931837923607266, "learning_rate": 2.722272423427896e-05, "loss": 0.0114, "step": 1568 }, { "epoch": 2.3523238380809595, "grad_norm": 0.04068516417645611, "learning_rate": 2.7103137257858868e-05, "loss": 0.0159, "step": 1569 }, { "epoch": 2.3538230884557723, "grad_norm": 0.0560986602506305, "learning_rate": 2.698377233792476e-05, "loss": 0.0253, "step": 1570 }, { "epoch": 2.3553223388305846, "grad_norm": 0.04599797703427196, "learning_rate": 2.6864629838082956e-05, "loss": 0.0159, "step": 1571 }, { "epoch": 2.3568215892053974, "grad_norm": 0.04078821680618464, "learning_rate": 2.6745710121262136e-05, "loss": 0.012, "step": 1572 }, { "epoch": 2.3583208395802098, "grad_norm": 0.06842020690415691, "learning_rate": 2.6627013549712355e-05, "loss": 0.0148, "step": 1573 }, { "epoch": 2.3598200899550226, "grad_norm": 0.03319032136524963, "learning_rate": 2.6508540485004006e-05, "loss": 0.0097, "step": 1574 }, { "epoch": 2.361319340329835, "grad_norm": 0.05139456085562142, "learning_rate": 2.639029128802657e-05, "loss": 0.0146, "step": 1575 }, { "epoch": 2.3628185907046477, "grad_norm": 0.06307061073372533, "learning_rate": 2.6272266318987603e-05, "loss": 0.0213, "step": 1576 }, { "epoch": 2.36431784107946, "grad_norm": 0.038160947406252387, "learning_rate": 2.615446593741161e-05, "loss": 0.0113, "step": 1577 }, { "epoch": 2.365817091454273, "grad_norm": 0.05381521540641214, "learning_rate": 2.603689050213902e-05, "loss": 0.0159, "step": 1578 }, { "epoch": 2.3673163418290857, "grad_norm": 0.07567632627062996, "learning_rate": 2.5919540371325e-05, "loss": 0.0218, "step": 1579 }, { "epoch": 2.368815592203898, "grad_norm": 0.04343802349186751, "learning_rate": 2.580241590243837e-05, "loss": 0.0069, "step": 1580 }, { "epoch": 2.370314842578711, "grad_norm": 0.035981995121494297, "learning_rate": 2.5685517452260567e-05, "loss": 0.0094, "step": 1581 }, { "epoch": 2.371814092953523, "grad_norm": 0.03327317547759289, "learning_rate": 2.5568845376884587e-05, "loss": 0.0065, "step": 1582 }, { "epoch": 2.373313343328336, "grad_norm": 0.04266424967115805, "learning_rate": 2.5452400031713785e-05, "loss": 0.0187, "step": 1583 }, { "epoch": 2.3748125937031483, "grad_norm": 0.04761399156659926, "learning_rate": 2.5336181771460876e-05, "loss": 0.0165, "step": 1584 }, { "epoch": 2.376311844077961, "grad_norm": 0.048747473323202256, "learning_rate": 2.5220190950146827e-05, "loss": 0.01, "step": 1585 }, { "epoch": 2.3778110944527735, "grad_norm": 0.039981752669350695, "learning_rate": 2.5104427921099782e-05, "loss": 0.0052, "step": 1586 }, { "epoch": 2.3793103448275863, "grad_norm": 0.04586322792145994, "learning_rate": 2.4988893036954043e-05, "loss": 0.0131, "step": 1587 }, { "epoch": 2.3808095952023987, "grad_norm": 0.03289757882784411, "learning_rate": 2.4873586649648894e-05, "loss": 0.0076, "step": 1588 }, { "epoch": 2.3823088455772115, "grad_norm": 0.053525818232516974, "learning_rate": 2.4758509110427575e-05, "loss": 0.0092, "step": 1589 }, { "epoch": 2.383808095952024, "grad_norm": 0.0260906798066175, "learning_rate": 2.464366076983623e-05, "loss": 0.0067, "step": 1590 }, { "epoch": 2.3853073463268366, "grad_norm": 0.04852958702388942, "learning_rate": 2.45290419777228e-05, "loss": 0.0088, "step": 1591 }, { "epoch": 2.386806596701649, "grad_norm": 0.05708515361745962, "learning_rate": 2.441465308323605e-05, "loss": 0.0151, "step": 1592 }, { "epoch": 2.3883058470764618, "grad_norm": 0.06690475513808632, "learning_rate": 2.4300494434824373e-05, "loss": 0.0141, "step": 1593 }, { "epoch": 2.3898050974512746, "grad_norm": 0.07034082481319692, "learning_rate": 2.4186566380234798e-05, "loss": 0.0183, "step": 1594 }, { "epoch": 2.391304347826087, "grad_norm": 0.04379949104159517, "learning_rate": 2.407286926651192e-05, "loss": 0.0104, "step": 1595 }, { "epoch": 2.3928035982008997, "grad_norm": 0.03916460924167042, "learning_rate": 2.3959403439996907e-05, "loss": 0.0109, "step": 1596 }, { "epoch": 2.394302848575712, "grad_norm": 0.044700534764359666, "learning_rate": 2.3846169246326343e-05, "loss": 0.0098, "step": 1597 }, { "epoch": 2.395802098950525, "grad_norm": 0.031537306628073504, "learning_rate": 2.373316703043119e-05, "loss": 0.0064, "step": 1598 }, { "epoch": 2.3973013493253372, "grad_norm": 0.059228194972760356, "learning_rate": 2.362039713653581e-05, "loss": 0.0112, "step": 1599 }, { "epoch": 2.39880059970015, "grad_norm": 0.03603607951514331, "learning_rate": 2.3507859908156827e-05, "loss": 0.0109, "step": 1600 }, { "epoch": 2.4002998500749624, "grad_norm": 0.037249803689051716, "learning_rate": 2.339555568810221e-05, "loss": 0.0165, "step": 1601 }, { "epoch": 2.401799100449775, "grad_norm": 0.040614489764046136, "learning_rate": 2.328348481847006e-05, "loss": 0.0092, "step": 1602 }, { "epoch": 2.4032983508245875, "grad_norm": 0.04025495516343567, "learning_rate": 2.3171647640647687e-05, "loss": 0.0096, "step": 1603 }, { "epoch": 2.4047976011994003, "grad_norm": 0.02884411856762933, "learning_rate": 2.3060044495310505e-05, "loss": 0.0055, "step": 1604 }, { "epoch": 2.406296851574213, "grad_norm": 0.05045960092506823, "learning_rate": 2.2948675722421086e-05, "loss": 0.0158, "step": 1605 }, { "epoch": 2.4077961019490255, "grad_norm": 0.0551056436749728, "learning_rate": 2.2837541661228025e-05, "loss": 0.0119, "step": 1606 }, { "epoch": 2.409295352323838, "grad_norm": 0.04833036551816059, "learning_rate": 2.2726642650264895e-05, "loss": 0.0214, "step": 1607 }, { "epoch": 2.4107946026986506, "grad_norm": 0.07648766070075265, "learning_rate": 2.2615979027349387e-05, "loss": 0.0167, "step": 1608 }, { "epoch": 2.4122938530734634, "grad_norm": 0.029987044457999906, "learning_rate": 2.2505551129582047e-05, "loss": 0.0073, "step": 1609 }, { "epoch": 2.413793103448276, "grad_norm": 0.03597204491386506, "learning_rate": 2.2395359293345396e-05, "loss": 0.0065, "step": 1610 }, { "epoch": 2.4152923538230886, "grad_norm": 0.023778253711961315, "learning_rate": 2.2285403854302912e-05, "loss": 0.0065, "step": 1611 }, { "epoch": 2.416791604197901, "grad_norm": 0.0644646469541785, "learning_rate": 2.2175685147397906e-05, "loss": 0.0119, "step": 1612 }, { "epoch": 2.4182908545727138, "grad_norm": 0.05440033122122723, "learning_rate": 2.2066203506852566e-05, "loss": 0.0189, "step": 1613 }, { "epoch": 2.419790104947526, "grad_norm": 0.04000431572529469, "learning_rate": 2.195695926616702e-05, "loss": 0.0135, "step": 1614 }, { "epoch": 2.421289355322339, "grad_norm": 0.03921868468833917, "learning_rate": 2.1847952758118117e-05, "loss": 0.0122, "step": 1615 }, { "epoch": 2.4227886056971513, "grad_norm": 0.037836915480475734, "learning_rate": 2.173918431475861e-05, "loss": 0.0144, "step": 1616 }, { "epoch": 2.424287856071964, "grad_norm": 0.03003865685433883, "learning_rate": 2.163065426741603e-05, "loss": 0.0075, "step": 1617 }, { "epoch": 2.4257871064467764, "grad_norm": 0.07674219516894219, "learning_rate": 2.1522362946691698e-05, "loss": 0.0198, "step": 1618 }, { "epoch": 2.427286356821589, "grad_norm": 0.04221774862827237, "learning_rate": 2.1414310682459802e-05, "loss": 0.0072, "step": 1619 }, { "epoch": 2.428785607196402, "grad_norm": 0.036107728447124024, "learning_rate": 2.1306497803866277e-05, "loss": 0.0226, "step": 1620 }, { "epoch": 2.4302848575712144, "grad_norm": 0.03614753125579904, "learning_rate": 2.119892463932781e-05, "loss": 0.0072, "step": 1621 }, { "epoch": 2.431784107946027, "grad_norm": 0.05264305843875394, "learning_rate": 2.1091591516530952e-05, "loss": 0.0367, "step": 1622 }, { "epoch": 2.4332833583208395, "grad_norm": 0.041034139959893585, "learning_rate": 2.098449876243096e-05, "loss": 0.0104, "step": 1623 }, { "epoch": 2.4347826086956523, "grad_norm": 0.0642574057820137, "learning_rate": 2.0877646703250996e-05, "loss": 0.0232, "step": 1624 }, { "epoch": 2.4362818590704647, "grad_norm": 0.058453318712321, "learning_rate": 2.0771035664480942e-05, "loss": 0.0077, "step": 1625 }, { "epoch": 2.4377811094452775, "grad_norm": 0.05995496409834707, "learning_rate": 2.0664665970876496e-05, "loss": 0.0099, "step": 1626 }, { "epoch": 2.43928035982009, "grad_norm": 0.03844761053713044, "learning_rate": 2.0558537946458177e-05, "loss": 0.0076, "step": 1627 }, { "epoch": 2.4407796101949026, "grad_norm": 0.024867934503397734, "learning_rate": 2.045265191451041e-05, "loss": 0.0044, "step": 1628 }, { "epoch": 2.442278860569715, "grad_norm": 0.05311255008445868, "learning_rate": 2.0347008197580374e-05, "loss": 0.0191, "step": 1629 }, { "epoch": 2.443778110944528, "grad_norm": 0.031827800856509665, "learning_rate": 2.024160711747717e-05, "loss": 0.0065, "step": 1630 }, { "epoch": 2.44527736131934, "grad_norm": 0.05131745770511056, "learning_rate": 2.013644899527074e-05, "loss": 0.0133, "step": 1631 }, { "epoch": 2.446776611694153, "grad_norm": 0.07284238985049785, "learning_rate": 2.0031534151290943e-05, "loss": 0.0272, "step": 1632 }, { "epoch": 2.4482758620689653, "grad_norm": 0.04697325704656067, "learning_rate": 1.9926862905126665e-05, "loss": 0.014, "step": 1633 }, { "epoch": 2.449775112443778, "grad_norm": 0.048057744885183065, "learning_rate": 1.9822435575624608e-05, "loss": 0.0147, "step": 1634 }, { "epoch": 2.451274362818591, "grad_norm": 0.045793680272017155, "learning_rate": 1.9718252480888566e-05, "loss": 0.0134, "step": 1635 }, { "epoch": 2.4527736131934033, "grad_norm": 0.04667645820248889, "learning_rate": 1.9614313938278272e-05, "loss": 0.0132, "step": 1636 }, { "epoch": 2.454272863568216, "grad_norm": 0.05174429739295553, "learning_rate": 1.9510620264408596e-05, "loss": 0.0243, "step": 1637 }, { "epoch": 2.4557721139430284, "grad_norm": 0.110231925442238, "learning_rate": 1.9407171775148436e-05, "loss": 0.021, "step": 1638 }, { "epoch": 2.457271364317841, "grad_norm": 0.03434362044805778, "learning_rate": 1.930396878561983e-05, "loss": 0.0138, "step": 1639 }, { "epoch": 2.4587706146926536, "grad_norm": 0.06327466377623622, "learning_rate": 1.9201011610196973e-05, "loss": 0.0173, "step": 1640 }, { "epoch": 2.4602698650674664, "grad_norm": 0.038432132972600765, "learning_rate": 1.9098300562505266e-05, "loss": 0.0093, "step": 1641 }, { "epoch": 2.4617691154422787, "grad_norm": 0.03794215355045494, "learning_rate": 1.8995835955420416e-05, "loss": 0.007, "step": 1642 }, { "epoch": 2.4632683658170915, "grad_norm": 0.0767443242117149, "learning_rate": 1.8893618101067355e-05, "loss": 0.0115, "step": 1643 }, { "epoch": 2.464767616191904, "grad_norm": 0.03859797682525315, "learning_rate": 1.879164731081937e-05, "loss": 0.0084, "step": 1644 }, { "epoch": 2.4662668665667167, "grad_norm": 0.042954509808475476, "learning_rate": 1.8689923895297245e-05, "loss": 0.0108, "step": 1645 }, { "epoch": 2.4677661169415295, "grad_norm": 0.03498277093832187, "learning_rate": 1.858844816436809e-05, "loss": 0.0047, "step": 1646 }, { "epoch": 2.469265367316342, "grad_norm": 0.03199505586240192, "learning_rate": 1.848722042714457e-05, "loss": 0.004, "step": 1647 }, { "epoch": 2.470764617691154, "grad_norm": 0.04456743778913846, "learning_rate": 1.838624099198397e-05, "loss": 0.0146, "step": 1648 }, { "epoch": 2.472263868065967, "grad_norm": 0.04820916494385257, "learning_rate": 1.8285510166487152e-05, "loss": 0.0121, "step": 1649 }, { "epoch": 2.47376311844078, "grad_norm": 0.06060334655449111, "learning_rate": 1.818502825749764e-05, "loss": 0.0171, "step": 1650 }, { "epoch": 2.475262368815592, "grad_norm": 0.0379451423014228, "learning_rate": 1.808479557110081e-05, "loss": 0.0061, "step": 1651 }, { "epoch": 2.476761619190405, "grad_norm": 0.04195253494434152, "learning_rate": 1.7984812412622787e-05, "loss": 0.0134, "step": 1652 }, { "epoch": 2.4782608695652173, "grad_norm": 0.044679392990216345, "learning_rate": 1.78850790866296e-05, "loss": 0.0045, "step": 1653 }, { "epoch": 2.47976011994003, "grad_norm": 0.05100088676782655, "learning_rate": 1.7785595896926265e-05, "loss": 0.011, "step": 1654 }, { "epoch": 2.4812593703148424, "grad_norm": 0.06387968500101993, "learning_rate": 1.7686363146555805e-05, "loss": 0.0092, "step": 1655 }, { "epoch": 2.4827586206896552, "grad_norm": 0.06041373504290301, "learning_rate": 1.7587381137798432e-05, "loss": 0.0176, "step": 1656 }, { "epoch": 2.4842578710644676, "grad_norm": 0.08163580517734162, "learning_rate": 1.7488650172170496e-05, "loss": 0.0141, "step": 1657 }, { "epoch": 2.4857571214392804, "grad_norm": 0.036870335902452075, "learning_rate": 1.7390170550423625e-05, "loss": 0.0082, "step": 1658 }, { "epoch": 2.4872563718140928, "grad_norm": 0.040709622085519565, "learning_rate": 1.7291942572543807e-05, "loss": 0.0153, "step": 1659 }, { "epoch": 2.4887556221889056, "grad_norm": 0.04456648108859674, "learning_rate": 1.719396653775056e-05, "loss": 0.0074, "step": 1660 }, { "epoch": 2.4902548725637184, "grad_norm": 0.04377882755814446, "learning_rate": 1.7096242744495837e-05, "loss": 0.0103, "step": 1661 }, { "epoch": 2.4917541229385307, "grad_norm": 0.0729066633203347, "learning_rate": 1.6998771490463262e-05, "loss": 0.0334, "step": 1662 }, { "epoch": 2.4932533733133435, "grad_norm": 0.05241713761579378, "learning_rate": 1.690155307256719e-05, "loss": 0.0103, "step": 1663 }, { "epoch": 2.494752623688156, "grad_norm": 0.0467056862888581, "learning_rate": 1.680458778695174e-05, "loss": 0.0084, "step": 1664 }, { "epoch": 2.4962518740629687, "grad_norm": 0.07045163060731827, "learning_rate": 1.6707875928990058e-05, "loss": 0.0222, "step": 1665 }, { "epoch": 2.497751124437781, "grad_norm": 0.055528118440165704, "learning_rate": 1.661141779328319e-05, "loss": 0.008, "step": 1666 }, { "epoch": 2.499250374812594, "grad_norm": 0.035917683449783605, "learning_rate": 1.6515213673659357e-05, "loss": 0.0077, "step": 1667 }, { "epoch": 2.500749625187406, "grad_norm": 0.056362883617270515, "learning_rate": 1.6419263863172997e-05, "loss": 0.0154, "step": 1668 }, { "epoch": 2.502248875562219, "grad_norm": 0.04328275913067473, "learning_rate": 1.632356865410384e-05, "loss": 0.013, "step": 1669 }, { "epoch": 2.5037481259370313, "grad_norm": 0.044862866976074015, "learning_rate": 1.622812833795613e-05, "loss": 0.0141, "step": 1670 }, { "epoch": 2.505247376311844, "grad_norm": 0.051915962820213685, "learning_rate": 1.6132943205457606e-05, "loss": 0.0088, "step": 1671 }, { "epoch": 2.506746626686657, "grad_norm": 0.04516719200013124, "learning_rate": 1.6038013546558696e-05, "loss": 0.0091, "step": 1672 }, { "epoch": 2.5082458770614693, "grad_norm": 0.050036405327217245, "learning_rate": 1.5943339650431576e-05, "loss": 0.0106, "step": 1673 }, { "epoch": 2.5097451274362816, "grad_norm": 0.0457058330617825, "learning_rate": 1.5848921805469397e-05, "loss": 0.0092, "step": 1674 }, { "epoch": 2.5112443778110944, "grad_norm": 0.03342504987080694, "learning_rate": 1.5754760299285252e-05, "loss": 0.0091, "step": 1675 }, { "epoch": 2.5127436281859072, "grad_norm": 0.040420268265012875, "learning_rate": 1.566085541871145e-05, "loss": 0.0052, "step": 1676 }, { "epoch": 2.5142428785607196, "grad_norm": 0.06164059710875758, "learning_rate": 1.5567207449798515e-05, "loss": 0.0258, "step": 1677 }, { "epoch": 2.5157421289355324, "grad_norm": 0.057605709783369166, "learning_rate": 1.547381667781439e-05, "loss": 0.0289, "step": 1678 }, { "epoch": 2.5172413793103448, "grad_norm": 0.028167540589282527, "learning_rate": 1.538068338724361e-05, "loss": 0.0053, "step": 1679 }, { "epoch": 2.5187406296851576, "grad_norm": 0.0422278998534822, "learning_rate": 1.528780786178631e-05, "loss": 0.0193, "step": 1680 }, { "epoch": 2.52023988005997, "grad_norm": 0.03314812743069891, "learning_rate": 1.5195190384357404e-05, "loss": 0.0094, "step": 1681 }, { "epoch": 2.5217391304347827, "grad_norm": 0.06727909068694563, "learning_rate": 1.5102831237085857e-05, "loss": 0.0146, "step": 1682 }, { "epoch": 2.523238380809595, "grad_norm": 0.04004516716489981, "learning_rate": 1.5010730701313625e-05, "loss": 0.0142, "step": 1683 }, { "epoch": 2.524737631184408, "grad_norm": 0.0477522580374889, "learning_rate": 1.4918889057594876e-05, "loss": 0.0094, "step": 1684 }, { "epoch": 2.52623688155922, "grad_norm": 0.046202178677567866, "learning_rate": 1.4827306585695234e-05, "loss": 0.0134, "step": 1685 }, { "epoch": 2.527736131934033, "grad_norm": 0.06037049282034485, "learning_rate": 1.4735983564590783e-05, "loss": 0.0109, "step": 1686 }, { "epoch": 2.529235382308846, "grad_norm": 0.058032801247956796, "learning_rate": 1.4644920272467244e-05, "loss": 0.0112, "step": 1687 }, { "epoch": 2.530734632683658, "grad_norm": 0.04254433435164772, "learning_rate": 1.4554116986719257e-05, "loss": 0.0175, "step": 1688 }, { "epoch": 2.5322338830584705, "grad_norm": 0.025951630484904446, "learning_rate": 1.4463573983949341e-05, "loss": 0.0048, "step": 1689 }, { "epoch": 2.5337331334332833, "grad_norm": 0.08259144802291749, "learning_rate": 1.4373291539967182e-05, "loss": 0.0642, "step": 1690 }, { "epoch": 2.535232383808096, "grad_norm": 0.039457640169434316, "learning_rate": 1.4283269929788779e-05, "loss": 0.0121, "step": 1691 }, { "epoch": 2.5367316341829085, "grad_norm": 0.03558390085116224, "learning_rate": 1.4193509427635543e-05, "loss": 0.0114, "step": 1692 }, { "epoch": 2.5382308845577213, "grad_norm": 0.03415480880425091, "learning_rate": 1.4104010306933557e-05, "loss": 0.0062, "step": 1693 }, { "epoch": 2.5397301349325336, "grad_norm": 0.045796622099045996, "learning_rate": 1.4014772840312663e-05, "loss": 0.0077, "step": 1694 }, { "epoch": 2.5412293853073464, "grad_norm": 0.054148746586993535, "learning_rate": 1.3925797299605647e-05, "loss": 0.015, "step": 1695 }, { "epoch": 2.542728635682159, "grad_norm": 0.024421237093400897, "learning_rate": 1.3837083955847418e-05, "loss": 0.0045, "step": 1696 }, { "epoch": 2.5442278860569716, "grad_norm": 0.06197500818735937, "learning_rate": 1.3748633079274253e-05, "loss": 0.0131, "step": 1697 }, { "epoch": 2.545727136431784, "grad_norm": 0.05264163996373128, "learning_rate": 1.3660444939322836e-05, "loss": 0.0227, "step": 1698 }, { "epoch": 2.5472263868065967, "grad_norm": 0.08067636964906463, "learning_rate": 1.3572519804629536e-05, "loss": 0.0237, "step": 1699 }, { "epoch": 2.548725637181409, "grad_norm": 0.030322320592028634, "learning_rate": 1.3484857943029572e-05, "loss": 0.0081, "step": 1700 }, { "epoch": 2.550224887556222, "grad_norm": 0.032101259741781796, "learning_rate": 1.339745962155613e-05, "loss": 0.0068, "step": 1701 }, { "epoch": 2.5517241379310347, "grad_norm": 0.06722494549676583, "learning_rate": 1.3310325106439726e-05, "loss": 0.0138, "step": 1702 }, { "epoch": 2.553223388305847, "grad_norm": 0.056474569171435375, "learning_rate": 1.3223454663107172e-05, "loss": 0.0134, "step": 1703 }, { "epoch": 2.5547226386806594, "grad_norm": 0.03764220074385009, "learning_rate": 1.3136848556180892e-05, "loss": 0.0092, "step": 1704 }, { "epoch": 2.556221889055472, "grad_norm": 0.0453564205498934, "learning_rate": 1.30505070494781e-05, "loss": 0.0202, "step": 1705 }, { "epoch": 2.557721139430285, "grad_norm": 0.04478968627228144, "learning_rate": 1.296443040601003e-05, "loss": 0.0104, "step": 1706 }, { "epoch": 2.5592203898050974, "grad_norm": 0.035851435868513, "learning_rate": 1.2878618887981064e-05, "loss": 0.0068, "step": 1707 }, { "epoch": 2.56071964017991, "grad_norm": 0.035598452987037586, "learning_rate": 1.279307275678795e-05, "loss": 0.0075, "step": 1708 }, { "epoch": 2.5622188905547225, "grad_norm": 0.03387765166868693, "learning_rate": 1.2707792273019048e-05, "loss": 0.0107, "step": 1709 }, { "epoch": 2.5637181409295353, "grad_norm": 0.05148687122929166, "learning_rate": 1.262277769645348e-05, "loss": 0.01, "step": 1710 }, { "epoch": 2.5652173913043477, "grad_norm": 0.10743393215957228, "learning_rate": 1.2538029286060426e-05, "loss": 0.0112, "step": 1711 }, { "epoch": 2.5667166416791605, "grad_norm": 0.040261118251164116, "learning_rate": 1.2453547299998225e-05, "loss": 0.015, "step": 1712 }, { "epoch": 2.5682158920539733, "grad_norm": 0.03742339657842725, "learning_rate": 1.2369331995613665e-05, "loss": 0.0041, "step": 1713 }, { "epoch": 2.5697151424287856, "grad_norm": 0.04417334383768286, "learning_rate": 1.228538362944115e-05, "loss": 0.0106, "step": 1714 }, { "epoch": 2.571214392803598, "grad_norm": 0.07433484391924158, "learning_rate": 1.2201702457201947e-05, "loss": 0.02, "step": 1715 }, { "epoch": 2.572713643178411, "grad_norm": 0.036073343776736266, "learning_rate": 1.2118288733803473e-05, "loss": 0.0103, "step": 1716 }, { "epoch": 2.5742128935532236, "grad_norm": 0.03372355593898115, "learning_rate": 1.2035142713338366e-05, "loss": 0.0065, "step": 1717 }, { "epoch": 2.575712143928036, "grad_norm": 0.03392962930913605, "learning_rate": 1.19522646490838e-05, "loss": 0.0117, "step": 1718 }, { "epoch": 2.5772113943028487, "grad_norm": 0.029621354833537627, "learning_rate": 1.1869654793500784e-05, "loss": 0.0094, "step": 1719 }, { "epoch": 2.578710644677661, "grad_norm": 0.04044294739240826, "learning_rate": 1.1787313398233235e-05, "loss": 0.0107, "step": 1720 }, { "epoch": 2.580209895052474, "grad_norm": 0.056658167645708614, "learning_rate": 1.1705240714107302e-05, "loss": 0.0096, "step": 1721 }, { "epoch": 2.5817091454272862, "grad_norm": 0.05178402629396156, "learning_rate": 1.1623436991130654e-05, "loss": 0.019, "step": 1722 }, { "epoch": 2.583208395802099, "grad_norm": 0.03463124057838358, "learning_rate": 1.1541902478491606e-05, "loss": 0.0074, "step": 1723 }, { "epoch": 2.5847076461769114, "grad_norm": 0.03151218959720064, "learning_rate": 1.1460637424558407e-05, "loss": 0.0074, "step": 1724 }, { "epoch": 2.586206896551724, "grad_norm": 0.05462315276876222, "learning_rate": 1.1379642076878527e-05, "loss": 0.0292, "step": 1725 }, { "epoch": 2.5877061469265366, "grad_norm": 0.048365182176495165, "learning_rate": 1.129891668217783e-05, "loss": 0.0244, "step": 1726 }, { "epoch": 2.5892053973013494, "grad_norm": 0.037880615857319395, "learning_rate": 1.1218461486359877e-05, "loss": 0.0107, "step": 1727 }, { "epoch": 2.590704647676162, "grad_norm": 0.057564791499358094, "learning_rate": 1.1138276734505104e-05, "loss": 0.0136, "step": 1728 }, { "epoch": 2.5922038980509745, "grad_norm": 0.03993609994631937, "learning_rate": 1.1058362670870249e-05, "loss": 0.0101, "step": 1729 }, { "epoch": 2.593703148425787, "grad_norm": 0.036187330343662796, "learning_rate": 1.0978719538887349e-05, "loss": 0.0061, "step": 1730 }, { "epoch": 2.5952023988005997, "grad_norm": 0.042479518521181626, "learning_rate": 1.0899347581163221e-05, "loss": 0.0112, "step": 1731 }, { "epoch": 2.5967016491754125, "grad_norm": 0.03879025283345282, "learning_rate": 1.0820247039478604e-05, "loss": 0.0103, "step": 1732 }, { "epoch": 2.598200899550225, "grad_norm": 0.04141190230052251, "learning_rate": 1.0741418154787442e-05, "loss": 0.0077, "step": 1733 }, { "epoch": 2.5997001499250376, "grad_norm": 0.037063806848575076, "learning_rate": 1.0662861167216243e-05, "loss": 0.007, "step": 1734 }, { "epoch": 2.60119940029985, "grad_norm": 0.03828686541339917, "learning_rate": 1.0584576316063188e-05, "loss": 0.0056, "step": 1735 }, { "epoch": 2.6026986506746628, "grad_norm": 0.042964401840021114, "learning_rate": 1.0506563839797501e-05, "loss": 0.0099, "step": 1736 }, { "epoch": 2.604197901049475, "grad_norm": 0.04612342334630918, "learning_rate": 1.042882397605871e-05, "loss": 0.0071, "step": 1737 }, { "epoch": 2.605697151424288, "grad_norm": 0.047317886442380205, "learning_rate": 1.0351356961655945e-05, "loss": 0.0139, "step": 1738 }, { "epoch": 2.6071964017991007, "grad_norm": 0.03784689635872378, "learning_rate": 1.0274163032567163e-05, "loss": 0.0131, "step": 1739 }, { "epoch": 2.608695652173913, "grad_norm": 0.048587304701118564, "learning_rate": 1.0197242423938446e-05, "loss": 0.0154, "step": 1740 }, { "epoch": 2.6101949025487254, "grad_norm": 0.06914053190744007, "learning_rate": 1.0120595370083318e-05, "loss": 0.02, "step": 1741 }, { "epoch": 2.6116941529235382, "grad_norm": 0.05823729620646619, "learning_rate": 1.0044222104481971e-05, "loss": 0.015, "step": 1742 }, { "epoch": 2.613193403298351, "grad_norm": 0.05047639300436932, "learning_rate": 9.968122859780648e-06, "loss": 0.0098, "step": 1743 }, { "epoch": 2.6146926536731634, "grad_norm": 0.04274580971147598, "learning_rate": 9.892297867790845e-06, "loss": 0.0099, "step": 1744 }, { "epoch": 2.6161919040479757, "grad_norm": 0.029820467191873472, "learning_rate": 9.816747359488632e-06, "loss": 0.0043, "step": 1745 }, { "epoch": 2.6176911544227885, "grad_norm": 0.032573845045774534, "learning_rate": 9.74147156501396e-06, "loss": 0.0084, "step": 1746 }, { "epoch": 2.6191904047976013, "grad_norm": 0.05315778753440274, "learning_rate": 9.666470713669918e-06, "loss": 0.0454, "step": 1747 }, { "epoch": 2.6206896551724137, "grad_norm": 0.05133148060466583, "learning_rate": 9.591745033922173e-06, "loss": 0.0125, "step": 1748 }, { "epoch": 2.6221889055472265, "grad_norm": 0.04994252154552361, "learning_rate": 9.517294753398064e-06, "loss": 0.0105, "step": 1749 }, { "epoch": 2.623688155922039, "grad_norm": 0.03487915845198506, "learning_rate": 9.443120098886061e-06, "loss": 0.0063, "step": 1750 }, { "epoch": 2.6251874062968517, "grad_norm": 0.0330666641708233, "learning_rate": 9.369221296335006e-06, "loss": 0.0073, "step": 1751 }, { "epoch": 2.626686656671664, "grad_norm": 0.04110843585872846, "learning_rate": 9.295598570853514e-06, "loss": 0.0129, "step": 1752 }, { "epoch": 2.628185907046477, "grad_norm": 0.06731552981788465, "learning_rate": 9.222252146709142e-06, "loss": 0.0113, "step": 1753 }, { "epoch": 2.6296851574212896, "grad_norm": 0.034593184604678175, "learning_rate": 9.149182247327837e-06, "loss": 0.0143, "step": 1754 }, { "epoch": 2.631184407796102, "grad_norm": 0.04157072194043167, "learning_rate": 9.076389095293148e-06, "loss": 0.0129, "step": 1755 }, { "epoch": 2.6326836581709143, "grad_norm": 0.049987902316825865, "learning_rate": 9.00387291234569e-06, "loss": 0.0132, "step": 1756 }, { "epoch": 2.634182908545727, "grad_norm": 0.06968717000059639, "learning_rate": 8.931633919382298e-06, "loss": 0.0354, "step": 1757 }, { "epoch": 2.63568215892054, "grad_norm": 0.06538147492519782, "learning_rate": 8.85967233645547e-06, "loss": 0.014, "step": 1758 }, { "epoch": 2.6371814092953523, "grad_norm": 0.038320456379008105, "learning_rate": 8.787988382772705e-06, "loss": 0.0096, "step": 1759 }, { "epoch": 2.638680659670165, "grad_norm": 0.04286853776086437, "learning_rate": 8.716582276695728e-06, "loss": 0.0093, "step": 1760 }, { "epoch": 2.6401799100449774, "grad_norm": 0.04684067305397859, "learning_rate": 8.645454235739903e-06, "loss": 0.0081, "step": 1761 }, { "epoch": 2.6416791604197902, "grad_norm": 0.05534917471477078, "learning_rate": 8.574604476573621e-06, "loss": 0.0146, "step": 1762 }, { "epoch": 2.6431784107946026, "grad_norm": 0.1147630988897454, "learning_rate": 8.504033215017527e-06, "loss": 0.0244, "step": 1763 }, { "epoch": 2.6446776611694154, "grad_norm": 0.05391917398956416, "learning_rate": 8.433740666043898e-06, "loss": 0.0123, "step": 1764 }, { "epoch": 2.6461769115442277, "grad_norm": 0.04389897470604268, "learning_rate": 8.363727043776038e-06, "loss": 0.0109, "step": 1765 }, { "epoch": 2.6476761619190405, "grad_norm": 0.03600481781401776, "learning_rate": 8.293992561487596e-06, "loss": 0.0211, "step": 1766 }, { "epoch": 2.649175412293853, "grad_norm": 0.065614713593463, "learning_rate": 8.224537431601886e-06, "loss": 0.0117, "step": 1767 }, { "epoch": 2.6506746626686657, "grad_norm": 0.03443156843989127, "learning_rate": 8.15536186569129e-06, "loss": 0.0125, "step": 1768 }, { "epoch": 2.6521739130434785, "grad_norm": 0.04909539497158153, "learning_rate": 8.086466074476563e-06, "loss": 0.0156, "step": 1769 }, { "epoch": 2.653673163418291, "grad_norm": 0.04289750280283566, "learning_rate": 8.017850267826232e-06, "loss": 0.0118, "step": 1770 }, { "epoch": 2.655172413793103, "grad_norm": 0.04777349214571409, "learning_rate": 7.949514654755962e-06, "loss": 0.019, "step": 1771 }, { "epoch": 2.656671664167916, "grad_norm": 0.04182182875761444, "learning_rate": 7.881459443427886e-06, "loss": 0.0135, "step": 1772 }, { "epoch": 2.658170914542729, "grad_norm": 0.03102000250987136, "learning_rate": 7.81368484114996e-06, "loss": 0.0103, "step": 1773 }, { "epoch": 2.659670164917541, "grad_norm": 0.034781231775033394, "learning_rate": 7.746191054375362e-06, "loss": 0.0124, "step": 1774 }, { "epoch": 2.661169415292354, "grad_norm": 0.03923775864833756, "learning_rate": 7.67897828870191e-06, "loss": 0.01, "step": 1775 }, { "epoch": 2.6626686656671663, "grad_norm": 0.0642892146750047, "learning_rate": 7.612046748871327e-06, "loss": 0.0119, "step": 1776 }, { "epoch": 2.664167916041979, "grad_norm": 0.0447824761347899, "learning_rate": 7.545396638768698e-06, "loss": 0.0108, "step": 1777 }, { "epoch": 2.6656671664167915, "grad_norm": 0.05105597909092663, "learning_rate": 7.479028161421797e-06, "loss": 0.0073, "step": 1778 }, { "epoch": 2.6671664167916043, "grad_norm": 0.050486427170508684, "learning_rate": 7.412941519000527e-06, "loss": 0.0163, "step": 1779 }, { "epoch": 2.668665667166417, "grad_norm": 0.038036072252586865, "learning_rate": 7.347136912816277e-06, "loss": 0.0166, "step": 1780 }, { "epoch": 2.6701649175412294, "grad_norm": 0.03715585736405845, "learning_rate": 7.281614543321269e-06, "loss": 0.0087, "step": 1781 }, { "epoch": 2.6716641679160418, "grad_norm": 0.04320048869501502, "learning_rate": 7.216374610108012e-06, "loss": 0.0137, "step": 1782 }, { "epoch": 2.6731634182908546, "grad_norm": 0.04494460270144113, "learning_rate": 7.151417311908648e-06, "loss": 0.0154, "step": 1783 }, { "epoch": 2.6746626686656674, "grad_norm": 0.07810748014534971, "learning_rate": 7.086742846594385e-06, "loss": 0.0152, "step": 1784 }, { "epoch": 2.6761619190404797, "grad_norm": 0.03839261813654216, "learning_rate": 7.022351411174866e-06, "loss": 0.0106, "step": 1785 }, { "epoch": 2.677661169415292, "grad_norm": 0.036246667899462175, "learning_rate": 6.958243201797554e-06, "loss": 0.0111, "step": 1786 }, { "epoch": 2.679160419790105, "grad_norm": 0.04340569330273567, "learning_rate": 6.894418413747183e-06, "loss": 0.0091, "step": 1787 }, { "epoch": 2.6806596701649177, "grad_norm": 0.048415298718281985, "learning_rate": 6.830877241445111e-06, "loss": 0.0057, "step": 1788 }, { "epoch": 2.68215892053973, "grad_norm": 0.05732039596715092, "learning_rate": 6.767619878448783e-06, "loss": 0.0131, "step": 1789 }, { "epoch": 2.683658170914543, "grad_norm": 0.04379961391087932, "learning_rate": 6.704646517451107e-06, "loss": 0.0111, "step": 1790 }, { "epoch": 2.685157421289355, "grad_norm": 0.036728589602393405, "learning_rate": 6.6419573502798374e-06, "loss": 0.0096, "step": 1791 }, { "epoch": 2.686656671664168, "grad_norm": 0.0260401880716117, "learning_rate": 6.579552567897051e-06, "loss": 0.0044, "step": 1792 }, { "epoch": 2.6881559220389803, "grad_norm": 0.033806979051743735, "learning_rate": 6.517432360398556e-06, "loss": 0.0179, "step": 1793 }, { "epoch": 2.689655172413793, "grad_norm": 0.036411546274605675, "learning_rate": 6.455596917013273e-06, "loss": 0.0064, "step": 1794 }, { "epoch": 2.691154422788606, "grad_norm": 0.056313097870667905, "learning_rate": 6.394046426102674e-06, "loss": 0.0116, "step": 1795 }, { "epoch": 2.6926536731634183, "grad_norm": 0.02955564277336811, "learning_rate": 6.332781075160243e-06, "loss": 0.0045, "step": 1796 }, { "epoch": 2.6941529235382307, "grad_norm": 0.07517964552445663, "learning_rate": 6.2718010508108545e-06, "loss": 0.0385, "step": 1797 }, { "epoch": 2.6956521739130435, "grad_norm": 0.07054952867220826, "learning_rate": 6.21110653881023e-06, "loss": 0.0143, "step": 1798 }, { "epoch": 2.6971514242878563, "grad_norm": 0.02278968590892955, "learning_rate": 6.1506977240444074e-06, "loss": 0.0025, "step": 1799 }, { "epoch": 2.6986506746626686, "grad_norm": 0.07562322753054954, "learning_rate": 6.090574790529091e-06, "loss": 0.012, "step": 1800 }, { "epoch": 2.7001499250374814, "grad_norm": 0.03915044026809977, "learning_rate": 6.030737921409169e-06, "loss": 0.0082, "step": 1801 }, { "epoch": 2.7016491754122938, "grad_norm": 0.04026227035700397, "learning_rate": 5.971187298958103e-06, "loss": 0.014, "step": 1802 }, { "epoch": 2.7031484257871066, "grad_norm": 0.05329942920332837, "learning_rate": 5.911923104577455e-06, "loss": 0.0243, "step": 1803 }, { "epoch": 2.704647676161919, "grad_norm": 0.042274055914819364, "learning_rate": 5.852945518796205e-06, "loss": 0.0102, "step": 1804 }, { "epoch": 2.7061469265367317, "grad_norm": 0.039769102199125106, "learning_rate": 5.7942547212703315e-06, "loss": 0.0105, "step": 1805 }, { "epoch": 2.707646176911544, "grad_norm": 0.03079199853425841, "learning_rate": 5.735850890782157e-06, "loss": 0.0108, "step": 1806 }, { "epoch": 2.709145427286357, "grad_norm": 0.05225735394957531, "learning_rate": 5.6777342052399045e-06, "loss": 0.0227, "step": 1807 }, { "epoch": 2.7106446776611692, "grad_norm": 0.039410722770702566, "learning_rate": 5.619904841677059e-06, "loss": 0.0082, "step": 1808 }, { "epoch": 2.712143928035982, "grad_norm": 0.05923834687684238, "learning_rate": 5.562362976251901e-06, "loss": 0.0319, "step": 1809 }, { "epoch": 2.713643178410795, "grad_norm": 0.03704362252483329, "learning_rate": 5.505108784246926e-06, "loss": 0.0096, "step": 1810 }, { "epoch": 2.715142428785607, "grad_norm": 0.05704483825936286, "learning_rate": 5.448142440068316e-06, "loss": 0.0075, "step": 1811 }, { "epoch": 2.7166416791604195, "grad_norm": 0.032415598037398075, "learning_rate": 5.39146411724547e-06, "loss": 0.0059, "step": 1812 }, { "epoch": 2.7181409295352323, "grad_norm": 0.04448043782757098, "learning_rate": 5.335073988430372e-06, "loss": 0.0134, "step": 1813 }, { "epoch": 2.719640179910045, "grad_norm": 0.04721482806994072, "learning_rate": 5.278972225397127e-06, "loss": 0.0174, "step": 1814 }, { "epoch": 2.7211394302848575, "grad_norm": 0.06187954508185074, "learning_rate": 5.223158999041444e-06, "loss": 0.0094, "step": 1815 }, { "epoch": 2.7226386806596703, "grad_norm": 0.04072401244893839, "learning_rate": 5.167634479380068e-06, "loss": 0.0091, "step": 1816 }, { "epoch": 2.7241379310344827, "grad_norm": 0.060146327033198416, "learning_rate": 5.1123988355503475e-06, "loss": 0.0157, "step": 1817 }, { "epoch": 2.7256371814092955, "grad_norm": 0.042979415062188345, "learning_rate": 5.057452235809624e-06, "loss": 0.0134, "step": 1818 }, { "epoch": 2.727136431784108, "grad_norm": 0.05208115661281863, "learning_rate": 5.002794847534764e-06, "loss": 0.0178, "step": 1819 }, { "epoch": 2.7286356821589206, "grad_norm": 0.037556004650360325, "learning_rate": 4.948426837221631e-06, "loss": 0.0132, "step": 1820 }, { "epoch": 2.7301349325337334, "grad_norm": 0.06584481168797356, "learning_rate": 4.8943483704846475e-06, "loss": 0.0212, "step": 1821 }, { "epoch": 2.7316341829085458, "grad_norm": 0.045335104655996494, "learning_rate": 4.840559612056183e-06, "loss": 0.0136, "step": 1822 }, { "epoch": 2.733133433283358, "grad_norm": 0.040007627404927436, "learning_rate": 4.7870607257861415e-06, "loss": 0.0154, "step": 1823 }, { "epoch": 2.734632683658171, "grad_norm": 0.05263971747375208, "learning_rate": 4.733851874641382e-06, "loss": 0.0121, "step": 1824 }, { "epoch": 2.7361319340329837, "grad_norm": 0.032287000653965, "learning_rate": 4.680933220705308e-06, "loss": 0.0124, "step": 1825 }, { "epoch": 2.737631184407796, "grad_norm": 0.03915014390264884, "learning_rate": 4.628304925177318e-06, "loss": 0.0105, "step": 1826 }, { "epoch": 2.7391304347826084, "grad_norm": 0.04634060905307104, "learning_rate": 4.575967148372317e-06, "loss": 0.0188, "step": 1827 }, { "epoch": 2.7406296851574212, "grad_norm": 0.05421939527734177, "learning_rate": 4.523920049720265e-06, "loss": 0.0365, "step": 1828 }, { "epoch": 2.742128935532234, "grad_norm": 0.041259117105296526, "learning_rate": 4.4721637877656375e-06, "loss": 0.01, "step": 1829 }, { "epoch": 2.7436281859070464, "grad_norm": 0.04327982215295031, "learning_rate": 4.420698520166988e-06, "loss": 0.0131, "step": 1830 }, { "epoch": 2.745127436281859, "grad_norm": 0.02980015816214023, "learning_rate": 4.369524403696457e-06, "loss": 0.0103, "step": 1831 }, { "epoch": 2.7466266866566715, "grad_norm": 0.05434799362597012, "learning_rate": 4.318641594239259e-06, "loss": 0.0078, "step": 1832 }, { "epoch": 2.7481259370314843, "grad_norm": 0.049891340313563876, "learning_rate": 4.268050246793276e-06, "loss": 0.0153, "step": 1833 }, { "epoch": 2.7496251874062967, "grad_norm": 0.056890926873373264, "learning_rate": 4.217750515468522e-06, "loss": 0.0107, "step": 1834 }, { "epoch": 2.7511244377811095, "grad_norm": 0.06475141113715238, "learning_rate": 4.167742553486675e-06, "loss": 0.0082, "step": 1835 }, { "epoch": 2.7526236881559223, "grad_norm": 0.07835675864803161, "learning_rate": 4.118026513180695e-06, "loss": 0.0121, "step": 1836 }, { "epoch": 2.7541229385307346, "grad_norm": 0.04135507086935162, "learning_rate": 4.068602545994249e-06, "loss": 0.0103, "step": 1837 }, { "epoch": 2.755622188905547, "grad_norm": 0.05907458281295751, "learning_rate": 4.019470802481307e-06, "loss": 0.0184, "step": 1838 }, { "epoch": 2.75712143928036, "grad_norm": 0.06866162598790887, "learning_rate": 3.970631432305694e-06, "loss": 0.0179, "step": 1839 }, { "epoch": 2.7586206896551726, "grad_norm": 0.046791981639323535, "learning_rate": 3.922084584240582e-06, "loss": 0.0148, "step": 1840 }, { "epoch": 2.760119940029985, "grad_norm": 0.03878652054957538, "learning_rate": 3.873830406168111e-06, "loss": 0.0069, "step": 1841 }, { "epoch": 2.7616191904047978, "grad_norm": 0.047360116489953655, "learning_rate": 3.825869045078867e-06, "loss": 0.0155, "step": 1842 }, { "epoch": 2.76311844077961, "grad_norm": 0.0552944209185456, "learning_rate": 3.7782006470714616e-06, "loss": 0.0092, "step": 1843 }, { "epoch": 2.764617691154423, "grad_norm": 0.046438722788018326, "learning_rate": 3.730825357352119e-06, "loss": 0.0058, "step": 1844 }, { "epoch": 2.7661169415292353, "grad_norm": 0.04634478473435023, "learning_rate": 3.68374332023419e-06, "loss": 0.007, "step": 1845 }, { "epoch": 2.767616191904048, "grad_norm": 0.08671161749407622, "learning_rate": 3.6369546791377052e-06, "loss": 0.0179, "step": 1846 }, { "epoch": 2.7691154422788604, "grad_norm": 0.03277404115960048, "learning_rate": 3.5904595765890005e-06, "loss": 0.0061, "step": 1847 }, { "epoch": 2.770614692653673, "grad_norm": 0.03467762097707111, "learning_rate": 3.5442581542201923e-06, "loss": 0.0079, "step": 1848 }, { "epoch": 2.7721139430284856, "grad_norm": 0.06453591656774282, "learning_rate": 3.4983505527688586e-06, "loss": 0.0113, "step": 1849 }, { "epoch": 2.7736131934032984, "grad_norm": 0.038371480429255554, "learning_rate": 3.452736912077503e-06, "loss": 0.0085, "step": 1850 }, { "epoch": 2.775112443778111, "grad_norm": 0.05088933410705421, "learning_rate": 3.40741737109318e-06, "loss": 0.0124, "step": 1851 }, { "epoch": 2.7766116941529235, "grad_norm": 0.05365402450828376, "learning_rate": 3.3623920678670597e-06, "loss": 0.0094, "step": 1852 }, { "epoch": 2.778110944527736, "grad_norm": 0.06433825349176588, "learning_rate": 3.3176611395540626e-06, "loss": 0.0218, "step": 1853 }, { "epoch": 2.7796101949025487, "grad_norm": 0.05478690192081729, "learning_rate": 3.273224722412327e-06, "loss": 0.0434, "step": 1854 }, { "epoch": 2.7811094452773615, "grad_norm": 0.031119522780049997, "learning_rate": 3.2290829518028862e-06, "loss": 0.0032, "step": 1855 }, { "epoch": 2.782608695652174, "grad_norm": 0.07248646957866157, "learning_rate": 3.1852359621892367e-06, "loss": 0.0126, "step": 1856 }, { "epoch": 2.7841079460269866, "grad_norm": 0.04506998953202122, "learning_rate": 3.1416838871368924e-06, "loss": 0.0132, "step": 1857 }, { "epoch": 2.785607196401799, "grad_norm": 0.0954270927188225, "learning_rate": 3.0984268593130528e-06, "loss": 0.0478, "step": 1858 }, { "epoch": 2.787106446776612, "grad_norm": 0.04707162322619969, "learning_rate": 3.0554650104861136e-06, "loss": 0.0126, "step": 1859 }, { "epoch": 2.788605697151424, "grad_norm": 0.05714555030108658, "learning_rate": 3.012798471525324e-06, "loss": 0.0134, "step": 1860 }, { "epoch": 2.790104947526237, "grad_norm": 0.025492994502564904, "learning_rate": 2.970427372400353e-06, "loss": 0.0049, "step": 1861 }, { "epoch": 2.7916041979010497, "grad_norm": 0.047324738517620916, "learning_rate": 2.928351842180921e-06, "loss": 0.0141, "step": 1862 }, { "epoch": 2.793103448275862, "grad_norm": 0.03796558848278133, "learning_rate": 2.8865720090364034e-06, "loss": 0.0063, "step": 1863 }, { "epoch": 2.7946026986506745, "grad_norm": 0.09223145975016753, "learning_rate": 2.845088000235396e-06, "loss": 0.0208, "step": 1864 }, { "epoch": 2.7961019490254873, "grad_norm": 0.02944884610745672, "learning_rate": 2.8038999421453826e-06, "loss": 0.005, "step": 1865 }, { "epoch": 2.7976011994003, "grad_norm": 0.03805115476289398, "learning_rate": 2.7630079602323442e-06, "loss": 0.0198, "step": 1866 }, { "epoch": 2.7991004497751124, "grad_norm": 0.045894195911977305, "learning_rate": 2.7224121790603517e-06, "loss": 0.0174, "step": 1867 }, { "epoch": 2.8005997001499248, "grad_norm": 0.0321259718686503, "learning_rate": 2.6821127222911857e-06, "loss": 0.006, "step": 1868 }, { "epoch": 2.8020989505247376, "grad_norm": 0.05824176413837105, "learning_rate": 2.6421097126839712e-06, "loss": 0.017, "step": 1869 }, { "epoch": 2.8035982008995504, "grad_norm": 0.03748203437706974, "learning_rate": 2.6024032720948443e-06, "loss": 0.0099, "step": 1870 }, { "epoch": 2.8050974512743627, "grad_norm": 0.046475582434025865, "learning_rate": 2.5629935214764865e-06, "loss": 0.0136, "step": 1871 }, { "epoch": 2.8065967016491755, "grad_norm": 0.03246901888350811, "learning_rate": 2.5238805808778242e-06, "loss": 0.0078, "step": 1872 }, { "epoch": 2.808095952023988, "grad_norm": 0.05771124749730396, "learning_rate": 2.4850645694436736e-06, "loss": 0.0163, "step": 1873 }, { "epoch": 2.8095952023988007, "grad_norm": 0.0808127898786902, "learning_rate": 2.446545605414341e-06, "loss": 0.015, "step": 1874 }, { "epoch": 2.811094452773613, "grad_norm": 0.06382369287303619, "learning_rate": 2.4083238061252567e-06, "loss": 0.0093, "step": 1875 }, { "epoch": 2.812593703148426, "grad_norm": 0.0504391572531248, "learning_rate": 2.3703992880066638e-06, "loss": 0.0132, "step": 1876 }, { "epoch": 2.8140929535232386, "grad_norm": 0.033216894153539456, "learning_rate": 2.332772166583208e-06, "loss": 0.006, "step": 1877 }, { "epoch": 2.815592203898051, "grad_norm": 0.05096305885370249, "learning_rate": 2.295442556473637e-06, "loss": 0.0174, "step": 1878 }, { "epoch": 2.8170914542728633, "grad_norm": 0.046175190454476704, "learning_rate": 2.2584105713904125e-06, "loss": 0.0133, "step": 1879 }, { "epoch": 2.818590704647676, "grad_norm": 0.0355997466646441, "learning_rate": 2.2216763241393767e-06, "loss": 0.0091, "step": 1880 }, { "epoch": 2.820089955022489, "grad_norm": 0.038485173123134625, "learning_rate": 2.1852399266194314e-06, "loss": 0.0141, "step": 1881 }, { "epoch": 2.8215892053973013, "grad_norm": 0.04108582616221667, "learning_rate": 2.1491014898221582e-06, "loss": 0.0091, "step": 1882 }, { "epoch": 2.823088455772114, "grad_norm": 0.032876936800594725, "learning_rate": 2.1132611238315003e-06, "loss": 0.0092, "step": 1883 }, { "epoch": 2.8245877061469264, "grad_norm": 0.05477658714603886, "learning_rate": 2.0777189378234143e-06, "loss": 0.01, "step": 1884 }, { "epoch": 2.8260869565217392, "grad_norm": 0.05136080138073991, "learning_rate": 2.0424750400655947e-06, "loss": 0.0178, "step": 1885 }, { "epoch": 2.8275862068965516, "grad_norm": 0.051708174632592536, "learning_rate": 2.0075295379170412e-06, "loss": 0.0195, "step": 1886 }, { "epoch": 2.8290854572713644, "grad_norm": 0.04229975085912354, "learning_rate": 1.9728825378278246e-06, "loss": 0.0102, "step": 1887 }, { "epoch": 2.8305847076461768, "grad_norm": 0.04996308582086851, "learning_rate": 1.9385341453386995e-06, "loss": 0.0074, "step": 1888 }, { "epoch": 2.8320839580209896, "grad_norm": 0.03789459557831022, "learning_rate": 1.904484465080847e-06, "loss": 0.012, "step": 1889 }, { "epoch": 2.833583208395802, "grad_norm": 0.08562598045821541, "learning_rate": 1.8707336007754873e-06, "loss": 0.0212, "step": 1890 }, { "epoch": 2.8350824587706147, "grad_norm": 0.06285167373390634, "learning_rate": 1.8372816552336026e-06, "loss": 0.0292, "step": 1891 }, { "epoch": 2.8365817091454275, "grad_norm": 0.050419940365418164, "learning_rate": 1.8041287303556364e-06, "loss": 0.0116, "step": 1892 }, { "epoch": 2.83808095952024, "grad_norm": 0.042285911613742205, "learning_rate": 1.771274927131139e-06, "loss": 0.0203, "step": 1893 }, { "epoch": 2.839580209895052, "grad_norm": 0.049022018563915075, "learning_rate": 1.7387203456384782e-06, "loss": 0.0145, "step": 1894 }, { "epoch": 2.841079460269865, "grad_norm": 0.045171833670496146, "learning_rate": 1.706465085044584e-06, "loss": 0.0129, "step": 1895 }, { "epoch": 2.842578710644678, "grad_norm": 0.039715327644927564, "learning_rate": 1.6745092436045494e-06, "loss": 0.0082, "step": 1896 }, { "epoch": 2.84407796101949, "grad_norm": 0.035305839810997684, "learning_rate": 1.6428529186614195e-06, "loss": 0.0057, "step": 1897 }, { "epoch": 2.845577211394303, "grad_norm": 0.02884773890983854, "learning_rate": 1.611496206645835e-06, "loss": 0.0074, "step": 1898 }, { "epoch": 2.8470764617691153, "grad_norm": 0.0574287405654098, "learning_rate": 1.580439203075812e-06, "loss": 0.012, "step": 1899 }, { "epoch": 2.848575712143928, "grad_norm": 0.052201211207994744, "learning_rate": 1.5496820025563409e-06, "loss": 0.0332, "step": 1900 }, { "epoch": 2.8500749625187405, "grad_norm": 0.04134364982677059, "learning_rate": 1.5192246987791981e-06, "loss": 0.0136, "step": 1901 }, { "epoch": 2.8515742128935533, "grad_norm": 0.05598242526509591, "learning_rate": 1.4890673845226133e-06, "loss": 0.015, "step": 1902 }, { "epoch": 2.853073463268366, "grad_norm": 0.04925546071669627, "learning_rate": 1.4592101516509914e-06, "loss": 0.0126, "step": 1903 }, { "epoch": 2.8545727136431784, "grad_norm": 0.04932675245740339, "learning_rate": 1.4296530911146466e-06, "loss": 0.0088, "step": 1904 }, { "epoch": 2.856071964017991, "grad_norm": 0.054060677512884445, "learning_rate": 1.400396292949513e-06, "loss": 0.0178, "step": 1905 }, { "epoch": 2.8575712143928036, "grad_norm": 0.06076418765692268, "learning_rate": 1.3714398462768563e-06, "loss": 0.0122, "step": 1906 }, { "epoch": 2.8590704647676164, "grad_norm": 0.03402198970460235, "learning_rate": 1.3427838393030633e-06, "loss": 0.0084, "step": 1907 }, { "epoch": 2.8605697151424287, "grad_norm": 0.04514894375645042, "learning_rate": 1.3144283593192752e-06, "loss": 0.0174, "step": 1908 }, { "epoch": 2.862068965517241, "grad_norm": 0.052315594232706215, "learning_rate": 1.2863734927012095e-06, "loss": 0.0126, "step": 1909 }, { "epoch": 2.863568215892054, "grad_norm": 0.0682162999204661, "learning_rate": 1.2586193249088608e-06, "loss": 0.0135, "step": 1910 }, { "epoch": 2.8650674662668667, "grad_norm": 0.12167475345782697, "learning_rate": 1.231165940486234e-06, "loss": 0.0405, "step": 1911 }, { "epoch": 2.866566716641679, "grad_norm": 0.05271276493946917, "learning_rate": 1.20401342306109e-06, "loss": 0.0118, "step": 1912 }, { "epoch": 2.868065967016492, "grad_norm": 0.04153845951568349, "learning_rate": 1.1771618553447216e-06, "loss": 0.0084, "step": 1913 }, { "epoch": 2.869565217391304, "grad_norm": 0.0619396612689843, "learning_rate": 1.1506113191316447e-06, "loss": 0.0136, "step": 1914 }, { "epoch": 2.871064467766117, "grad_norm": 0.049638642472926986, "learning_rate": 1.1243618952994195e-06, "loss": 0.0129, "step": 1915 }, { "epoch": 2.8725637181409294, "grad_norm": 0.03889444306335509, "learning_rate": 1.0984136638083177e-06, "loss": 0.0084, "step": 1916 }, { "epoch": 2.874062968515742, "grad_norm": 0.05884561350496651, "learning_rate": 1.0727667037011668e-06, "loss": 0.0201, "step": 1917 }, { "epoch": 2.875562218890555, "grad_norm": 0.05667222439249693, "learning_rate": 1.0474210931030514e-06, "loss": 0.0272, "step": 1918 }, { "epoch": 2.8770614692653673, "grad_norm": 0.0344349427646189, "learning_rate": 1.0223769092211012e-06, "loss": 0.0146, "step": 1919 }, { "epoch": 2.8785607196401797, "grad_norm": 0.04845218395463286, "learning_rate": 9.976342283442463e-07, "loss": 0.0099, "step": 1920 }, { "epoch": 2.8800599700149925, "grad_norm": 0.04516383648126532, "learning_rate": 9.731931258429638e-07, "loss": 0.0109, "step": 1921 }, { "epoch": 2.8815592203898053, "grad_norm": 0.039101125756830266, "learning_rate": 9.490536761691204e-07, "loss": 0.0127, "step": 1922 }, { "epoch": 2.8830584707646176, "grad_norm": 0.0348598580784241, "learning_rate": 9.252159528556403e-07, "loss": 0.0064, "step": 1923 }, { "epoch": 2.8845577211394304, "grad_norm": 0.056794174363373306, "learning_rate": 9.016800285163718e-07, "loss": 0.0134, "step": 1924 }, { "epoch": 2.886056971514243, "grad_norm": 0.04997582538816562, "learning_rate": 8.784459748458318e-07, "loss": 0.0107, "step": 1925 }, { "epoch": 2.8875562218890556, "grad_norm": 0.029698873545562304, "learning_rate": 8.555138626189618e-07, "loss": 0.0063, "step": 1926 }, { "epoch": 2.889055472263868, "grad_norm": 0.052799586718671646, "learning_rate": 8.328837616909613e-07, "loss": 0.0123, "step": 1927 }, { "epoch": 2.8905547226386807, "grad_norm": 0.060340451537960624, "learning_rate": 8.105557409970432e-07, "loss": 0.0058, "step": 1928 }, { "epoch": 2.892053973013493, "grad_norm": 0.04169678291317928, "learning_rate": 7.885298685522235e-07, "loss": 0.0131, "step": 1929 }, { "epoch": 2.893553223388306, "grad_norm": 0.031006864063695138, "learning_rate": 7.668062114511321e-07, "loss": 0.0062, "step": 1930 }, { "epoch": 2.8950524737631183, "grad_norm": 0.037579581122970544, "learning_rate": 7.453848358678017e-07, "loss": 0.0057, "step": 1931 }, { "epoch": 2.896551724137931, "grad_norm": 0.05444793834468081, "learning_rate": 7.242658070554464e-07, "loss": 0.0178, "step": 1932 }, { "epoch": 2.898050974512744, "grad_norm": 0.03736257572000215, "learning_rate": 7.034491893463058e-07, "loss": 0.014, "step": 1933 }, { "epoch": 2.899550224887556, "grad_norm": 0.050629155821355354, "learning_rate": 6.829350461514006e-07, "loss": 0.0141, "step": 1934 }, { "epoch": 2.9010494752623686, "grad_norm": 0.05873417473569679, "learning_rate": 6.627234399603555e-07, "loss": 0.0124, "step": 1935 }, { "epoch": 2.9025487256371814, "grad_norm": 0.034113384932770746, "learning_rate": 6.428144323412544e-07, "loss": 0.0077, "step": 1936 }, { "epoch": 2.904047976011994, "grad_norm": 0.03951450139165464, "learning_rate": 6.232080839403631e-07, "loss": 0.0136, "step": 1937 }, { "epoch": 2.9055472263868065, "grad_norm": 0.0609884773357238, "learning_rate": 6.039044544820404e-07, "loss": 0.0156, "step": 1938 }, { "epoch": 2.9070464767616193, "grad_norm": 0.07346959965077916, "learning_rate": 5.849036027684606e-07, "loss": 0.0137, "step": 1939 }, { "epoch": 2.9085457271364317, "grad_norm": 0.0396757737108472, "learning_rate": 5.662055866795357e-07, "loss": 0.0081, "step": 1940 }, { "epoch": 2.9100449775112445, "grad_norm": 0.03226220803926473, "learning_rate": 5.478104631726711e-07, "loss": 0.0078, "step": 1941 }, { "epoch": 2.911544227886057, "grad_norm": 0.061890838617855416, "learning_rate": 5.297182882825879e-07, "loss": 0.0285, "step": 1942 }, { "epoch": 2.9130434782608696, "grad_norm": 0.04818178406688492, "learning_rate": 5.119291171211793e-07, "loss": 0.0238, "step": 1943 }, { "epoch": 2.9145427286356824, "grad_norm": 0.03875041789389296, "learning_rate": 4.944430038773762e-07, "loss": 0.0083, "step": 1944 }, { "epoch": 2.9160419790104948, "grad_norm": 0.03421383965254478, "learning_rate": 4.772600018168816e-07, "loss": 0.0121, "step": 1945 }, { "epoch": 2.917541229385307, "grad_norm": 0.029783687728112736, "learning_rate": 4.6038016328211476e-07, "loss": 0.0098, "step": 1946 }, { "epoch": 2.91904047976012, "grad_norm": 0.04334039207917819, "learning_rate": 4.438035396920004e-07, "loss": 0.0166, "step": 1947 }, { "epoch": 2.9205397301349327, "grad_norm": 0.04401175252419814, "learning_rate": 4.275301815417909e-07, "loss": 0.0073, "step": 1948 }, { "epoch": 2.922038980509745, "grad_norm": 0.04209525296246078, "learning_rate": 4.115601384029666e-07, "loss": 0.0125, "step": 1949 }, { "epoch": 2.923538230884558, "grad_norm": 0.031899540772813685, "learning_rate": 3.958934589230467e-07, "loss": 0.0076, "step": 1950 }, { "epoch": 2.9250374812593702, "grad_norm": 0.04162407460425198, "learning_rate": 3.805301908254455e-07, "loss": 0.0134, "step": 1951 }, { "epoch": 2.926536731634183, "grad_norm": 0.02690287458247992, "learning_rate": 3.654703809093607e-07, "loss": 0.0053, "step": 1952 }, { "epoch": 2.9280359820089954, "grad_norm": 0.042896647658177786, "learning_rate": 3.50714075049563e-07, "loss": 0.006, "step": 1953 }, { "epoch": 2.929535232383808, "grad_norm": 0.0696480415710855, "learning_rate": 3.362613181963403e-07, "loss": 0.0243, "step": 1954 }, { "epoch": 2.9310344827586206, "grad_norm": 0.03572318317050368, "learning_rate": 3.2211215437528694e-07, "loss": 0.0105, "step": 1955 }, { "epoch": 2.9325337331334334, "grad_norm": 0.06683991641909554, "learning_rate": 3.0826662668720364e-07, "loss": 0.018, "step": 1956 }, { "epoch": 2.9340329835082457, "grad_norm": 0.0440778804252679, "learning_rate": 2.947247773079753e-07, "loss": 0.0128, "step": 1957 }, { "epoch": 2.9355322338830585, "grad_norm": 0.03918100651382146, "learning_rate": 2.81486647488427e-07, "loss": 0.0076, "step": 1958 }, { "epoch": 2.9370314842578713, "grad_norm": 0.037521901035567344, "learning_rate": 2.685522775541904e-07, "loss": 0.0123, "step": 1959 }, { "epoch": 2.9385307346326837, "grad_norm": 0.02751199505821018, "learning_rate": 2.5592170690560414e-07, "loss": 0.005, "step": 1960 }, { "epoch": 2.940029985007496, "grad_norm": 0.03734913153943532, "learning_rate": 2.4359497401758024e-07, "loss": 0.0094, "step": 1961 }, { "epoch": 2.941529235382309, "grad_norm": 0.05757025217118907, "learning_rate": 2.315721164394713e-07, "loss": 0.0183, "step": 1962 }, { "epoch": 2.9430284857571216, "grad_norm": 0.03149321116250488, "learning_rate": 2.1985317079500356e-07, "loss": 0.0085, "step": 1963 }, { "epoch": 2.944527736131934, "grad_norm": 0.03809558330191986, "learning_rate": 2.0843817278209942e-07, "loss": 0.0099, "step": 1964 }, { "epoch": 2.9460269865067468, "grad_norm": 0.04365279664933326, "learning_rate": 1.973271571728441e-07, "loss": 0.0131, "step": 1965 }, { "epoch": 2.947526236881559, "grad_norm": 0.06624961251115817, "learning_rate": 1.86520157813308e-07, "loss": 0.0085, "step": 1966 }, { "epoch": 2.949025487256372, "grad_norm": 0.05085077331950606, "learning_rate": 1.7601720762346897e-07, "loss": 0.0144, "step": 1967 }, { "epoch": 2.9505247376311843, "grad_norm": 0.04696188689378594, "learning_rate": 1.6581833859716788e-07, "loss": 0.0117, "step": 1968 }, { "epoch": 2.952023988005997, "grad_norm": 0.03716989035056519, "learning_rate": 1.5592358180189782e-07, "loss": 0.0076, "step": 1969 }, { "epoch": 2.9535232383808094, "grad_norm": 0.04255459156532792, "learning_rate": 1.4633296737882607e-07, "loss": 0.0095, "step": 1970 }, { "epoch": 2.9550224887556222, "grad_norm": 0.03752019994795145, "learning_rate": 1.3704652454261668e-07, "loss": 0.0059, "step": 1971 }, { "epoch": 2.9565217391304346, "grad_norm": 0.041200745192899624, "learning_rate": 1.2806428158138596e-07, "loss": 0.0091, "step": 1972 }, { "epoch": 2.9580209895052474, "grad_norm": 0.06162388641909606, "learning_rate": 1.193862658566025e-07, "loss": 0.0144, "step": 1973 }, { "epoch": 2.95952023988006, "grad_norm": 0.0589071951900631, "learning_rate": 1.1101250380300965e-07, "loss": 0.0291, "step": 1974 }, { "epoch": 2.9610194902548725, "grad_norm": 0.05002592563470208, "learning_rate": 1.0294302092853647e-07, "loss": 0.0119, "step": 1975 }, { "epoch": 2.962518740629685, "grad_norm": 0.030537467131411956, "learning_rate": 9.517784181422019e-08, "loss": 0.0079, "step": 1976 }, { "epoch": 2.9640179910044977, "grad_norm": 0.04020417423763941, "learning_rate": 8.771699011416168e-08, "loss": 0.006, "step": 1977 }, { "epoch": 2.9655172413793105, "grad_norm": 0.057186391102222636, "learning_rate": 8.056048855540343e-08, "loss": 0.0432, "step": 1978 }, { "epoch": 2.967016491754123, "grad_norm": 0.0641737485717754, "learning_rate": 7.370835893788508e-08, "loss": 0.017, "step": 1979 }, { "epoch": 2.9685157421289357, "grad_norm": 0.04127807538853771, "learning_rate": 6.71606221343768e-08, "loss": 0.0111, "step": 1980 }, { "epoch": 2.970014992503748, "grad_norm": 0.04366547017507273, "learning_rate": 6.09172980904238e-08, "loss": 0.0181, "step": 1981 }, { "epoch": 2.971514242878561, "grad_norm": 0.06613765452494803, "learning_rate": 5.497840582429081e-08, "loss": 0.0159, "step": 1982 }, { "epoch": 2.973013493253373, "grad_norm": 0.043680343247550456, "learning_rate": 4.934396342684e-08, "loss": 0.0098, "step": 1983 }, { "epoch": 2.974512743628186, "grad_norm": 0.035972782091150166, "learning_rate": 4.401398806159751e-08, "loss": 0.0096, "step": 1984 }, { "epoch": 2.9760119940029988, "grad_norm": 0.05290085181070818, "learning_rate": 3.898849596456478e-08, "loss": 0.0125, "step": 1985 }, { "epoch": 2.977511244377811, "grad_norm": 0.053528326356848906, "learning_rate": 3.4267502444274015e-08, "loss": 0.0083, "step": 1986 }, { "epoch": 2.9790104947526235, "grad_norm": 0.0388492299519708, "learning_rate": 2.985102188168831e-08, "loss": 0.0069, "step": 1987 }, { "epoch": 2.9805097451274363, "grad_norm": 0.0412347074122606, "learning_rate": 2.573906773016832e-08, "loss": 0.0096, "step": 1988 }, { "epoch": 2.982008995502249, "grad_norm": 0.05110960746173951, "learning_rate": 2.193165251545004e-08, "loss": 0.0248, "step": 1989 }, { "epoch": 2.9835082458770614, "grad_norm": 0.04776321571386053, "learning_rate": 1.842878783557822e-08, "loss": 0.0077, "step": 1990 }, { "epoch": 2.9850074962518742, "grad_norm": 0.017427881242845993, "learning_rate": 1.5230484360873044e-08, "loss": 0.0022, "step": 1991 }, { "epoch": 2.9865067466266866, "grad_norm": 0.046443204162746166, "learning_rate": 1.2336751833941229e-08, "loss": 0.0101, "step": 1992 }, { "epoch": 2.9880059970014994, "grad_norm": 0.05322748811455273, "learning_rate": 9.747599069576119e-09, "loss": 0.0349, "step": 1993 }, { "epoch": 2.9895052473763117, "grad_norm": 0.03808612444789204, "learning_rate": 7.463033954802078e-09, "loss": 0.0181, "step": 1994 }, { "epoch": 2.9910044977511245, "grad_norm": 0.0566683908530055, "learning_rate": 5.483063448785686e-09, "loss": 0.019, "step": 1995 }, { "epoch": 2.992503748125937, "grad_norm": 0.06682256849285167, "learning_rate": 3.807693582869032e-09, "loss": 0.0134, "step": 1996 }, { "epoch": 2.9940029985007497, "grad_norm": 0.04095732409855308, "learning_rate": 2.4369294605253166e-09, "loss": 0.0116, "step": 1997 }, { "epoch": 2.995502248875562, "grad_norm": 0.06718811977451918, "learning_rate": 1.3707752573255405e-09, "loss": 0.0145, "step": 1998 }, { "epoch": 2.997001499250375, "grad_norm": 0.058751406120225766, "learning_rate": 6.092342209607083e-10, "loss": 0.0137, "step": 1999 }, { "epoch": 2.9985007496251876, "grad_norm": 0.046584715389230986, "learning_rate": 1.5230867123072756e-10, "loss": 0.0114, "step": 2000 }, { "epoch": 3.0, "grad_norm": 0.048620158092767395, "learning_rate": 0.0, "loss": 0.009, "step": 2001 }, { "epoch": 3.0, "eval_loss": 0.04013560339808464, "eval_runtime": 96.8817, "eval_samples_per_second": 11.591, "eval_steps_per_second": 0.733, "step": 2001 }, { "epoch": 3.0, "step": 2001, "total_flos": 3988562654658560.0, "train_loss": 0.004510109983980358, "train_runtime": 5871.4929, "train_samples_per_second": 10.905, "train_steps_per_second": 0.341 } ], "logging_steps": 1, "max_steps": 2001, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3988562654658560.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }