{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9133519673188562, "eval_steps": 500, "global_step": 531, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0017200602021070737, "grad_norm": 9.863033294677734, "learning_rate": 1.0000000000000001e-07, "loss": 46.5684, "step": 1 }, { "epoch": 0.0034401204042141475, "grad_norm": 9.501362800598145, "learning_rate": 2.0000000000000002e-07, "loss": 46.7594, "step": 2 }, { "epoch": 0.005160180606321221, "grad_norm": 10.306024551391602, "learning_rate": 3.0000000000000004e-07, "loss": 47.7285, "step": 3 }, { "epoch": 0.006880240808428295, "grad_norm": 10.35928726196289, "learning_rate": 4.0000000000000003e-07, "loss": 46.2416, "step": 4 }, { "epoch": 0.008600301010535369, "grad_norm": 10.284590721130371, "learning_rate": 5.000000000000001e-07, "loss": 46.4638, "step": 5 }, { "epoch": 0.010320361212642442, "grad_norm": 9.508557319641113, "learning_rate": 6.000000000000001e-07, "loss": 47.0967, "step": 6 }, { "epoch": 0.012040421414749517, "grad_norm": 9.53107738494873, "learning_rate": 7.000000000000001e-07, "loss": 46.8945, "step": 7 }, { "epoch": 0.01376048161685659, "grad_norm": 8.917387008666992, "learning_rate": 8.000000000000001e-07, "loss": 46.0318, "step": 8 }, { "epoch": 0.015480541818963663, "grad_norm": 9.046576499938965, "learning_rate": 9.000000000000001e-07, "loss": 46.7626, "step": 9 }, { "epoch": 0.017200602021070738, "grad_norm": 8.346121788024902, "learning_rate": 1.0000000000000002e-06, "loss": 46.4023, "step": 10 }, { "epoch": 0.01892066222317781, "grad_norm": 7.368544578552246, "learning_rate": 1.1e-06, "loss": 46.9599, "step": 11 }, { "epoch": 0.020640722425284884, "grad_norm": 7.151632308959961, "learning_rate": 1.2000000000000002e-06, "loss": 46.8859, "step": 12 }, { "epoch": 0.022360782627391957, "grad_norm": 6.858343124389648, "learning_rate": 1.3e-06, "loss": 47.0505, "step": 13 }, { "epoch": 0.024080842829499034, "grad_norm": 6.4683918952941895, "learning_rate": 1.4000000000000001e-06, "loss": 46.6439, "step": 14 }, { "epoch": 0.025800903031606107, "grad_norm": 8.405121803283691, "learning_rate": 1.5e-06, "loss": 46.8736, "step": 15 }, { "epoch": 0.02752096323371318, "grad_norm": 10.191279411315918, "learning_rate": 1.6000000000000001e-06, "loss": 46.2945, "step": 16 }, { "epoch": 0.029241023435820253, "grad_norm": 10.24738597869873, "learning_rate": 1.7000000000000002e-06, "loss": 46.0618, "step": 17 }, { "epoch": 0.030961083637927326, "grad_norm": 9.236139297485352, "learning_rate": 1.8000000000000001e-06, "loss": 47.0773, "step": 18 }, { "epoch": 0.0326811438400344, "grad_norm": 8.437068939208984, "learning_rate": 1.9000000000000002e-06, "loss": 46.6797, "step": 19 }, { "epoch": 0.034401204042141476, "grad_norm": 8.35542106628418, "learning_rate": 2.0000000000000003e-06, "loss": 46.8348, "step": 20 }, { "epoch": 0.03612126424424855, "grad_norm": 9.210644721984863, "learning_rate": 2.1000000000000002e-06, "loss": 46.7521, "step": 21 }, { "epoch": 0.03784132444635562, "grad_norm": 9.216105461120605, "learning_rate": 2.2e-06, "loss": 46.4541, "step": 22 }, { "epoch": 0.0395613846484627, "grad_norm": 8.33311939239502, "learning_rate": 2.3000000000000004e-06, "loss": 46.7192, "step": 23 }, { "epoch": 0.04128144485056977, "grad_norm": 7.9267449378967285, "learning_rate": 2.4000000000000003e-06, "loss": 46.7559, "step": 24 }, { "epoch": 0.043001505052676844, "grad_norm": 7.363073348999023, "learning_rate": 2.5e-06, "loss": 45.9578, "step": 25 }, { "epoch": 0.044721565254783914, "grad_norm": 7.073836326599121, "learning_rate": 2.6e-06, "loss": 45.9155, "step": 26 }, { "epoch": 0.04644162545689099, "grad_norm": 7.0409088134765625, "learning_rate": 2.7000000000000004e-06, "loss": 46.6551, "step": 27 }, { "epoch": 0.04816168565899807, "grad_norm": 6.685385227203369, "learning_rate": 2.8000000000000003e-06, "loss": 46.3889, "step": 28 }, { "epoch": 0.04988174586110514, "grad_norm": 6.540524959564209, "learning_rate": 2.9e-06, "loss": 47.0662, "step": 29 }, { "epoch": 0.05160180606321221, "grad_norm": 6.627730846405029, "learning_rate": 3e-06, "loss": 46.101, "step": 30 }, { "epoch": 0.05332186626531928, "grad_norm": 7.054595470428467, "learning_rate": 3.1000000000000004e-06, "loss": 46.7379, "step": 31 }, { "epoch": 0.05504192646742636, "grad_norm": 6.73618221282959, "learning_rate": 3.2000000000000003e-06, "loss": 46.3172, "step": 32 }, { "epoch": 0.056761986669533436, "grad_norm": 5.943539619445801, "learning_rate": 3.3000000000000006e-06, "loss": 47.0245, "step": 33 }, { "epoch": 0.058482046871640506, "grad_norm": 6.056912899017334, "learning_rate": 3.4000000000000005e-06, "loss": 46.0208, "step": 34 }, { "epoch": 0.06020210707374758, "grad_norm": 5.7798309326171875, "learning_rate": 3.5e-06, "loss": 46.0746, "step": 35 }, { "epoch": 0.06192216727585465, "grad_norm": 5.896692276000977, "learning_rate": 3.6000000000000003e-06, "loss": 46.4911, "step": 36 }, { "epoch": 0.06364222747796174, "grad_norm": 5.496098518371582, "learning_rate": 3.7e-06, "loss": 45.9957, "step": 37 }, { "epoch": 0.0653622876800688, "grad_norm": 5.253308296203613, "learning_rate": 3.8000000000000005e-06, "loss": 45.7145, "step": 38 }, { "epoch": 0.06708234788217587, "grad_norm": 5.340756416320801, "learning_rate": 3.900000000000001e-06, "loss": 46.7068, "step": 39 }, { "epoch": 0.06880240808428295, "grad_norm": 5.312371730804443, "learning_rate": 4.000000000000001e-06, "loss": 46.4172, "step": 40 }, { "epoch": 0.07052246828639003, "grad_norm": 5.484511375427246, "learning_rate": 4.1e-06, "loss": 45.6433, "step": 41 }, { "epoch": 0.0722425284884971, "grad_norm": 5.260024547576904, "learning_rate": 4.2000000000000004e-06, "loss": 46.1259, "step": 42 }, { "epoch": 0.07396258869060417, "grad_norm": 5.4440999031066895, "learning_rate": 4.3e-06, "loss": 46.2947, "step": 43 }, { "epoch": 0.07568264889271124, "grad_norm": 5.153471946716309, "learning_rate": 4.4e-06, "loss": 46.2882, "step": 44 }, { "epoch": 0.07740270909481832, "grad_norm": 5.6796488761901855, "learning_rate": 4.5e-06, "loss": 46.3314, "step": 45 }, { "epoch": 0.0791227692969254, "grad_norm": 5.253461837768555, "learning_rate": 4.600000000000001e-06, "loss": 46.4205, "step": 46 }, { "epoch": 0.08084282949903247, "grad_norm": 5.475714683532715, "learning_rate": 4.7e-06, "loss": 45.9359, "step": 47 }, { "epoch": 0.08256288970113954, "grad_norm": 5.880488872528076, "learning_rate": 4.800000000000001e-06, "loss": 46.0665, "step": 48 }, { "epoch": 0.08428294990324661, "grad_norm": 5.69149923324585, "learning_rate": 4.9000000000000005e-06, "loss": 45.8811, "step": 49 }, { "epoch": 0.08600301010535369, "grad_norm": 5.086974143981934, "learning_rate": 5e-06, "loss": 46.1759, "step": 50 }, { "epoch": 0.08772307030746077, "grad_norm": 5.902041435241699, "learning_rate": 5.1e-06, "loss": 46.4326, "step": 51 }, { "epoch": 0.08944313050956783, "grad_norm": 5.740163803100586, "learning_rate": 5.2e-06, "loss": 45.607, "step": 52 }, { "epoch": 0.0911631907116749, "grad_norm": 5.547687530517578, "learning_rate": 5.300000000000001e-06, "loss": 45.8864, "step": 53 }, { "epoch": 0.09288325091378198, "grad_norm": 6.20143461227417, "learning_rate": 5.400000000000001e-06, "loss": 45.7776, "step": 54 }, { "epoch": 0.09460331111588906, "grad_norm": 5.544395446777344, "learning_rate": 5.500000000000001e-06, "loss": 46.3334, "step": 55 }, { "epoch": 0.09632337131799613, "grad_norm": 5.583694934844971, "learning_rate": 5.600000000000001e-06, "loss": 45.5733, "step": 56 }, { "epoch": 0.0980434315201032, "grad_norm": 5.751343250274658, "learning_rate": 5.7e-06, "loss": 45.2805, "step": 57 }, { "epoch": 0.09976349172221027, "grad_norm": 6.024663925170898, "learning_rate": 5.8e-06, "loss": 45.6569, "step": 58 }, { "epoch": 0.10148355192431735, "grad_norm": 5.834673881530762, "learning_rate": 5.9e-06, "loss": 45.7854, "step": 59 }, { "epoch": 0.10320361212642443, "grad_norm": 6.796127796173096, "learning_rate": 6e-06, "loss": 46.4184, "step": 60 }, { "epoch": 0.1049236723285315, "grad_norm": 6.303890705108643, "learning_rate": 6.1e-06, "loss": 46.0877, "step": 61 }, { "epoch": 0.10664373253063857, "grad_norm": 6.320569038391113, "learning_rate": 6.200000000000001e-06, "loss": 45.8802, "step": 62 }, { "epoch": 0.10836379273274564, "grad_norm": 6.015869617462158, "learning_rate": 6.300000000000001e-06, "loss": 45.6733, "step": 63 }, { "epoch": 0.11008385293485272, "grad_norm": 6.314846515655518, "learning_rate": 6.4000000000000006e-06, "loss": 45.8531, "step": 64 }, { "epoch": 0.1118039131369598, "grad_norm": 6.106888294219971, "learning_rate": 6.5000000000000004e-06, "loss": 45.2966, "step": 65 }, { "epoch": 0.11352397333906687, "grad_norm": 6.093003273010254, "learning_rate": 6.600000000000001e-06, "loss": 46.0552, "step": 66 }, { "epoch": 0.11524403354117393, "grad_norm": 5.634897708892822, "learning_rate": 6.700000000000001e-06, "loss": 45.3367, "step": 67 }, { "epoch": 0.11696409374328101, "grad_norm": 6.261721134185791, "learning_rate": 6.800000000000001e-06, "loss": 45.7655, "step": 68 }, { "epoch": 0.11868415394538809, "grad_norm": 5.926329135894775, "learning_rate": 6.9e-06, "loss": 45.1391, "step": 69 }, { "epoch": 0.12040421414749516, "grad_norm": 5.887923717498779, "learning_rate": 7e-06, "loss": 45.43, "step": 70 }, { "epoch": 0.12212427434960224, "grad_norm": 5.79194450378418, "learning_rate": 7.100000000000001e-06, "loss": 45.763, "step": 71 }, { "epoch": 0.1238443345517093, "grad_norm": 6.2670722007751465, "learning_rate": 7.2000000000000005e-06, "loss": 45.137, "step": 72 }, { "epoch": 0.1255643947538164, "grad_norm": 5.60247278213501, "learning_rate": 7.3e-06, "loss": 45.7123, "step": 73 }, { "epoch": 0.12728445495592347, "grad_norm": 7.346188545227051, "learning_rate": 7.4e-06, "loss": 45.5004, "step": 74 }, { "epoch": 0.12900451515803052, "grad_norm": 6.078243255615234, "learning_rate": 7.500000000000001e-06, "loss": 45.5731, "step": 75 }, { "epoch": 0.1307245753601376, "grad_norm": 7.248181343078613, "learning_rate": 7.600000000000001e-06, "loss": 45.7506, "step": 76 }, { "epoch": 0.13244463556224467, "grad_norm": 5.837612628936768, "learning_rate": 7.7e-06, "loss": 45.7486, "step": 77 }, { "epoch": 0.13416469576435175, "grad_norm": 7.661535739898682, "learning_rate": 7.800000000000002e-06, "loss": 45.5743, "step": 78 }, { "epoch": 0.13588475596645883, "grad_norm": 5.925168991088867, "learning_rate": 7.9e-06, "loss": 45.0878, "step": 79 }, { "epoch": 0.1376048161685659, "grad_norm": 6.47566032409668, "learning_rate": 8.000000000000001e-06, "loss": 45.4567, "step": 80 }, { "epoch": 0.13932487637067298, "grad_norm": 5.9641852378845215, "learning_rate": 8.1e-06, "loss": 45.0619, "step": 81 }, { "epoch": 0.14104493657278006, "grad_norm": 6.4052324295043945, "learning_rate": 8.2e-06, "loss": 45.9375, "step": 82 }, { "epoch": 0.14276499677488713, "grad_norm": 6.149839401245117, "learning_rate": 8.3e-06, "loss": 45.5695, "step": 83 }, { "epoch": 0.1444850569769942, "grad_norm": 6.546520233154297, "learning_rate": 8.400000000000001e-06, "loss": 45.8453, "step": 84 }, { "epoch": 0.14620511717910126, "grad_norm": 6.221002578735352, "learning_rate": 8.5e-06, "loss": 44.9717, "step": 85 }, { "epoch": 0.14792517738120833, "grad_norm": 6.772427558898926, "learning_rate": 8.6e-06, "loss": 45.8218, "step": 86 }, { "epoch": 0.1496452375833154, "grad_norm": 7.193835258483887, "learning_rate": 8.700000000000001e-06, "loss": 45.4445, "step": 87 }, { "epoch": 0.1513652977854225, "grad_norm": 6.1274895668029785, "learning_rate": 8.8e-06, "loss": 45.127, "step": 88 }, { "epoch": 0.15308535798752956, "grad_norm": 7.984500885009766, "learning_rate": 8.900000000000001e-06, "loss": 44.9928, "step": 89 }, { "epoch": 0.15480541818963664, "grad_norm": 6.346322059631348, "learning_rate": 9e-06, "loss": 45.4616, "step": 90 }, { "epoch": 0.15652547839174372, "grad_norm": 7.460362911224365, "learning_rate": 9.100000000000001e-06, "loss": 45.2488, "step": 91 }, { "epoch": 0.1582455385938508, "grad_norm": 7.510838508605957, "learning_rate": 9.200000000000002e-06, "loss": 45.0087, "step": 92 }, { "epoch": 0.15996559879595787, "grad_norm": 6.567174911499023, "learning_rate": 9.3e-06, "loss": 44.3216, "step": 93 }, { "epoch": 0.16168565899806495, "grad_norm": 8.357160568237305, "learning_rate": 9.4e-06, "loss": 45.1102, "step": 94 }, { "epoch": 0.163405719200172, "grad_norm": 6.190718650817871, "learning_rate": 9.5e-06, "loss": 45.0187, "step": 95 }, { "epoch": 0.16512577940227907, "grad_norm": 6.47848653793335, "learning_rate": 9.600000000000001e-06, "loss": 45.9968, "step": 96 }, { "epoch": 0.16684583960438615, "grad_norm": 6.271050453186035, "learning_rate": 9.7e-06, "loss": 45.6172, "step": 97 }, { "epoch": 0.16856589980649322, "grad_norm": 7.013180255889893, "learning_rate": 9.800000000000001e-06, "loss": 45.1628, "step": 98 }, { "epoch": 0.1702859600086003, "grad_norm": 6.151601314544678, "learning_rate": 9.9e-06, "loss": 45.3331, "step": 99 }, { "epoch": 0.17200602021070738, "grad_norm": 6.522064685821533, "learning_rate": 1e-05, "loss": 45.01, "step": 100 }, { "epoch": 0.17372608041281445, "grad_norm": 6.936015605926514, "learning_rate": 9.999990859614876e-06, "loss": 44.7838, "step": 101 }, { "epoch": 0.17544614061492153, "grad_norm": 6.696622371673584, "learning_rate": 9.99996343849292e-06, "loss": 44.4096, "step": 102 }, { "epoch": 0.1771662008170286, "grad_norm": 6.770718574523926, "learning_rate": 9.999917736734387e-06, "loss": 44.6306, "step": 103 }, { "epoch": 0.17888626101913566, "grad_norm": 8.121861457824707, "learning_rate": 9.999853754506375e-06, "loss": 45.3421, "step": 104 }, { "epoch": 0.18060632122124273, "grad_norm": 7.101470947265625, "learning_rate": 9.999771492042807e-06, "loss": 45.415, "step": 105 }, { "epoch": 0.1823263814233498, "grad_norm": 8.22966194152832, "learning_rate": 9.99967094964445e-06, "loss": 45.3149, "step": 106 }, { "epoch": 0.18404644162545689, "grad_norm": 10.975841522216797, "learning_rate": 9.9995521276789e-06, "loss": 45.2939, "step": 107 }, { "epoch": 0.18576650182756396, "grad_norm": 7.771969795227051, "learning_rate": 9.999415026580592e-06, "loss": 45.2433, "step": 108 }, { "epoch": 0.18748656202967104, "grad_norm": 10.387533187866211, "learning_rate": 9.999259646850787e-06, "loss": 45.2594, "step": 109 }, { "epoch": 0.18920662223177812, "grad_norm": 8.546263694763184, "learning_rate": 9.999085989057578e-06, "loss": 45.0587, "step": 110 }, { "epoch": 0.1909266824338852, "grad_norm": 9.245797157287598, "learning_rate": 9.998894053835883e-06, "loss": 45.2273, "step": 111 }, { "epoch": 0.19264674263599227, "grad_norm": 9.295574188232422, "learning_rate": 9.998683841887449e-06, "loss": 45.5471, "step": 112 }, { "epoch": 0.19436680283809935, "grad_norm": 8.650074005126953, "learning_rate": 9.99845535398084e-06, "loss": 45.2802, "step": 113 }, { "epoch": 0.1960868630402064, "grad_norm": 11.102032661437988, "learning_rate": 9.998208590951449e-06, "loss": 45.3066, "step": 114 }, { "epoch": 0.19780692324231347, "grad_norm": 6.938677787780762, "learning_rate": 9.99794355370147e-06, "loss": 45.1454, "step": 115 }, { "epoch": 0.19952698344442055, "grad_norm": 16.639892578125, "learning_rate": 9.997660243199928e-06, "loss": 45.6125, "step": 116 }, { "epoch": 0.20124704364652762, "grad_norm": 11.0328369140625, "learning_rate": 9.99735866048265e-06, "loss": 45.4087, "step": 117 }, { "epoch": 0.2029671038486347, "grad_norm": 13.419515609741211, "learning_rate": 9.997038806652264e-06, "loss": 45.8139, "step": 118 }, { "epoch": 0.20468716405074178, "grad_norm": 13.69460678100586, "learning_rate": 9.996700682878206e-06, "loss": 45.304, "step": 119 }, { "epoch": 0.20640722425284885, "grad_norm": 7.643240928649902, "learning_rate": 9.996344290396713e-06, "loss": 45.1699, "step": 120 }, { "epoch": 0.20812728445495593, "grad_norm": 9.692140579223633, "learning_rate": 9.995969630510805e-06, "loss": 44.7309, "step": 121 }, { "epoch": 0.209847344657063, "grad_norm": 7.550359725952148, "learning_rate": 9.995576704590299e-06, "loss": 44.6299, "step": 122 }, { "epoch": 0.21156740485917008, "grad_norm": 6.425361156463623, "learning_rate": 9.995165514071793e-06, "loss": 44.8296, "step": 123 }, { "epoch": 0.21328746506127713, "grad_norm": 8.089837074279785, "learning_rate": 9.994736060458665e-06, "loss": 45.2622, "step": 124 }, { "epoch": 0.2150075252633842, "grad_norm": 6.491065979003906, "learning_rate": 9.994288345321059e-06, "loss": 44.3288, "step": 125 }, { "epoch": 0.21672758546549128, "grad_norm": 5.423704147338867, "learning_rate": 9.993822370295892e-06, "loss": 44.7233, "step": 126 }, { "epoch": 0.21844764566759836, "grad_norm": 7.901766777038574, "learning_rate": 9.993338137086841e-06, "loss": 45.0159, "step": 127 }, { "epoch": 0.22016770586970544, "grad_norm": 7.50944709777832, "learning_rate": 9.992835647464339e-06, "loss": 44.8785, "step": 128 }, { "epoch": 0.22188776607181251, "grad_norm": 7.400681972503662, "learning_rate": 9.992314903265561e-06, "loss": 45.3053, "step": 129 }, { "epoch": 0.2236078262739196, "grad_norm": 8.214972496032715, "learning_rate": 9.991775906394434e-06, "loss": 44.8066, "step": 130 }, { "epoch": 0.22532788647602667, "grad_norm": 7.00181770324707, "learning_rate": 9.991218658821609e-06, "loss": 44.5013, "step": 131 }, { "epoch": 0.22704794667813374, "grad_norm": 6.811863422393799, "learning_rate": 9.990643162584467e-06, "loss": 44.7327, "step": 132 }, { "epoch": 0.22876800688024082, "grad_norm": 8.760518074035645, "learning_rate": 9.99004941978712e-06, "loss": 44.6706, "step": 133 }, { "epoch": 0.23048806708234787, "grad_norm": 7.204503536224365, "learning_rate": 9.989437432600373e-06, "loss": 44.8113, "step": 134 }, { "epoch": 0.23220812728445495, "grad_norm": 8.635207176208496, "learning_rate": 9.988807203261752e-06, "loss": 45.3002, "step": 135 }, { "epoch": 0.23392818748656202, "grad_norm": 7.5076823234558105, "learning_rate": 9.988158734075468e-06, "loss": 44.919, "step": 136 }, { "epoch": 0.2356482476886691, "grad_norm": 8.329002380371094, "learning_rate": 9.98749202741243e-06, "loss": 44.9676, "step": 137 }, { "epoch": 0.23736830789077618, "grad_norm": 7.5155158042907715, "learning_rate": 9.986807085710213e-06, "loss": 45.167, "step": 138 }, { "epoch": 0.23908836809288325, "grad_norm": 8.341093063354492, "learning_rate": 9.986103911473075e-06, "loss": 45.1884, "step": 139 }, { "epoch": 0.24080842829499033, "grad_norm": 7.4146342277526855, "learning_rate": 9.985382507271928e-06, "loss": 44.8973, "step": 140 }, { "epoch": 0.2425284884970974, "grad_norm": 8.28603458404541, "learning_rate": 9.984642875744338e-06, "loss": 44.6793, "step": 141 }, { "epoch": 0.24424854869920448, "grad_norm": 7.7817583084106445, "learning_rate": 9.983885019594506e-06, "loss": 44.7361, "step": 142 }, { "epoch": 0.24596860890131156, "grad_norm": 8.71711254119873, "learning_rate": 9.983108941593277e-06, "loss": 44.514, "step": 143 }, { "epoch": 0.2476886691034186, "grad_norm": 8.746935844421387, "learning_rate": 9.982314644578111e-06, "loss": 45.4915, "step": 144 }, { "epoch": 0.24940872930552568, "grad_norm": 8.616266250610352, "learning_rate": 9.981502131453077e-06, "loss": 44.4955, "step": 145 }, { "epoch": 0.2511287895076328, "grad_norm": 8.73255443572998, "learning_rate": 9.980671405188852e-06, "loss": 44.7087, "step": 146 }, { "epoch": 0.25284884970973986, "grad_norm": 7.063310623168945, "learning_rate": 9.979822468822696e-06, "loss": 44.7221, "step": 147 }, { "epoch": 0.25456890991184694, "grad_norm": 8.635971069335938, "learning_rate": 9.978955325458453e-06, "loss": 44.5993, "step": 148 }, { "epoch": 0.25628897011395396, "grad_norm": 7.771418571472168, "learning_rate": 9.978069978266534e-06, "loss": 45.1399, "step": 149 }, { "epoch": 0.25800903031606104, "grad_norm": 9.317761421203613, "learning_rate": 9.9771664304839e-06, "loss": 44.8672, "step": 150 }, { "epoch": 0.2597290905181681, "grad_norm": 8.102662086486816, "learning_rate": 9.976244685414065e-06, "loss": 45.4121, "step": 151 }, { "epoch": 0.2614491507202752, "grad_norm": 12.453920364379883, "learning_rate": 9.97530474642707e-06, "loss": 44.6091, "step": 152 }, { "epoch": 0.26316921092238227, "grad_norm": 9.156878471374512, "learning_rate": 9.974346616959476e-06, "loss": 45.1213, "step": 153 }, { "epoch": 0.26488927112448934, "grad_norm": 11.89529037475586, "learning_rate": 9.973370300514353e-06, "loss": 44.5162, "step": 154 }, { "epoch": 0.2666093313265964, "grad_norm": 9.946409225463867, "learning_rate": 9.972375800661264e-06, "loss": 44.106, "step": 155 }, { "epoch": 0.2683293915287035, "grad_norm": 10.854037284851074, "learning_rate": 9.971363121036252e-06, "loss": 44.9188, "step": 156 }, { "epoch": 0.2700494517308106, "grad_norm": 9.196579933166504, "learning_rate": 9.970332265341833e-06, "loss": 44.6235, "step": 157 }, { "epoch": 0.27176951193291765, "grad_norm": 11.79194164276123, "learning_rate": 9.969283237346973e-06, "loss": 45.7572, "step": 158 }, { "epoch": 0.2734895721350247, "grad_norm": 7.789967060089111, "learning_rate": 9.968216040887078e-06, "loss": 44.8069, "step": 159 }, { "epoch": 0.2752096323371318, "grad_norm": 10.540818214416504, "learning_rate": 9.967130679863984e-06, "loss": 44.7341, "step": 160 }, { "epoch": 0.2769296925392389, "grad_norm": 7.4835405349731445, "learning_rate": 9.966027158245939e-06, "loss": 45.4445, "step": 161 }, { "epoch": 0.27864975274134596, "grad_norm": 9.472413063049316, "learning_rate": 9.964905480067585e-06, "loss": 44.5498, "step": 162 }, { "epoch": 0.28036981294345303, "grad_norm": 9.155404090881348, "learning_rate": 9.963765649429954e-06, "loss": 44.3932, "step": 163 }, { "epoch": 0.2820898731455601, "grad_norm": 8.14942455291748, "learning_rate": 9.962607670500442e-06, "loss": 44.5368, "step": 164 }, { "epoch": 0.2838099333476672, "grad_norm": 10.103939056396484, "learning_rate": 9.961431547512794e-06, "loss": 44.1877, "step": 165 }, { "epoch": 0.28552999354977426, "grad_norm": 7.972369194030762, "learning_rate": 9.960237284767103e-06, "loss": 44.775, "step": 166 }, { "epoch": 0.28725005375188134, "grad_norm": 13.627598762512207, "learning_rate": 9.959024886629772e-06, "loss": 44.3529, "step": 167 }, { "epoch": 0.2889701139539884, "grad_norm": 9.467161178588867, "learning_rate": 9.957794357533518e-06, "loss": 44.7574, "step": 168 }, { "epoch": 0.29069017415609544, "grad_norm": 12.242958068847656, "learning_rate": 9.956545701977347e-06, "loss": 44.5439, "step": 169 }, { "epoch": 0.2924102343582025, "grad_norm": 10.142478942871094, "learning_rate": 9.955278924526532e-06, "loss": 44.4951, "step": 170 }, { "epoch": 0.2941302945603096, "grad_norm": 12.610379219055176, "learning_rate": 9.95399402981261e-06, "loss": 44.7777, "step": 171 }, { "epoch": 0.29585035476241667, "grad_norm": 9.542120933532715, "learning_rate": 9.952691022533352e-06, "loss": 44.5877, "step": 172 }, { "epoch": 0.29757041496452374, "grad_norm": 13.538004875183105, "learning_rate": 9.951369907452752e-06, "loss": 44.4161, "step": 173 }, { "epoch": 0.2992904751666308, "grad_norm": 11.80547046661377, "learning_rate": 9.950030689401014e-06, "loss": 44.8214, "step": 174 }, { "epoch": 0.3010105353687379, "grad_norm": 11.983033180236816, "learning_rate": 9.948673373274523e-06, "loss": 44.7073, "step": 175 }, { "epoch": 0.302730595570845, "grad_norm": 10.844096183776855, "learning_rate": 9.947297964035837e-06, "loss": 44.4097, "step": 176 }, { "epoch": 0.30445065577295205, "grad_norm": 11.820161819458008, "learning_rate": 9.94590446671366e-06, "loss": 44.7555, "step": 177 }, { "epoch": 0.3061707159750591, "grad_norm": 11.587363243103027, "learning_rate": 9.94449288640284e-06, "loss": 44.2532, "step": 178 }, { "epoch": 0.3078907761771662, "grad_norm": 10.775352478027344, "learning_rate": 9.943063228264327e-06, "loss": 44.8222, "step": 179 }, { "epoch": 0.3096108363792733, "grad_norm": 12.077004432678223, "learning_rate": 9.941615497525172e-06, "loss": 45.0694, "step": 180 }, { "epoch": 0.31133089658138036, "grad_norm": 9.174921035766602, "learning_rate": 9.940149699478502e-06, "loss": 44.4424, "step": 181 }, { "epoch": 0.31305095678348743, "grad_norm": 9.949912071228027, "learning_rate": 9.938665839483503e-06, "loss": 44.82, "step": 182 }, { "epoch": 0.3147710169855945, "grad_norm": 10.341442108154297, "learning_rate": 9.937163922965394e-06, "loss": 43.993, "step": 183 }, { "epoch": 0.3164910771877016, "grad_norm": 7.373370170593262, "learning_rate": 9.93564395541541e-06, "loss": 44.6727, "step": 184 }, { "epoch": 0.31821113738980866, "grad_norm": 11.212616920471191, "learning_rate": 9.93410594239079e-06, "loss": 44.3283, "step": 185 }, { "epoch": 0.31993119759191574, "grad_norm": 8.05331802368164, "learning_rate": 9.932549889514747e-06, "loss": 44.6646, "step": 186 }, { "epoch": 0.3216512577940228, "grad_norm": 11.395249366760254, "learning_rate": 9.930975802476448e-06, "loss": 43.9711, "step": 187 }, { "epoch": 0.3233713179961299, "grad_norm": 7.730330944061279, "learning_rate": 9.929383687030999e-06, "loss": 45.2283, "step": 188 }, { "epoch": 0.3250913781982369, "grad_norm": 11.927477836608887, "learning_rate": 9.927773548999419e-06, "loss": 44.723, "step": 189 }, { "epoch": 0.326811438400344, "grad_norm": 8.933055877685547, "learning_rate": 9.92614539426862e-06, "loss": 44.5058, "step": 190 }, { "epoch": 0.32853149860245107, "grad_norm": 13.985485076904297, "learning_rate": 9.924499228791387e-06, "loss": 44.8917, "step": 191 }, { "epoch": 0.33025155880455814, "grad_norm": 13.561887741088867, "learning_rate": 9.922835058586353e-06, "loss": 44.6659, "step": 192 }, { "epoch": 0.3319716190066652, "grad_norm": 8.802722930908203, "learning_rate": 9.921152889737985e-06, "loss": 43.9996, "step": 193 }, { "epoch": 0.3336916792087723, "grad_norm": 10.387024879455566, "learning_rate": 9.919452728396548e-06, "loss": 44.7691, "step": 194 }, { "epoch": 0.3354117394108794, "grad_norm": 8.743358612060547, "learning_rate": 9.917734580778094e-06, "loss": 45.2034, "step": 195 }, { "epoch": 0.33713179961298645, "grad_norm": 7.972978591918945, "learning_rate": 9.915998453164435e-06, "loss": 45.5303, "step": 196 }, { "epoch": 0.3388518598150935, "grad_norm": 8.630827903747559, "learning_rate": 9.914244351903122e-06, "loss": 44.7464, "step": 197 }, { "epoch": 0.3405719200172006, "grad_norm": 6.4518609046936035, "learning_rate": 9.912472283407421e-06, "loss": 44.2976, "step": 198 }, { "epoch": 0.3422919802193077, "grad_norm": 8.886955261230469, "learning_rate": 9.910682254156284e-06, "loss": 44.8556, "step": 199 }, { "epoch": 0.34401204042141476, "grad_norm": 7.261415004730225, "learning_rate": 9.908874270694337e-06, "loss": 44.466, "step": 200 }, { "epoch": 0.34573210062352183, "grad_norm": 8.031500816345215, "learning_rate": 9.907048339631843e-06, "loss": 44.4753, "step": 201 }, { "epoch": 0.3474521608256289, "grad_norm": 7.2853569984436035, "learning_rate": 9.905204467644688e-06, "loss": 45.0669, "step": 202 }, { "epoch": 0.349172221027736, "grad_norm": 8.290436744689941, "learning_rate": 9.903342661474355e-06, "loss": 44.7901, "step": 203 }, { "epoch": 0.35089228122984306, "grad_norm": 9.512991905212402, "learning_rate": 9.901462927927891e-06, "loss": 44.7863, "step": 204 }, { "epoch": 0.35261234143195014, "grad_norm": 8.04033374786377, "learning_rate": 9.899565273877892e-06, "loss": 44.9776, "step": 205 }, { "epoch": 0.3543324016340572, "grad_norm": 11.140141487121582, "learning_rate": 9.897649706262474e-06, "loss": 45.1048, "step": 206 }, { "epoch": 0.3560524618361643, "grad_norm": 8.036794662475586, "learning_rate": 9.895716232085247e-06, "loss": 44.763, "step": 207 }, { "epoch": 0.3577725220382713, "grad_norm": 11.508170127868652, "learning_rate": 9.89376485841529e-06, "loss": 44.7094, "step": 208 }, { "epoch": 0.3594925822403784, "grad_norm": 8.577386856079102, "learning_rate": 9.891795592387127e-06, "loss": 44.5892, "step": 209 }, { "epoch": 0.36121264244248547, "grad_norm": 11.701868057250977, "learning_rate": 9.889808441200697e-06, "loss": 44.4664, "step": 210 }, { "epoch": 0.36293270264459254, "grad_norm": 7.955048561096191, "learning_rate": 9.887803412121331e-06, "loss": 44.5424, "step": 211 }, { "epoch": 0.3646527628466996, "grad_norm": 11.340240478515625, "learning_rate": 9.885780512479725e-06, "loss": 44.3322, "step": 212 }, { "epoch": 0.3663728230488067, "grad_norm": 8.020219802856445, "learning_rate": 9.88373974967191e-06, "loss": 44.2046, "step": 213 }, { "epoch": 0.36809288325091377, "grad_norm": 10.230839729309082, "learning_rate": 9.881681131159232e-06, "loss": 44.2038, "step": 214 }, { "epoch": 0.36981294345302085, "grad_norm": 8.711820602416992, "learning_rate": 9.879604664468315e-06, "loss": 44.0065, "step": 215 }, { "epoch": 0.3715330036551279, "grad_norm": 10.792113304138184, "learning_rate": 9.877510357191042e-06, "loss": 44.3272, "step": 216 }, { "epoch": 0.373253063857235, "grad_norm": 8.520330429077148, "learning_rate": 9.875398216984521e-06, "loss": 44.403, "step": 217 }, { "epoch": 0.3749731240593421, "grad_norm": 12.095443725585938, "learning_rate": 9.873268251571065e-06, "loss": 44.8801, "step": 218 }, { "epoch": 0.37669318426144915, "grad_norm": 10.553284645080566, "learning_rate": 9.871120468738156e-06, "loss": 44.1855, "step": 219 }, { "epoch": 0.37841324446355623, "grad_norm": 10.677011489868164, "learning_rate": 9.868954876338414e-06, "loss": 44.9765, "step": 220 }, { "epoch": 0.3801333046656633, "grad_norm": 10.681044578552246, "learning_rate": 9.866771482289585e-06, "loss": 44.3767, "step": 221 }, { "epoch": 0.3818533648677704, "grad_norm": 8.905034065246582, "learning_rate": 9.86457029457449e-06, "loss": 44.5735, "step": 222 }, { "epoch": 0.38357342506987746, "grad_norm": 11.448368072509766, "learning_rate": 9.86235132124101e-06, "loss": 44.3242, "step": 223 }, { "epoch": 0.38529348527198454, "grad_norm": 8.515959739685059, "learning_rate": 9.860114570402055e-06, "loss": 45.0004, "step": 224 }, { "epoch": 0.3870135454740916, "grad_norm": 11.419853210449219, "learning_rate": 9.85786005023553e-06, "loss": 44.7329, "step": 225 }, { "epoch": 0.3887336056761987, "grad_norm": 10.274301528930664, "learning_rate": 9.855587768984308e-06, "loss": 44.5693, "step": 226 }, { "epoch": 0.39045366587830577, "grad_norm": 9.408884048461914, "learning_rate": 9.8532977349562e-06, "loss": 44.2397, "step": 227 }, { "epoch": 0.3921737260804128, "grad_norm": 8.234622955322266, "learning_rate": 9.850989956523922e-06, "loss": 44.3023, "step": 228 }, { "epoch": 0.39389378628251986, "grad_norm": 10.358318328857422, "learning_rate": 9.848664442125068e-06, "loss": 44.6614, "step": 229 }, { "epoch": 0.39561384648462694, "grad_norm": 9.553954124450684, "learning_rate": 9.846321200262079e-06, "loss": 44.8496, "step": 230 }, { "epoch": 0.397333906686734, "grad_norm": 9.689641952514648, "learning_rate": 9.843960239502205e-06, "loss": 44.1763, "step": 231 }, { "epoch": 0.3990539668888411, "grad_norm": 8.881220817565918, "learning_rate": 9.841581568477483e-06, "loss": 44.5679, "step": 232 }, { "epoch": 0.40077402709094817, "grad_norm": 8.874262809753418, "learning_rate": 9.839185195884702e-06, "loss": 45.0497, "step": 233 }, { "epoch": 0.40249408729305525, "grad_norm": 8.762188911437988, "learning_rate": 9.836771130485367e-06, "loss": 44.415, "step": 234 }, { "epoch": 0.4042141474951623, "grad_norm": 7.233129501342773, "learning_rate": 9.834339381105676e-06, "loss": 44.4637, "step": 235 }, { "epoch": 0.4059342076972694, "grad_norm": 10.540146827697754, "learning_rate": 9.831889956636478e-06, "loss": 44.525, "step": 236 }, { "epoch": 0.4076542678993765, "grad_norm": 5.345295429229736, "learning_rate": 9.829422866033246e-06, "loss": 43.6553, "step": 237 }, { "epoch": 0.40937432810148355, "grad_norm": 9.131731033325195, "learning_rate": 9.826938118316044e-06, "loss": 44.2395, "step": 238 }, { "epoch": 0.41109438830359063, "grad_norm": 6.4219560623168945, "learning_rate": 9.82443572256949e-06, "loss": 44.541, "step": 239 }, { "epoch": 0.4128144485056977, "grad_norm": 7.843993663787842, "learning_rate": 9.821915687942729e-06, "loss": 44.6975, "step": 240 }, { "epoch": 0.4145345087078048, "grad_norm": 7.926616668701172, "learning_rate": 9.8193780236494e-06, "loss": 43.983, "step": 241 }, { "epoch": 0.41625456890991186, "grad_norm": 7.5367045402526855, "learning_rate": 9.81682273896759e-06, "loss": 43.9646, "step": 242 }, { "epoch": 0.41797462911201894, "grad_norm": 10.298775672912598, "learning_rate": 9.814249843239816e-06, "loss": 44.0679, "step": 243 }, { "epoch": 0.419694689314126, "grad_norm": 8.142918586730957, "learning_rate": 9.811659345872979e-06, "loss": 44.9597, "step": 244 }, { "epoch": 0.4214147495162331, "grad_norm": 11.297587394714355, "learning_rate": 9.809051256338338e-06, "loss": 44.3569, "step": 245 }, { "epoch": 0.42313480971834017, "grad_norm": 8.743409156799316, "learning_rate": 9.806425584171468e-06, "loss": 43.6217, "step": 246 }, { "epoch": 0.42485486992044724, "grad_norm": 9.554738998413086, "learning_rate": 9.803782338972235e-06, "loss": 44.3762, "step": 247 }, { "epoch": 0.42657493012255426, "grad_norm": 8.766114234924316, "learning_rate": 9.801121530404746e-06, "loss": 44.1824, "step": 248 }, { "epoch": 0.42829499032466134, "grad_norm": 8.465466499328613, "learning_rate": 9.798443168197332e-06, "loss": 44.0283, "step": 249 }, { "epoch": 0.4300150505267684, "grad_norm": 8.999267578125, "learning_rate": 9.795747262142494e-06, "loss": 44.1171, "step": 250 }, { "epoch": 0.4317351107288755, "grad_norm": 6.678277492523193, "learning_rate": 9.79303382209688e-06, "loss": 44.7172, "step": 251 }, { "epoch": 0.43345517093098257, "grad_norm": 9.430837631225586, "learning_rate": 9.790302857981247e-06, "loss": 44.3632, "step": 252 }, { "epoch": 0.43517523113308965, "grad_norm": 6.532567501068115, "learning_rate": 9.787554379780417e-06, "loss": 44.2348, "step": 253 }, { "epoch": 0.4368952913351967, "grad_norm": 9.008966445922852, "learning_rate": 9.784788397543254e-06, "loss": 43.9189, "step": 254 }, { "epoch": 0.4386153515373038, "grad_norm": 7.171030521392822, "learning_rate": 9.782004921382612e-06, "loss": 44.719, "step": 255 }, { "epoch": 0.4403354117394109, "grad_norm": 8.457947731018066, "learning_rate": 9.77920396147531e-06, "loss": 44.3203, "step": 256 }, { "epoch": 0.44205547194151795, "grad_norm": 8.303704261779785, "learning_rate": 9.77638552806209e-06, "loss": 44.6251, "step": 257 }, { "epoch": 0.44377553214362503, "grad_norm": 7.793144702911377, "learning_rate": 9.773549631447576e-06, "loss": 44.4527, "step": 258 }, { "epoch": 0.4454955923457321, "grad_norm": 9.074666976928711, "learning_rate": 9.770696282000245e-06, "loss": 44.4602, "step": 259 }, { "epoch": 0.4472156525478392, "grad_norm": 7.790366172790527, "learning_rate": 9.767825490152381e-06, "loss": 44.0525, "step": 260 }, { "epoch": 0.44893571274994626, "grad_norm": 6.790820598602295, "learning_rate": 9.764937266400042e-06, "loss": 44.3677, "step": 261 }, { "epoch": 0.45065577295205334, "grad_norm": 7.48856782913208, "learning_rate": 9.76203162130302e-06, "loss": 44.281, "step": 262 }, { "epoch": 0.4523758331541604, "grad_norm": 6.4276814460754395, "learning_rate": 9.759108565484796e-06, "loss": 44.5151, "step": 263 }, { "epoch": 0.4540958933562675, "grad_norm": 7.4707350730896, "learning_rate": 9.756168109632519e-06, "loss": 44.4653, "step": 264 }, { "epoch": 0.45581595355837456, "grad_norm": 6.78653621673584, "learning_rate": 9.753210264496943e-06, "loss": 45.1088, "step": 265 }, { "epoch": 0.45753601376048164, "grad_norm": 6.732356548309326, "learning_rate": 9.75023504089241e-06, "loss": 44.3737, "step": 266 }, { "epoch": 0.4592560739625887, "grad_norm": 6.069101810455322, "learning_rate": 9.747242449696794e-06, "loss": 44.5619, "step": 267 }, { "epoch": 0.46097613416469574, "grad_norm": 6.123370170593262, "learning_rate": 9.74423250185147e-06, "loss": 44.3021, "step": 268 }, { "epoch": 0.4626961943668028, "grad_norm": 6.104144096374512, "learning_rate": 9.74120520836127e-06, "loss": 44.2148, "step": 269 }, { "epoch": 0.4644162545689099, "grad_norm": 8.752744674682617, "learning_rate": 9.738160580294444e-06, "loss": 44.6205, "step": 270 }, { "epoch": 0.46613631477101697, "grad_norm": 7.354405879974365, "learning_rate": 9.735098628782624e-06, "loss": 44.459, "step": 271 }, { "epoch": 0.46785637497312405, "grad_norm": 8.801732063293457, "learning_rate": 9.732019365020778e-06, "loss": 44.635, "step": 272 }, { "epoch": 0.4695764351752311, "grad_norm": 7.110331058502197, "learning_rate": 9.728922800267162e-06, "loss": 44.3402, "step": 273 }, { "epoch": 0.4712964953773382, "grad_norm": 8.147953033447266, "learning_rate": 9.7258089458433e-06, "loss": 44.6291, "step": 274 }, { "epoch": 0.4730165555794453, "grad_norm": 7.08845853805542, "learning_rate": 9.722677813133921e-06, "loss": 45.0577, "step": 275 }, { "epoch": 0.47473661578155235, "grad_norm": 6.4095025062561035, "learning_rate": 9.719529413586928e-06, "loss": 43.9258, "step": 276 }, { "epoch": 0.47645667598365943, "grad_norm": 7.270499229431152, "learning_rate": 9.716363758713357e-06, "loss": 44.8198, "step": 277 }, { "epoch": 0.4781767361857665, "grad_norm": 7.040411949157715, "learning_rate": 9.713180860087328e-06, "loss": 44.1966, "step": 278 }, { "epoch": 0.4798967963878736, "grad_norm": 6.831189155578613, "learning_rate": 9.709980729346009e-06, "loss": 44.88, "step": 279 }, { "epoch": 0.48161685658998066, "grad_norm": 7.017982482910156, "learning_rate": 9.706763378189571e-06, "loss": 44.3914, "step": 280 }, { "epoch": 0.48333691679208773, "grad_norm": 7.631150245666504, "learning_rate": 9.703528818381144e-06, "loss": 44.3798, "step": 281 }, { "epoch": 0.4850569769941948, "grad_norm": 6.717972278594971, "learning_rate": 9.70027706174678e-06, "loss": 43.8038, "step": 282 }, { "epoch": 0.4867770371963019, "grad_norm": 8.252500534057617, "learning_rate": 9.697008120175402e-06, "loss": 44.4889, "step": 283 }, { "epoch": 0.48849709739840896, "grad_norm": 7.6612420082092285, "learning_rate": 9.693722005618763e-06, "loss": 44.2001, "step": 284 }, { "epoch": 0.49021715760051604, "grad_norm": 6.846263408660889, "learning_rate": 9.690418730091403e-06, "loss": 44.6985, "step": 285 }, { "epoch": 0.4919372178026231, "grad_norm": 6.461937427520752, "learning_rate": 9.687098305670606e-06, "loss": 44.2506, "step": 286 }, { "epoch": 0.49365727800473014, "grad_norm": 6.650141716003418, "learning_rate": 9.683760744496356e-06, "loss": 44.3858, "step": 287 }, { "epoch": 0.4953773382068372, "grad_norm": 6.5828986167907715, "learning_rate": 9.68040605877129e-06, "loss": 43.6322, "step": 288 }, { "epoch": 0.4970973984089443, "grad_norm": 6.385183334350586, "learning_rate": 9.677034260760658e-06, "loss": 44.7745, "step": 289 }, { "epoch": 0.49881745861105137, "grad_norm": 7.130415916442871, "learning_rate": 9.673645362792273e-06, "loss": 44.1543, "step": 290 }, { "epoch": 0.5005375188131584, "grad_norm": 6.580416202545166, "learning_rate": 9.670239377256467e-06, "loss": 43.8422, "step": 291 }, { "epoch": 0.5022575790152656, "grad_norm": 7.959731101989746, "learning_rate": 9.666816316606044e-06, "loss": 44.4367, "step": 292 }, { "epoch": 0.5039776392173726, "grad_norm": 6.089702606201172, "learning_rate": 9.663376193356249e-06, "loss": 43.9484, "step": 293 }, { "epoch": 0.5056976994194797, "grad_norm": 8.458806037902832, "learning_rate": 9.659919020084695e-06, "loss": 44.1408, "step": 294 }, { "epoch": 0.5074177596215868, "grad_norm": 6.733780860900879, "learning_rate": 9.656444809431344e-06, "loss": 43.9267, "step": 295 }, { "epoch": 0.5091378198236939, "grad_norm": 8.281241416931152, "learning_rate": 9.652953574098444e-06, "loss": 44.7447, "step": 296 }, { "epoch": 0.5108578800258009, "grad_norm": 7.2893195152282715, "learning_rate": 9.649445326850491e-06, "loss": 44.1749, "step": 297 }, { "epoch": 0.5125779402279079, "grad_norm": 8.188138961791992, "learning_rate": 9.645920080514176e-06, "loss": 44.5725, "step": 298 }, { "epoch": 0.5142980004300151, "grad_norm": 8.285508155822754, "learning_rate": 9.642377847978343e-06, "loss": 44.4519, "step": 299 }, { "epoch": 0.5160180606321221, "grad_norm": 12.107803344726562, "learning_rate": 9.638818642193939e-06, "loss": 43.6642, "step": 300 }, { "epoch": 0.5177381208342292, "grad_norm": 10.988150596618652, "learning_rate": 9.63524247617397e-06, "loss": 43.9385, "step": 301 }, { "epoch": 0.5194581810363362, "grad_norm": 12.65985107421875, "learning_rate": 9.631649362993447e-06, "loss": 44.304, "step": 302 }, { "epoch": 0.5211782412384434, "grad_norm": 12.63979721069336, "learning_rate": 9.62803931578935e-06, "loss": 44.2028, "step": 303 }, { "epoch": 0.5228983014405504, "grad_norm": 7.90657377243042, "learning_rate": 9.624412347760564e-06, "loss": 44.1649, "step": 304 }, { "epoch": 0.5246183616426575, "grad_norm": 9.31624698638916, "learning_rate": 9.620768472167844e-06, "loss": 43.996, "step": 305 }, { "epoch": 0.5263384218447645, "grad_norm": 8.557055473327637, "learning_rate": 9.61710770233376e-06, "loss": 44.3358, "step": 306 }, { "epoch": 0.5280584820468717, "grad_norm": 7.057743549346924, "learning_rate": 9.613430051642652e-06, "loss": 44.583, "step": 307 }, { "epoch": 0.5297785422489787, "grad_norm": 7.244456768035889, "learning_rate": 9.609735533540576e-06, "loss": 43.7423, "step": 308 }, { "epoch": 0.5314986024510858, "grad_norm": 6.6239333152771, "learning_rate": 9.606024161535261e-06, "loss": 43.719, "step": 309 }, { "epoch": 0.5332186626531928, "grad_norm": 7.109512805938721, "learning_rate": 9.602295949196052e-06, "loss": 43.8263, "step": 310 }, { "epoch": 0.5349387228553, "grad_norm": 7.938423156738281, "learning_rate": 9.59855091015387e-06, "loss": 43.513, "step": 311 }, { "epoch": 0.536658783057407, "grad_norm": 7.159519195556641, "learning_rate": 9.594789058101154e-06, "loss": 43.7476, "step": 312 }, { "epoch": 0.5383788432595141, "grad_norm": 9.051861763000488, "learning_rate": 9.591010406791814e-06, "loss": 44.6808, "step": 313 }, { "epoch": 0.5400989034616211, "grad_norm": 7.437441825866699, "learning_rate": 9.587214970041181e-06, "loss": 44.2018, "step": 314 }, { "epoch": 0.5418189636637283, "grad_norm": 8.533609390258789, "learning_rate": 9.58340276172596e-06, "loss": 44.35, "step": 315 }, { "epoch": 0.5435390238658353, "grad_norm": 7.41975736618042, "learning_rate": 9.579573795784167e-06, "loss": 44.1627, "step": 316 }, { "epoch": 0.5452590840679423, "grad_norm": 6.5013580322265625, "learning_rate": 9.575728086215093e-06, "loss": 44.0411, "step": 317 }, { "epoch": 0.5469791442700495, "grad_norm": 8.311059951782227, "learning_rate": 9.571865647079246e-06, "loss": 44.6953, "step": 318 }, { "epoch": 0.5486992044721565, "grad_norm": 5.960739612579346, "learning_rate": 9.567986492498299e-06, "loss": 44.1261, "step": 319 }, { "epoch": 0.5504192646742636, "grad_norm": 7.7150959968566895, "learning_rate": 9.564090636655033e-06, "loss": 44.0052, "step": 320 }, { "epoch": 0.5521393248763706, "grad_norm": 7.0516815185546875, "learning_rate": 9.560178093793304e-06, "loss": 44.4024, "step": 321 }, { "epoch": 0.5538593850784778, "grad_norm": 6.510403633117676, "learning_rate": 9.55624887821797e-06, "loss": 44.3171, "step": 322 }, { "epoch": 0.5555794452805848, "grad_norm": 6.586174488067627, "learning_rate": 9.552303004294845e-06, "loss": 44.1694, "step": 323 }, { "epoch": 0.5572995054826919, "grad_norm": 7.093349456787109, "learning_rate": 9.548340486450656e-06, "loss": 43.9714, "step": 324 }, { "epoch": 0.5590195656847989, "grad_norm": 5.638337135314941, "learning_rate": 9.544361339172976e-06, "loss": 43.9597, "step": 325 }, { "epoch": 0.5607396258869061, "grad_norm": 6.339056491851807, "learning_rate": 9.54036557701018e-06, "loss": 43.9718, "step": 326 }, { "epoch": 0.5624596860890131, "grad_norm": 6.259551525115967, "learning_rate": 9.536353214571393e-06, "loss": 44.1765, "step": 327 }, { "epoch": 0.5641797462911202, "grad_norm": 6.071033477783203, "learning_rate": 9.53232426652643e-06, "loss": 44.1298, "step": 328 }, { "epoch": 0.5658998064932272, "grad_norm": 5.966522216796875, "learning_rate": 9.528278747605741e-06, "loss": 43.5899, "step": 329 }, { "epoch": 0.5676198666953344, "grad_norm": 6.373861312866211, "learning_rate": 9.52421667260037e-06, "loss": 43.9747, "step": 330 }, { "epoch": 0.5693399268974414, "grad_norm": 7.79583215713501, "learning_rate": 9.52013805636189e-06, "loss": 44.1957, "step": 331 }, { "epoch": 0.5710599870995485, "grad_norm": 6.730131149291992, "learning_rate": 9.516042913802349e-06, "loss": 44.5105, "step": 332 }, { "epoch": 0.5727800473016555, "grad_norm": 6.649820804595947, "learning_rate": 9.511931259894219e-06, "loss": 43.4763, "step": 333 }, { "epoch": 0.5745001075037627, "grad_norm": 8.12640380859375, "learning_rate": 9.507803109670337e-06, "loss": 43.5421, "step": 334 }, { "epoch": 0.5762201677058697, "grad_norm": 7.1897783279418945, "learning_rate": 9.503658478223862e-06, "loss": 43.661, "step": 335 }, { "epoch": 0.5779402279079768, "grad_norm": 6.837246417999268, "learning_rate": 9.499497380708202e-06, "loss": 43.7997, "step": 336 }, { "epoch": 0.5796602881100839, "grad_norm": 8.631741523742676, "learning_rate": 9.495319832336969e-06, "loss": 43.8287, "step": 337 }, { "epoch": 0.5813803483121909, "grad_norm": 8.00960636138916, "learning_rate": 9.491125848383926e-06, "loss": 43.6861, "step": 338 }, { "epoch": 0.583100408514298, "grad_norm": 6.487185001373291, "learning_rate": 9.486915444182926e-06, "loss": 43.8275, "step": 339 }, { "epoch": 0.584820468716405, "grad_norm": 7.411306381225586, "learning_rate": 9.482688635127849e-06, "loss": 43.4639, "step": 340 }, { "epoch": 0.5865405289185122, "grad_norm": 6.86546516418457, "learning_rate": 9.478445436672566e-06, "loss": 44.094, "step": 341 }, { "epoch": 0.5882605891206192, "grad_norm": 8.094916343688965, "learning_rate": 9.474185864330861e-06, "loss": 43.8569, "step": 342 }, { "epoch": 0.5899806493227263, "grad_norm": 7.7384138107299805, "learning_rate": 9.469909933676388e-06, "loss": 43.8937, "step": 343 }, { "epoch": 0.5917007095248333, "grad_norm": 8.301685333251953, "learning_rate": 9.46561766034261e-06, "loss": 44.288, "step": 344 }, { "epoch": 0.5934207697269405, "grad_norm": 10.100611686706543, "learning_rate": 9.461309060022737e-06, "loss": 43.9889, "step": 345 }, { "epoch": 0.5951408299290475, "grad_norm": 6.778660774230957, "learning_rate": 9.456984148469674e-06, "loss": 43.697, "step": 346 }, { "epoch": 0.5968608901311546, "grad_norm": 11.071110725402832, "learning_rate": 9.452642941495967e-06, "loss": 44.3392, "step": 347 }, { "epoch": 0.5985809503332616, "grad_norm": 7.321798324584961, "learning_rate": 9.448285454973739e-06, "loss": 43.5908, "step": 348 }, { "epoch": 0.6003010105353688, "grad_norm": 10.584439277648926, "learning_rate": 9.443911704834624e-06, "loss": 44.094, "step": 349 }, { "epoch": 0.6020210707374758, "grad_norm": 7.291213035583496, "learning_rate": 9.439521707069737e-06, "loss": 43.7771, "step": 350 }, { "epoch": 0.6037411309395829, "grad_norm": 10.645991325378418, "learning_rate": 9.435115477729577e-06, "loss": 44.2799, "step": 351 }, { "epoch": 0.60546119114169, "grad_norm": 8.463363647460938, "learning_rate": 9.430693032924003e-06, "loss": 43.9206, "step": 352 }, { "epoch": 0.6071812513437971, "grad_norm": 11.510597229003906, "learning_rate": 9.426254388822152e-06, "loss": 43.8839, "step": 353 }, { "epoch": 0.6089013115459041, "grad_norm": 9.736458778381348, "learning_rate": 9.421799561652391e-06, "loss": 44.6627, "step": 354 }, { "epoch": 0.6106213717480112, "grad_norm": 11.647321701049805, "learning_rate": 9.417328567702256e-06, "loss": 43.6068, "step": 355 }, { "epoch": 0.6123414319501183, "grad_norm": 12.150289535522461, "learning_rate": 9.412841423318386e-06, "loss": 43.6395, "step": 356 }, { "epoch": 0.6140614921522253, "grad_norm": 9.2246675491333, "learning_rate": 9.408338144906475e-06, "loss": 44.1518, "step": 357 }, { "epoch": 0.6157815523543324, "grad_norm": 10.310601234436035, "learning_rate": 9.403818748931201e-06, "loss": 43.9735, "step": 358 }, { "epoch": 0.6175016125564394, "grad_norm": 8.350564002990723, "learning_rate": 9.399283251916174e-06, "loss": 43.7255, "step": 359 }, { "epoch": 0.6192216727585466, "grad_norm": 9.116833686828613, "learning_rate": 9.394731670443869e-06, "loss": 43.6732, "step": 360 }, { "epoch": 0.6209417329606536, "grad_norm": 8.664816856384277, "learning_rate": 9.390164021155568e-06, "loss": 43.6788, "step": 361 }, { "epoch": 0.6226617931627607, "grad_norm": 8.421276092529297, "learning_rate": 9.385580320751301e-06, "loss": 43.7357, "step": 362 }, { "epoch": 0.6243818533648677, "grad_norm": 7.4578680992126465, "learning_rate": 9.380980585989782e-06, "loss": 43.7452, "step": 363 }, { "epoch": 0.6261019135669749, "grad_norm": 8.058576583862305, "learning_rate": 9.376364833688352e-06, "loss": 43.839, "step": 364 }, { "epoch": 0.6278219737690819, "grad_norm": 7.826845169067383, "learning_rate": 9.371733080722911e-06, "loss": 43.8903, "step": 365 }, { "epoch": 0.629542033971189, "grad_norm": 7.76792573928833, "learning_rate": 9.367085344027862e-06, "loss": 44.3217, "step": 366 }, { "epoch": 0.631262094173296, "grad_norm": 8.252405166625977, "learning_rate": 9.362421640596044e-06, "loss": 44.0905, "step": 367 }, { "epoch": 0.6329821543754032, "grad_norm": 7.431006908416748, "learning_rate": 9.35774198747868e-06, "loss": 43.864, "step": 368 }, { "epoch": 0.6347022145775102, "grad_norm": 9.33600902557373, "learning_rate": 9.353046401785297e-06, "loss": 43.0342, "step": 369 }, { "epoch": 0.6364222747796173, "grad_norm": 7.454495906829834, "learning_rate": 9.348334900683685e-06, "loss": 43.7442, "step": 370 }, { "epoch": 0.6381423349817243, "grad_norm": 8.429414749145508, "learning_rate": 9.343607501399812e-06, "loss": 43.7836, "step": 371 }, { "epoch": 0.6398623951838315, "grad_norm": 7.481090545654297, "learning_rate": 9.338864221217783e-06, "loss": 43.9994, "step": 372 }, { "epoch": 0.6415824553859385, "grad_norm": 7.106781482696533, "learning_rate": 9.33410507747976e-06, "loss": 43.6922, "step": 373 }, { "epoch": 0.6433025155880456, "grad_norm": 7.0425615310668945, "learning_rate": 9.329330087585905e-06, "loss": 44.5278, "step": 374 }, { "epoch": 0.6450225757901527, "grad_norm": 7.197376728057861, "learning_rate": 9.324539268994317e-06, "loss": 43.8955, "step": 375 }, { "epoch": 0.6467426359922598, "grad_norm": 6.9038286209106445, "learning_rate": 9.319732639220965e-06, "loss": 43.9268, "step": 376 }, { "epoch": 0.6484626961943668, "grad_norm": 7.032724380493164, "learning_rate": 9.31491021583963e-06, "loss": 43.5221, "step": 377 }, { "epoch": 0.6501827563964738, "grad_norm": 7.234856605529785, "learning_rate": 9.310072016481832e-06, "loss": 43.4951, "step": 378 }, { "epoch": 0.651902816598581, "grad_norm": 6.546868801116943, "learning_rate": 9.305218058836778e-06, "loss": 44.0876, "step": 379 }, { "epoch": 0.653622876800688, "grad_norm": 6.904932498931885, "learning_rate": 9.300348360651282e-06, "loss": 44.5423, "step": 380 }, { "epoch": 0.6553429370027951, "grad_norm": 7.505612850189209, "learning_rate": 9.295462939729711e-06, "loss": 43.9986, "step": 381 }, { "epoch": 0.6570629972049021, "grad_norm": 5.859259605407715, "learning_rate": 9.290561813933916e-06, "loss": 43.8683, "step": 382 }, { "epoch": 0.6587830574070093, "grad_norm": 8.047765731811523, "learning_rate": 9.285645001183167e-06, "loss": 44.4658, "step": 383 }, { "epoch": 0.6605031176091163, "grad_norm": 6.570570945739746, "learning_rate": 9.280712519454092e-06, "loss": 43.6115, "step": 384 }, { "epoch": 0.6622231778112234, "grad_norm": 6.266587734222412, "learning_rate": 9.2757643867806e-06, "loss": 43.6822, "step": 385 }, { "epoch": 0.6639432380133304, "grad_norm": 7.281513214111328, "learning_rate": 9.270800621253833e-06, "loss": 43.8285, "step": 386 }, { "epoch": 0.6656632982154376, "grad_norm": 6.563234806060791, "learning_rate": 9.265821241022074e-06, "loss": 43.6976, "step": 387 }, { "epoch": 0.6673833584175446, "grad_norm": 6.870432376861572, "learning_rate": 9.26082626429071e-06, "loss": 43.9566, "step": 388 }, { "epoch": 0.6691034186196517, "grad_norm": 8.11976432800293, "learning_rate": 9.255815709322142e-06, "loss": 43.8613, "step": 389 }, { "epoch": 0.6708234788217587, "grad_norm": 7.040714740753174, "learning_rate": 9.250789594435735e-06, "loss": 43.3387, "step": 390 }, { "epoch": 0.6725435390238659, "grad_norm": 6.891185283660889, "learning_rate": 9.245747938007734e-06, "loss": 43.596, "step": 391 }, { "epoch": 0.6742635992259729, "grad_norm": 7.045391082763672, "learning_rate": 9.240690758471216e-06, "loss": 43.1001, "step": 392 }, { "epoch": 0.67598365942808, "grad_norm": 6.838486194610596, "learning_rate": 9.235618074316005e-06, "loss": 44.0918, "step": 393 }, { "epoch": 0.677703719630187, "grad_norm": 8.006799697875977, "learning_rate": 9.230529904088621e-06, "loss": 43.4563, "step": 394 }, { "epoch": 0.6794237798322942, "grad_norm": 7.786087989807129, "learning_rate": 9.225426266392191e-06, "loss": 44.1002, "step": 395 }, { "epoch": 0.6811438400344012, "grad_norm": 7.782168388366699, "learning_rate": 9.220307179886408e-06, "loss": 44.15, "step": 396 }, { "epoch": 0.6828639002365082, "grad_norm": 7.179986953735352, "learning_rate": 9.215172663287435e-06, "loss": 43.8326, "step": 397 }, { "epoch": 0.6845839604386154, "grad_norm": 8.045145988464355, "learning_rate": 9.210022735367857e-06, "loss": 43.3196, "step": 398 }, { "epoch": 0.6863040206407224, "grad_norm": 7.906603813171387, "learning_rate": 9.204857414956606e-06, "loss": 44.152, "step": 399 }, { "epoch": 0.6880240808428295, "grad_norm": 8.266923904418945, "learning_rate": 9.199676720938886e-06, "loss": 44.158, "step": 400 }, { "epoch": 0.6897441410449365, "grad_norm": 7.465760231018066, "learning_rate": 9.194480672256117e-06, "loss": 43.9078, "step": 401 }, { "epoch": 0.6914642012470437, "grad_norm": 7.522243499755859, "learning_rate": 9.189269287905849e-06, "loss": 43.3097, "step": 402 }, { "epoch": 0.6931842614491507, "grad_norm": 6.484007358551025, "learning_rate": 9.184042586941708e-06, "loss": 43.9014, "step": 403 }, { "epoch": 0.6949043216512578, "grad_norm": 6.548778533935547, "learning_rate": 9.178800588473317e-06, "loss": 43.9104, "step": 404 }, { "epoch": 0.6966243818533648, "grad_norm": 6.578863620758057, "learning_rate": 9.17354331166623e-06, "loss": 44.0022, "step": 405 }, { "epoch": 0.698344442055472, "grad_norm": 7.083658695220947, "learning_rate": 9.168270775741863e-06, "loss": 43.7902, "step": 406 }, { "epoch": 0.700064502257579, "grad_norm": 7.241711139678955, "learning_rate": 9.162982999977417e-06, "loss": 43.9464, "step": 407 }, { "epoch": 0.7017845624596861, "grad_norm": 6.329436779022217, "learning_rate": 9.157680003705816e-06, "loss": 44.1072, "step": 408 }, { "epoch": 0.7035046226617931, "grad_norm": 6.435650825500488, "learning_rate": 9.15236180631563e-06, "loss": 43.392, "step": 409 }, { "epoch": 0.7052246828639003, "grad_norm": 5.681223392486572, "learning_rate": 9.14702842725101e-06, "loss": 44.3148, "step": 410 }, { "epoch": 0.7069447430660073, "grad_norm": 6.71289587020874, "learning_rate": 9.14167988601161e-06, "loss": 43.8893, "step": 411 }, { "epoch": 0.7086648032681144, "grad_norm": 6.497440814971924, "learning_rate": 9.13631620215252e-06, "loss": 44.1776, "step": 412 }, { "epoch": 0.7103848634702215, "grad_norm": 7.291422367095947, "learning_rate": 9.130937395284199e-06, "loss": 43.8195, "step": 413 }, { "epoch": 0.7121049236723286, "grad_norm": 6.935153961181641, "learning_rate": 9.125543485072386e-06, "loss": 43.9977, "step": 414 }, { "epoch": 0.7138249838744356, "grad_norm": 6.302245140075684, "learning_rate": 9.120134491238054e-06, "loss": 43.677, "step": 415 }, { "epoch": 0.7155450440765426, "grad_norm": 6.205868244171143, "learning_rate": 9.114710433557314e-06, "loss": 43.8423, "step": 416 }, { "epoch": 0.7172651042786498, "grad_norm": 5.34831428527832, "learning_rate": 9.109271331861361e-06, "loss": 43.6707, "step": 417 }, { "epoch": 0.7189851644807568, "grad_norm": 7.174152374267578, "learning_rate": 9.103817206036383e-06, "loss": 43.3579, "step": 418 }, { "epoch": 0.7207052246828639, "grad_norm": 6.666977882385254, "learning_rate": 9.098348076023506e-06, "loss": 43.8424, "step": 419 }, { "epoch": 0.7224252848849709, "grad_norm": 7.491025924682617, "learning_rate": 9.092863961818715e-06, "loss": 44.4333, "step": 420 }, { "epoch": 0.7241453450870781, "grad_norm": 6.508261680603027, "learning_rate": 9.087364883472774e-06, "loss": 43.7001, "step": 421 }, { "epoch": 0.7258654052891851, "grad_norm": 7.418080806732178, "learning_rate": 9.08185086109116e-06, "loss": 44.0001, "step": 422 }, { "epoch": 0.7275854654912922, "grad_norm": 6.983603000640869, "learning_rate": 9.076321914833988e-06, "loss": 44.6241, "step": 423 }, { "epoch": 0.7293055256933992, "grad_norm": 8.667305946350098, "learning_rate": 9.070778064915937e-06, "loss": 44.1089, "step": 424 }, { "epoch": 0.7310255858955064, "grad_norm": 7.419984340667725, "learning_rate": 9.065219331606182e-06, "loss": 43.9046, "step": 425 }, { "epoch": 0.7327456460976134, "grad_norm": 7.34318733215332, "learning_rate": 9.0596457352283e-06, "loss": 43.794, "step": 426 }, { "epoch": 0.7344657062997205, "grad_norm": 7.931493759155273, "learning_rate": 9.054057296160221e-06, "loss": 44.6317, "step": 427 }, { "epoch": 0.7361857665018275, "grad_norm": 6.583981037139893, "learning_rate": 9.048454034834143e-06, "loss": 43.5199, "step": 428 }, { "epoch": 0.7379058267039347, "grad_norm": 8.499653816223145, "learning_rate": 9.042835971736446e-06, "loss": 43.8616, "step": 429 }, { "epoch": 0.7396258869060417, "grad_norm": 6.757936000823975, "learning_rate": 9.037203127407642e-06, "loss": 44.0385, "step": 430 }, { "epoch": 0.7413459471081488, "grad_norm": 8.30978012084961, "learning_rate": 9.031555522442268e-06, "loss": 43.4628, "step": 431 }, { "epoch": 0.7430660073102558, "grad_norm": 7.357321262359619, "learning_rate": 9.025893177488848e-06, "loss": 43.6677, "step": 432 }, { "epoch": 0.744786067512363, "grad_norm": 8.4613676071167, "learning_rate": 9.02021611324978e-06, "loss": 43.2542, "step": 433 }, { "epoch": 0.74650612771447, "grad_norm": 8.782477378845215, "learning_rate": 9.014524350481287e-06, "loss": 44.0515, "step": 434 }, { "epoch": 0.7482261879165771, "grad_norm": 6.701351165771484, "learning_rate": 9.008817909993332e-06, "loss": 44.0643, "step": 435 }, { "epoch": 0.7499462481186842, "grad_norm": 7.1624884605407715, "learning_rate": 9.00309681264954e-06, "loss": 44.0639, "step": 436 }, { "epoch": 0.7516663083207912, "grad_norm": 6.229190349578857, "learning_rate": 8.997361079367124e-06, "loss": 43.9093, "step": 437 }, { "epoch": 0.7533863685228983, "grad_norm": 7.180543422698975, "learning_rate": 8.991610731116808e-06, "loss": 44.2652, "step": 438 }, { "epoch": 0.7551064287250053, "grad_norm": 6.3346781730651855, "learning_rate": 8.985845788922753e-06, "loss": 43.2561, "step": 439 }, { "epoch": 0.7568264889271125, "grad_norm": 7.270414352416992, "learning_rate": 8.980066273862473e-06, "loss": 43.9074, "step": 440 }, { "epoch": 0.7585465491292195, "grad_norm": 6.431163311004639, "learning_rate": 8.974272207066767e-06, "loss": 43.9343, "step": 441 }, { "epoch": 0.7602666093313266, "grad_norm": 6.415679931640625, "learning_rate": 8.968463609719636e-06, "loss": 44.2067, "step": 442 }, { "epoch": 0.7619866695334336, "grad_norm": 6.083033561706543, "learning_rate": 8.962640503058206e-06, "loss": 43.7967, "step": 443 }, { "epoch": 0.7637067297355408, "grad_norm": 6.942599773406982, "learning_rate": 8.956802908372652e-06, "loss": 43.4928, "step": 444 }, { "epoch": 0.7654267899376478, "grad_norm": 6.515557765960693, "learning_rate": 8.95095084700612e-06, "loss": 43.6577, "step": 445 }, { "epoch": 0.7671468501397549, "grad_norm": 7.167238235473633, "learning_rate": 8.945084340354646e-06, "loss": 43.4542, "step": 446 }, { "epoch": 0.7688669103418619, "grad_norm": 7.30296516418457, "learning_rate": 8.939203409867084e-06, "loss": 43.123, "step": 447 }, { "epoch": 0.7705869705439691, "grad_norm": 7.387278079986572, "learning_rate": 8.933308077045022e-06, "loss": 43.6603, "step": 448 }, { "epoch": 0.7723070307460761, "grad_norm": 7.026780128479004, "learning_rate": 8.927398363442705e-06, "loss": 44.0497, "step": 449 }, { "epoch": 0.7740270909481832, "grad_norm": 7.03558874130249, "learning_rate": 8.921474290666955e-06, "loss": 43.1461, "step": 450 }, { "epoch": 0.7757471511502902, "grad_norm": 8.48353099822998, "learning_rate": 8.915535880377096e-06, "loss": 43.6771, "step": 451 }, { "epoch": 0.7774672113523974, "grad_norm": 7.250082015991211, "learning_rate": 8.909583154284868e-06, "loss": 43.9369, "step": 452 }, { "epoch": 0.7791872715545044, "grad_norm": 8.958197593688965, "learning_rate": 8.90361613415436e-06, "loss": 43.7911, "step": 453 }, { "epoch": 0.7809073317566115, "grad_norm": 8.168319702148438, "learning_rate": 8.897634841801911e-06, "loss": 43.3905, "step": 454 }, { "epoch": 0.7826273919587186, "grad_norm": 8.520408630371094, "learning_rate": 8.891639299096051e-06, "loss": 43.3708, "step": 455 }, { "epoch": 0.7843474521608256, "grad_norm": 8.194758415222168, "learning_rate": 8.885629527957407e-06, "loss": 43.3692, "step": 456 }, { "epoch": 0.7860675123629327, "grad_norm": 7.554206848144531, "learning_rate": 8.879605550358627e-06, "loss": 43.5693, "step": 457 }, { "epoch": 0.7877875725650397, "grad_norm": 7.146202087402344, "learning_rate": 8.873567388324302e-06, "loss": 43.8261, "step": 458 }, { "epoch": 0.7895076327671469, "grad_norm": 6.808493137359619, "learning_rate": 8.867515063930881e-06, "loss": 43.3648, "step": 459 }, { "epoch": 0.7912276929692539, "grad_norm": 7.165658950805664, "learning_rate": 8.861448599306597e-06, "loss": 43.0367, "step": 460 }, { "epoch": 0.792947753171361, "grad_norm": 6.527984142303467, "learning_rate": 8.855368016631377e-06, "loss": 43.4491, "step": 461 }, { "epoch": 0.794667813373468, "grad_norm": 6.912752628326416, "learning_rate": 8.849273338136772e-06, "loss": 43.6405, "step": 462 }, { "epoch": 0.7963878735755752, "grad_norm": 6.334918975830078, "learning_rate": 8.84316458610586e-06, "loss": 44.4637, "step": 463 }, { "epoch": 0.7981079337776822, "grad_norm": 7.22133207321167, "learning_rate": 8.837041782873182e-06, "loss": 43.2829, "step": 464 }, { "epoch": 0.7998279939797893, "grad_norm": 6.233572006225586, "learning_rate": 8.83090495082465e-06, "loss": 43.3993, "step": 465 }, { "epoch": 0.8015480541818963, "grad_norm": 6.681156635284424, "learning_rate": 8.824754112397467e-06, "loss": 43.2356, "step": 466 }, { "epoch": 0.8032681143840035, "grad_norm": 7.240959167480469, "learning_rate": 8.818589290080043e-06, "loss": 42.8966, "step": 467 }, { "epoch": 0.8049881745861105, "grad_norm": 6.253081798553467, "learning_rate": 8.812410506411925e-06, "loss": 43.9822, "step": 468 }, { "epoch": 0.8067082347882176, "grad_norm": 7.541505813598633, "learning_rate": 8.806217783983693e-06, "loss": 43.9604, "step": 469 }, { "epoch": 0.8084282949903246, "grad_norm": 7.4928483963012695, "learning_rate": 8.800011145436893e-06, "loss": 43.8446, "step": 470 }, { "epoch": 0.8101483551924318, "grad_norm": 6.140499591827393, "learning_rate": 8.793790613463956e-06, "loss": 43.6913, "step": 471 }, { "epoch": 0.8118684153945388, "grad_norm": 7.944373607635498, "learning_rate": 8.787556210808101e-06, "loss": 43.4474, "step": 472 }, { "epoch": 0.8135884755966459, "grad_norm": 6.9422101974487305, "learning_rate": 8.781307960263267e-06, "loss": 43.293, "step": 473 }, { "epoch": 0.815308535798753, "grad_norm": 6.664095878601074, "learning_rate": 8.77504588467402e-06, "loss": 43.8227, "step": 474 }, { "epoch": 0.81702859600086, "grad_norm": 7.298461437225342, "learning_rate": 8.768770006935475e-06, "loss": 43.7175, "step": 475 }, { "epoch": 0.8187486562029671, "grad_norm": 6.43251895904541, "learning_rate": 8.762480349993204e-06, "loss": 43.143, "step": 476 }, { "epoch": 0.8204687164050741, "grad_norm": 6.303859233856201, "learning_rate": 8.756176936843161e-06, "loss": 43.7655, "step": 477 }, { "epoch": 0.8221887766071813, "grad_norm": 6.824503421783447, "learning_rate": 8.749859790531601e-06, "loss": 43.5909, "step": 478 }, { "epoch": 0.8239088368092883, "grad_norm": 6.232965469360352, "learning_rate": 8.743528934154982e-06, "loss": 43.6798, "step": 479 }, { "epoch": 0.8256288970113954, "grad_norm": 6.288873672485352, "learning_rate": 8.737184390859887e-06, "loss": 43.4713, "step": 480 }, { "epoch": 0.8273489572135024, "grad_norm": 6.1072306632995605, "learning_rate": 8.730826183842947e-06, "loss": 43.4521, "step": 481 }, { "epoch": 0.8290690174156096, "grad_norm": 7.3213701248168945, "learning_rate": 8.724454336350742e-06, "loss": 43.9662, "step": 482 }, { "epoch": 0.8307890776177166, "grad_norm": 6.282354354858398, "learning_rate": 8.718068871679735e-06, "loss": 44.1781, "step": 483 }, { "epoch": 0.8325091378198237, "grad_norm": 7.692941188812256, "learning_rate": 8.711669813176165e-06, "loss": 43.5585, "step": 484 }, { "epoch": 0.8342291980219307, "grad_norm": 6.070176124572754, "learning_rate": 8.705257184235973e-06, "loss": 43.843, "step": 485 }, { "epoch": 0.8359492582240379, "grad_norm": 7.584023952484131, "learning_rate": 8.698831008304723e-06, "loss": 43.5888, "step": 486 }, { "epoch": 0.8376693184261449, "grad_norm": 8.037973403930664, "learning_rate": 8.6923913088775e-06, "loss": 43.4765, "step": 487 }, { "epoch": 0.839389378628252, "grad_norm": 6.745630741119385, "learning_rate": 8.685938109498839e-06, "loss": 44.0438, "step": 488 }, { "epoch": 0.841109438830359, "grad_norm": 6.660660743713379, "learning_rate": 8.679471433762633e-06, "loss": 43.5884, "step": 489 }, { "epoch": 0.8428294990324662, "grad_norm": 7.205166339874268, "learning_rate": 8.672991305312042e-06, "loss": 43.5902, "step": 490 }, { "epoch": 0.8445495592345732, "grad_norm": 6.969662666320801, "learning_rate": 8.666497747839413e-06, "loss": 43.339, "step": 491 }, { "epoch": 0.8462696194366803, "grad_norm": 6.3066229820251465, "learning_rate": 8.659990785086195e-06, "loss": 43.8102, "step": 492 }, { "epoch": 0.8479896796387874, "grad_norm": 7.804117202758789, "learning_rate": 8.653470440842847e-06, "loss": 43.7162, "step": 493 }, { "epoch": 0.8497097398408945, "grad_norm": 6.339798450469971, "learning_rate": 8.646936738948747e-06, "loss": 43.3229, "step": 494 }, { "epoch": 0.8514298000430015, "grad_norm": 8.31767749786377, "learning_rate": 8.64038970329212e-06, "loss": 43.8772, "step": 495 }, { "epoch": 0.8531498602451085, "grad_norm": 7.365615367889404, "learning_rate": 8.633829357809937e-06, "loss": 43.2881, "step": 496 }, { "epoch": 0.8548699204472157, "grad_norm": 8.474952697753906, "learning_rate": 8.627255726487831e-06, "loss": 43.7939, "step": 497 }, { "epoch": 0.8565899806493227, "grad_norm": 6.960323333740234, "learning_rate": 8.620668833360009e-06, "loss": 43.499, "step": 498 }, { "epoch": 0.8583100408514298, "grad_norm": 7.6085734367370605, "learning_rate": 8.614068702509169e-06, "loss": 43.2025, "step": 499 }, { "epoch": 0.8600301010535368, "grad_norm": 5.805634498596191, "learning_rate": 8.607455358066404e-06, "loss": 44.0489, "step": 500 }, { "epoch": 0.861750161255644, "grad_norm": 7.452625274658203, "learning_rate": 8.600828824211122e-06, "loss": 43.147, "step": 501 }, { "epoch": 0.863470221457751, "grad_norm": 6.4528584480285645, "learning_rate": 8.594189125170952e-06, "loss": 43.5228, "step": 502 }, { "epoch": 0.8651902816598581, "grad_norm": 7.375026702880859, "learning_rate": 8.587536285221656e-06, "loss": 44.272, "step": 503 }, { "epoch": 0.8669103418619651, "grad_norm": 6.547872543334961, "learning_rate": 8.580870328687041e-06, "loss": 43.4759, "step": 504 }, { "epoch": 0.8686304020640723, "grad_norm": 7.088862419128418, "learning_rate": 8.574191279938872e-06, "loss": 43.5528, "step": 505 }, { "epoch": 0.8703504622661793, "grad_norm": 6.6691083908081055, "learning_rate": 8.567499163396777e-06, "loss": 43.4488, "step": 506 }, { "epoch": 0.8720705224682864, "grad_norm": 7.416652202606201, "learning_rate": 8.560794003528171e-06, "loss": 43.5291, "step": 507 }, { "epoch": 0.8737905826703934, "grad_norm": 5.969050884246826, "learning_rate": 8.554075824848146e-06, "loss": 43.5905, "step": 508 }, { "epoch": 0.8755106428725006, "grad_norm": 7.501400470733643, "learning_rate": 8.5473446519194e-06, "loss": 44.0266, "step": 509 }, { "epoch": 0.8772307030746076, "grad_norm": 6.1612548828125, "learning_rate": 8.540600509352139e-06, "loss": 44.2164, "step": 510 }, { "epoch": 0.8789507632767147, "grad_norm": 7.144975662231445, "learning_rate": 8.533843421803985e-06, "loss": 43.1628, "step": 511 }, { "epoch": 0.8806708234788218, "grad_norm": 6.905309200286865, "learning_rate": 8.527073413979894e-06, "loss": 43.8717, "step": 512 }, { "epoch": 0.8823908836809289, "grad_norm": 7.095192909240723, "learning_rate": 8.520290510632055e-06, "loss": 43.612, "step": 513 }, { "epoch": 0.8841109438830359, "grad_norm": 6.804945468902588, "learning_rate": 8.51349473655981e-06, "loss": 43.4807, "step": 514 }, { "epoch": 0.8858310040851429, "grad_norm": 8.927321434020996, "learning_rate": 8.506686116609553e-06, "loss": 43.5119, "step": 515 }, { "epoch": 0.8875510642872501, "grad_norm": 6.946136474609375, "learning_rate": 8.499864675674648e-06, "loss": 43.1227, "step": 516 }, { "epoch": 0.8892711244893571, "grad_norm": 11.003009796142578, "learning_rate": 8.493030438695336e-06, "loss": 43.2844, "step": 517 }, { "epoch": 0.8909911846914642, "grad_norm": 9.151321411132812, "learning_rate": 8.486183430658639e-06, "loss": 43.9351, "step": 518 }, { "epoch": 0.8927112448935712, "grad_norm": 10.419197082519531, "learning_rate": 8.479323676598271e-06, "loss": 43.4703, "step": 519 }, { "epoch": 0.8944313050956784, "grad_norm": 9.973617553710938, "learning_rate": 8.472451201594556e-06, "loss": 43.2196, "step": 520 }, { "epoch": 0.8961513652977854, "grad_norm": 7.706507682800293, "learning_rate": 8.465566030774314e-06, "loss": 43.3206, "step": 521 }, { "epoch": 0.8978714254998925, "grad_norm": 7.455542087554932, "learning_rate": 8.458668189310793e-06, "loss": 43.2529, "step": 522 }, { "epoch": 0.8995914857019995, "grad_norm": 7.595444202423096, "learning_rate": 8.451757702423566e-06, "loss": 43.5217, "step": 523 }, { "epoch": 0.9013115459041067, "grad_norm": 7.0413899421691895, "learning_rate": 8.444834595378434e-06, "loss": 43.6686, "step": 524 }, { "epoch": 0.9030316061062137, "grad_norm": 7.505328178405762, "learning_rate": 8.437898893487345e-06, "loss": 43.508, "step": 525 }, { "epoch": 0.9047516663083208, "grad_norm": 6.636236667633057, "learning_rate": 8.430950622108292e-06, "loss": 43.6455, "step": 526 }, { "epoch": 0.9064717265104278, "grad_norm": 8.014334678649902, "learning_rate": 8.42398980664523e-06, "loss": 43.9419, "step": 527 }, { "epoch": 0.908191786712535, "grad_norm": 6.861055374145508, "learning_rate": 8.417016472547968e-06, "loss": 44.0091, "step": 528 }, { "epoch": 0.909911846914642, "grad_norm": 7.528046607971191, "learning_rate": 8.41003064531209e-06, "loss": 43.9412, "step": 529 }, { "epoch": 0.9116319071167491, "grad_norm": 6.380741596221924, "learning_rate": 8.403032350478857e-06, "loss": 43.1688, "step": 530 }, { "epoch": 0.9133519673188562, "grad_norm": 8.248329162597656, "learning_rate": 8.396021613635116e-06, "loss": 43.241, "step": 531 } ], "logging_steps": 1, "max_steps": 1743, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 59, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.448505275235225e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }