{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9990999099909991, "eval_steps": 500, "global_step": 555, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0018001800180018, "grad_norm": 28.85651397705078, "learning_rate": 5.000000000000001e-07, "loss": 14.6669, "step": 1 }, { "epoch": 0.0036003600360036, "grad_norm": 20.232847213745117, "learning_rate": 1.0000000000000002e-06, "loss": 12.6879, "step": 2 }, { "epoch": 0.0054005400540054005, "grad_norm": 20.170488357543945, "learning_rate": 1.5e-06, "loss": 9.9005, "step": 3 }, { "epoch": 0.0072007200720072, "grad_norm": 18.635835647583008, "learning_rate": 2.0000000000000003e-06, "loss": 9.5553, "step": 4 }, { "epoch": 0.009000900090009001, "grad_norm": 17.15140724182129, "learning_rate": 2.5e-06, "loss": 8.7361, "step": 5 }, { "epoch": 0.010801080108010801, "grad_norm": 14.188899993896484, "learning_rate": 3e-06, "loss": 8.5991, "step": 6 }, { "epoch": 0.012601260126012601, "grad_norm": 16.08758544921875, "learning_rate": 3.5000000000000004e-06, "loss": 8.1545, "step": 7 }, { "epoch": 0.0144014401440144, "grad_norm": 15.893816947937012, "learning_rate": 4.000000000000001e-06, "loss": 7.9389, "step": 8 }, { "epoch": 0.016201620162016202, "grad_norm": 13.540562629699707, "learning_rate": 4.5e-06, "loss": 7.9457, "step": 9 }, { "epoch": 0.018001800180018002, "grad_norm": 12.593961715698242, "learning_rate": 5e-06, "loss": 7.9617, "step": 10 }, { "epoch": 0.019801980198019802, "grad_norm": 14.274964332580566, "learning_rate": 5.500000000000001e-06, "loss": 7.7485, "step": 11 }, { "epoch": 0.021602160216021602, "grad_norm": 13.785284996032715, "learning_rate": 6e-06, "loss": 8.1615, "step": 12 }, { "epoch": 0.023402340234023402, "grad_norm": 12.692747116088867, "learning_rate": 6.5000000000000004e-06, "loss": 7.9709, "step": 13 }, { "epoch": 0.025202520252025202, "grad_norm": 11.758572578430176, "learning_rate": 7.000000000000001e-06, "loss": 7.5982, "step": 14 }, { "epoch": 0.027002700270027002, "grad_norm": 11.451689720153809, "learning_rate": 7.5e-06, "loss": 7.8241, "step": 15 }, { "epoch": 0.0288028802880288, "grad_norm": 13.09874153137207, "learning_rate": 8.000000000000001e-06, "loss": 7.679, "step": 16 }, { "epoch": 0.0306030603060306, "grad_norm": 11.725943565368652, "learning_rate": 8.500000000000002e-06, "loss": 7.8199, "step": 17 }, { "epoch": 0.032403240324032405, "grad_norm": 12.393065452575684, "learning_rate": 9e-06, "loss": 7.8977, "step": 18 }, { "epoch": 0.034203420342034205, "grad_norm": 11.96495246887207, "learning_rate": 9.5e-06, "loss": 7.4002, "step": 19 }, { "epoch": 0.036003600360036005, "grad_norm": 10.148777961730957, "learning_rate": 1e-05, "loss": 7.6521, "step": 20 }, { "epoch": 0.037803780378037805, "grad_norm": 9.769454956054688, "learning_rate": 1.05e-05, "loss": 7.5508, "step": 21 }, { "epoch": 0.039603960396039604, "grad_norm": 8.934900283813477, "learning_rate": 1.1000000000000001e-05, "loss": 7.263, "step": 22 }, { "epoch": 0.041404140414041404, "grad_norm": 8.807928085327148, "learning_rate": 1.1500000000000002e-05, "loss": 7.4388, "step": 23 }, { "epoch": 0.043204320432043204, "grad_norm": 10.690240859985352, "learning_rate": 1.2e-05, "loss": 6.5005, "step": 24 }, { "epoch": 0.045004500450045004, "grad_norm": 8.784053802490234, "learning_rate": 1.25e-05, "loss": 6.7989, "step": 25 }, { "epoch": 0.046804680468046804, "grad_norm": 8.401625633239746, "learning_rate": 1.3000000000000001e-05, "loss": 6.8476, "step": 26 }, { "epoch": 0.048604860486048604, "grad_norm": 7.974181175231934, "learning_rate": 1.3500000000000001e-05, "loss": 6.8067, "step": 27 }, { "epoch": 0.050405040504050404, "grad_norm": 6.8001909255981445, "learning_rate": 1.4000000000000001e-05, "loss": 6.8886, "step": 28 }, { "epoch": 0.052205220522052204, "grad_norm": 7.849862098693848, "learning_rate": 1.45e-05, "loss": 6.7455, "step": 29 }, { "epoch": 0.054005400540054004, "grad_norm": 7.1335625648498535, "learning_rate": 1.5e-05, "loss": 6.1861, "step": 30 }, { "epoch": 0.0558055805580558, "grad_norm": 6.963135719299316, "learning_rate": 1.55e-05, "loss": 6.2953, "step": 31 }, { "epoch": 0.0576057605760576, "grad_norm": 6.415379524230957, "learning_rate": 1.6000000000000003e-05, "loss": 6.4054, "step": 32 }, { "epoch": 0.0594059405940594, "grad_norm": 6.006982326507568, "learning_rate": 1.65e-05, "loss": 6.6365, "step": 33 }, { "epoch": 0.0612061206120612, "grad_norm": 5.94774866104126, "learning_rate": 1.7000000000000003e-05, "loss": 6.3694, "step": 34 }, { "epoch": 0.063006300630063, "grad_norm": 5.748636722564697, "learning_rate": 1.75e-05, "loss": 6.0299, "step": 35 }, { "epoch": 0.06480648064806481, "grad_norm": 5.941126346588135, "learning_rate": 1.8e-05, "loss": 6.1121, "step": 36 }, { "epoch": 0.0666066606660666, "grad_norm": 5.799063682556152, "learning_rate": 1.85e-05, "loss": 6.1198, "step": 37 }, { "epoch": 0.06840684068406841, "grad_norm": 6.232558250427246, "learning_rate": 1.9e-05, "loss": 5.8275, "step": 38 }, { "epoch": 0.0702070207020702, "grad_norm": 5.436257362365723, "learning_rate": 1.9500000000000003e-05, "loss": 6.6235, "step": 39 }, { "epoch": 0.07200720072007201, "grad_norm": 5.369800090789795, "learning_rate": 2e-05, "loss": 6.3276, "step": 40 }, { "epoch": 0.0738073807380738, "grad_norm": 5.262800693511963, "learning_rate": 2.05e-05, "loss": 5.5985, "step": 41 }, { "epoch": 0.07560756075607561, "grad_norm": 5.9995903968811035, "learning_rate": 2.1e-05, "loss": 7.043, "step": 42 }, { "epoch": 0.0774077407740774, "grad_norm": 5.5932183265686035, "learning_rate": 2.15e-05, "loss": 5.9072, "step": 43 }, { "epoch": 0.07920792079207921, "grad_norm": 5.312334060668945, "learning_rate": 2.2000000000000003e-05, "loss": 7.0407, "step": 44 }, { "epoch": 0.081008100810081, "grad_norm": 5.7829155921936035, "learning_rate": 2.25e-05, "loss": 6.006, "step": 45 }, { "epoch": 0.08280828082808281, "grad_norm": 5.04926872253418, "learning_rate": 2.3000000000000003e-05, "loss": 6.7709, "step": 46 }, { "epoch": 0.0846084608460846, "grad_norm": 5.008914470672607, "learning_rate": 2.35e-05, "loss": 7.0494, "step": 47 }, { "epoch": 0.08640864086408641, "grad_norm": 4.904027938842773, "learning_rate": 2.4e-05, "loss": 6.2418, "step": 48 }, { "epoch": 0.08820882088208822, "grad_norm": 5.208195686340332, "learning_rate": 2.45e-05, "loss": 6.1476, "step": 49 }, { "epoch": 0.09000900090009001, "grad_norm": 5.478446006774902, "learning_rate": 2.5e-05, "loss": 6.1826, "step": 50 }, { "epoch": 0.09180918091809182, "grad_norm": 18.321847915649414, "learning_rate": 2.5500000000000003e-05, "loss": 11.8766, "step": 51 }, { "epoch": 0.09360936093609361, "grad_norm": 22.275039672851562, "learning_rate": 2.6000000000000002e-05, "loss": 9.0741, "step": 52 }, { "epoch": 0.09540954095409541, "grad_norm": 18.914152145385742, "learning_rate": 2.6500000000000004e-05, "loss": 6.9466, "step": 53 }, { "epoch": 0.09720972097209721, "grad_norm": 10.226181030273438, "learning_rate": 2.7000000000000002e-05, "loss": 5.7314, "step": 54 }, { "epoch": 0.09900990099009901, "grad_norm": 7.28556489944458, "learning_rate": 2.7500000000000004e-05, "loss": 4.9557, "step": 55 }, { "epoch": 0.10081008100810081, "grad_norm": 6.028027057647705, "learning_rate": 2.8000000000000003e-05, "loss": 5.4561, "step": 56 }, { "epoch": 0.10261026102610261, "grad_norm": 6.356996536254883, "learning_rate": 2.8499999999999998e-05, "loss": 5.1208, "step": 57 }, { "epoch": 0.10441044104410441, "grad_norm": 5.628063201904297, "learning_rate": 2.9e-05, "loss": 5.31, "step": 58 }, { "epoch": 0.10621062106210621, "grad_norm": 5.183619022369385, "learning_rate": 2.95e-05, "loss": 4.8533, "step": 59 }, { "epoch": 0.10801080108010801, "grad_norm": 5.032276630401611, "learning_rate": 3e-05, "loss": 4.5937, "step": 60 }, { "epoch": 0.10981098109810981, "grad_norm": 4.725021839141846, "learning_rate": 3.05e-05, "loss": 4.8969, "step": 61 }, { "epoch": 0.1116111611161116, "grad_norm": 4.868443965911865, "learning_rate": 3.1e-05, "loss": 5.03, "step": 62 }, { "epoch": 0.11341134113411341, "grad_norm": 4.929288864135742, "learning_rate": 3.15e-05, "loss": 4.9148, "step": 63 }, { "epoch": 0.1152115211521152, "grad_norm": 5.14013671875, "learning_rate": 3.2000000000000005e-05, "loss": 5.2505, "step": 64 }, { "epoch": 0.11701170117011701, "grad_norm": 5.411927223205566, "learning_rate": 3.2500000000000004e-05, "loss": 4.91, "step": 65 }, { "epoch": 0.1188118811881188, "grad_norm": 4.524367809295654, "learning_rate": 3.3e-05, "loss": 5.1703, "step": 66 }, { "epoch": 0.12061206120612061, "grad_norm": 4.979662895202637, "learning_rate": 3.35e-05, "loss": 5.1811, "step": 67 }, { "epoch": 0.1224122412241224, "grad_norm": 4.6250457763671875, "learning_rate": 3.4000000000000007e-05, "loss": 4.8259, "step": 68 }, { "epoch": 0.12421242124212421, "grad_norm": 4.706791400909424, "learning_rate": 3.45e-05, "loss": 4.4952, "step": 69 }, { "epoch": 0.126012601260126, "grad_norm": 4.518699645996094, "learning_rate": 3.5e-05, "loss": 4.7958, "step": 70 }, { "epoch": 0.1278127812781278, "grad_norm": 3.9459586143493652, "learning_rate": 3.55e-05, "loss": 5.0545, "step": 71 }, { "epoch": 0.12961296129612962, "grad_norm": 3.954882860183716, "learning_rate": 3.6e-05, "loss": 4.9021, "step": 72 }, { "epoch": 0.13141314131413143, "grad_norm": 4.027074813842773, "learning_rate": 3.65e-05, "loss": 4.6918, "step": 73 }, { "epoch": 0.1332133213321332, "grad_norm": 4.441482067108154, "learning_rate": 3.7e-05, "loss": 4.9625, "step": 74 }, { "epoch": 0.135013501350135, "grad_norm": 4.631096839904785, "learning_rate": 3.7500000000000003e-05, "loss": 4.7489, "step": 75 }, { "epoch": 0.13681368136813682, "grad_norm": 4.464879989624023, "learning_rate": 3.8e-05, "loss": 4.4667, "step": 76 }, { "epoch": 0.13861386138613863, "grad_norm": 4.17475700378418, "learning_rate": 3.85e-05, "loss": 5.517, "step": 77 }, { "epoch": 0.1404140414041404, "grad_norm": 4.135541915893555, "learning_rate": 3.9000000000000006e-05, "loss": 5.1446, "step": 78 }, { "epoch": 0.1422142214221422, "grad_norm": 4.521326541900635, "learning_rate": 3.9500000000000005e-05, "loss": 5.0185, "step": 79 }, { "epoch": 0.14401440144014402, "grad_norm": 4.14456844329834, "learning_rate": 4e-05, "loss": 4.5989, "step": 80 }, { "epoch": 0.14581458145814583, "grad_norm": 4.918880462646484, "learning_rate": 4.05e-05, "loss": 5.1453, "step": 81 }, { "epoch": 0.1476147614761476, "grad_norm": 3.9074041843414307, "learning_rate": 4.1e-05, "loss": 4.5343, "step": 82 }, { "epoch": 0.1494149414941494, "grad_norm": 4.088869094848633, "learning_rate": 4.15e-05, "loss": 5.2312, "step": 83 }, { "epoch": 0.15121512151215122, "grad_norm": 4.230950355529785, "learning_rate": 4.2e-05, "loss": 5.2262, "step": 84 }, { "epoch": 0.15301530153015303, "grad_norm": 3.9280920028686523, "learning_rate": 4.25e-05, "loss": 5.0867, "step": 85 }, { "epoch": 0.1548154815481548, "grad_norm": 4.369322776794434, "learning_rate": 4.3e-05, "loss": 5.6706, "step": 86 }, { "epoch": 0.1566156615661566, "grad_norm": 4.20168924331665, "learning_rate": 4.35e-05, "loss": 5.2653, "step": 87 }, { "epoch": 0.15841584158415842, "grad_norm": 4.246135711669922, "learning_rate": 4.4000000000000006e-05, "loss": 5.3541, "step": 88 }, { "epoch": 0.16021602160216022, "grad_norm": 4.878535270690918, "learning_rate": 4.4500000000000004e-05, "loss": 5.2614, "step": 89 }, { "epoch": 0.162016201620162, "grad_norm": 4.4837799072265625, "learning_rate": 4.5e-05, "loss": 5.7189, "step": 90 }, { "epoch": 0.1638163816381638, "grad_norm": 4.485711574554443, "learning_rate": 4.55e-05, "loss": 6.1362, "step": 91 }, { "epoch": 0.16561656165616562, "grad_norm": 4.198153495788574, "learning_rate": 4.600000000000001e-05, "loss": 5.5297, "step": 92 }, { "epoch": 0.16741674167416742, "grad_norm": 5.264278411865234, "learning_rate": 4.6500000000000005e-05, "loss": 5.475, "step": 93 }, { "epoch": 0.1692169216921692, "grad_norm": 4.983798503875732, "learning_rate": 4.7e-05, "loss": 5.8982, "step": 94 }, { "epoch": 0.171017101710171, "grad_norm": 4.64500617980957, "learning_rate": 4.75e-05, "loss": 5.4172, "step": 95 }, { "epoch": 0.17281728172817282, "grad_norm": 4.246720314025879, "learning_rate": 4.8e-05, "loss": 6.1043, "step": 96 }, { "epoch": 0.17461746174617462, "grad_norm": 4.144988536834717, "learning_rate": 4.85e-05, "loss": 6.1926, "step": 97 }, { "epoch": 0.17641764176417643, "grad_norm": 4.693668842315674, "learning_rate": 4.9e-05, "loss": 6.3127, "step": 98 }, { "epoch": 0.1782178217821782, "grad_norm": 4.599040985107422, "learning_rate": 4.9500000000000004e-05, "loss": 5.5763, "step": 99 }, { "epoch": 0.18001800180018002, "grad_norm": 4.811699390411377, "learning_rate": 5e-05, "loss": 6.2599, "step": 100 }, { "epoch": 0.18181818181818182, "grad_norm": 23.242341995239258, "learning_rate": 5.05e-05, "loss": 10.2299, "step": 101 }, { "epoch": 0.18361836183618363, "grad_norm": 24.24691390991211, "learning_rate": 5.1000000000000006e-05, "loss": 6.5731, "step": 102 }, { "epoch": 0.1854185418541854, "grad_norm": 13.8417387008667, "learning_rate": 5.1500000000000005e-05, "loss": 4.5894, "step": 103 }, { "epoch": 0.18721872187218722, "grad_norm": 7.606777667999268, "learning_rate": 5.2000000000000004e-05, "loss": 4.422, "step": 104 }, { "epoch": 0.18901890189018902, "grad_norm": 7.0894598960876465, "learning_rate": 5.25e-05, "loss": 4.4822, "step": 105 }, { "epoch": 0.19081908190819083, "grad_norm": 4.703729152679443, "learning_rate": 5.300000000000001e-05, "loss": 4.0663, "step": 106 }, { "epoch": 0.1926192619261926, "grad_norm": 3.9182114601135254, "learning_rate": 5.3500000000000006e-05, "loss": 3.7222, "step": 107 }, { "epoch": 0.19441944194419442, "grad_norm": 4.467214584350586, "learning_rate": 5.4000000000000005e-05, "loss": 4.4871, "step": 108 }, { "epoch": 0.19621962196219622, "grad_norm": 4.333056926727295, "learning_rate": 5.45e-05, "loss": 4.1119, "step": 109 }, { "epoch": 0.19801980198019803, "grad_norm": 4.1896467208862305, "learning_rate": 5.500000000000001e-05, "loss": 3.8633, "step": 110 }, { "epoch": 0.1998199819981998, "grad_norm": 4.409817218780518, "learning_rate": 5.550000000000001e-05, "loss": 4.1133, "step": 111 }, { "epoch": 0.20162016201620162, "grad_norm": 4.587196350097656, "learning_rate": 5.6000000000000006e-05, "loss": 4.085, "step": 112 }, { "epoch": 0.20342034203420342, "grad_norm": 4.2752556800842285, "learning_rate": 5.65e-05, "loss": 4.3178, "step": 113 }, { "epoch": 0.20522052205220523, "grad_norm": 4.687012672424316, "learning_rate": 5.6999999999999996e-05, "loss": 4.4426, "step": 114 }, { "epoch": 0.207020702070207, "grad_norm": 4.1435017585754395, "learning_rate": 5.7499999999999995e-05, "loss": 4.268, "step": 115 }, { "epoch": 0.20882088208820881, "grad_norm": 4.161345481872559, "learning_rate": 5.8e-05, "loss": 4.4851, "step": 116 }, { "epoch": 0.21062106210621062, "grad_norm": 3.927691698074341, "learning_rate": 5.85e-05, "loss": 4.0147, "step": 117 }, { "epoch": 0.21242124212421243, "grad_norm": 4.2265944480896, "learning_rate": 5.9e-05, "loss": 4.3646, "step": 118 }, { "epoch": 0.21422142214221424, "grad_norm": 4.2369704246521, "learning_rate": 5.95e-05, "loss": 4.3241, "step": 119 }, { "epoch": 0.21602160216021601, "grad_norm": 4.384726047515869, "learning_rate": 6e-05, "loss": 4.4394, "step": 120 }, { "epoch": 0.21782178217821782, "grad_norm": 4.316718578338623, "learning_rate": 6.05e-05, "loss": 4.0843, "step": 121 }, { "epoch": 0.21962196219621963, "grad_norm": 3.7887487411499023, "learning_rate": 6.1e-05, "loss": 4.1607, "step": 122 }, { "epoch": 0.22142214221422143, "grad_norm": 4.181461334228516, "learning_rate": 6.15e-05, "loss": 4.2074, "step": 123 }, { "epoch": 0.2232223222322232, "grad_norm": 3.9978582859039307, "learning_rate": 6.2e-05, "loss": 4.4763, "step": 124 }, { "epoch": 0.22502250225022502, "grad_norm": 4.156275749206543, "learning_rate": 6.25e-05, "loss": 4.2295, "step": 125 }, { "epoch": 0.22682268226822683, "grad_norm": 3.867684841156006, "learning_rate": 6.3e-05, "loss": 4.4481, "step": 126 }, { "epoch": 0.22862286228622863, "grad_norm": 4.037459373474121, "learning_rate": 6.35e-05, "loss": 4.157, "step": 127 }, { "epoch": 0.2304230423042304, "grad_norm": 4.304506778717041, "learning_rate": 6.400000000000001e-05, "loss": 4.5553, "step": 128 }, { "epoch": 0.23222322232223222, "grad_norm": 4.406405925750732, "learning_rate": 6.450000000000001e-05, "loss": 4.7441, "step": 129 }, { "epoch": 0.23402340234023403, "grad_norm": 4.19051456451416, "learning_rate": 6.500000000000001e-05, "loss": 4.7459, "step": 130 }, { "epoch": 0.23582358235823583, "grad_norm": 4.138574600219727, "learning_rate": 6.55e-05, "loss": 4.6568, "step": 131 }, { "epoch": 0.2376237623762376, "grad_norm": 4.434947967529297, "learning_rate": 6.6e-05, "loss": 4.8297, "step": 132 }, { "epoch": 0.23942394239423942, "grad_norm": 3.5757968425750732, "learning_rate": 6.65e-05, "loss": 4.2438, "step": 133 }, { "epoch": 0.24122412241224123, "grad_norm": 4.133842945098877, "learning_rate": 6.7e-05, "loss": 4.5298, "step": 134 }, { "epoch": 0.24302430243024303, "grad_norm": 4.089065074920654, "learning_rate": 6.750000000000001e-05, "loss": 5.4232, "step": 135 }, { "epoch": 0.2448244824482448, "grad_norm": 4.211568832397461, "learning_rate": 6.800000000000001e-05, "loss": 5.3946, "step": 136 }, { "epoch": 0.24662466246624662, "grad_norm": 4.204323768615723, "learning_rate": 6.850000000000001e-05, "loss": 4.9143, "step": 137 }, { "epoch": 0.24842484248424843, "grad_norm": 4.3052873611450195, "learning_rate": 6.9e-05, "loss": 5.5809, "step": 138 }, { "epoch": 0.2502250225022502, "grad_norm": 3.8890812397003174, "learning_rate": 6.95e-05, "loss": 5.0493, "step": 139 }, { "epoch": 0.252025202520252, "grad_norm": 4.273584842681885, "learning_rate": 7e-05, "loss": 5.45, "step": 140 }, { "epoch": 0.2538253825382538, "grad_norm": 4.8961358070373535, "learning_rate": 7.05e-05, "loss": 5.5538, "step": 141 }, { "epoch": 0.2556255625562556, "grad_norm": 4.320993423461914, "learning_rate": 7.1e-05, "loss": 5.3233, "step": 142 }, { "epoch": 0.25742574257425743, "grad_norm": 3.967057704925537, "learning_rate": 7.15e-05, "loss": 5.1942, "step": 143 }, { "epoch": 0.25922592259225924, "grad_norm": 4.173534870147705, "learning_rate": 7.2e-05, "loss": 5.1138, "step": 144 }, { "epoch": 0.26102610261026105, "grad_norm": 4.54023551940918, "learning_rate": 7.25e-05, "loss": 5.6945, "step": 145 }, { "epoch": 0.26282628262826285, "grad_norm": 4.46029806137085, "learning_rate": 7.3e-05, "loss": 5.5752, "step": 146 }, { "epoch": 0.2646264626462646, "grad_norm": 4.0620574951171875, "learning_rate": 7.35e-05, "loss": 5.2059, "step": 147 }, { "epoch": 0.2664266426642664, "grad_norm": 4.501194477081299, "learning_rate": 7.4e-05, "loss": 5.6742, "step": 148 }, { "epoch": 0.2682268226822682, "grad_norm": 4.779675483703613, "learning_rate": 7.450000000000001e-05, "loss": 6.3616, "step": 149 }, { "epoch": 0.27002700270027, "grad_norm": 5.094762802124023, "learning_rate": 7.500000000000001e-05, "loss": 5.8473, "step": 150 }, { "epoch": 0.27182718271827183, "grad_norm": 30.64379119873047, "learning_rate": 7.55e-05, "loss": 6.9989, "step": 151 }, { "epoch": 0.27362736273627364, "grad_norm": 11.478928565979004, "learning_rate": 7.6e-05, "loss": 3.2922, "step": 152 }, { "epoch": 0.27542754275427545, "grad_norm": 5.4645466804504395, "learning_rate": 7.65e-05, "loss": 3.4407, "step": 153 }, { "epoch": 0.27722772277227725, "grad_norm": 5.904770374298096, "learning_rate": 7.7e-05, "loss": 3.5843, "step": 154 }, { "epoch": 0.279027902790279, "grad_norm": 4.119198322296143, "learning_rate": 7.75e-05, "loss": 3.1653, "step": 155 }, { "epoch": 0.2808280828082808, "grad_norm": 4.040031909942627, "learning_rate": 7.800000000000001e-05, "loss": 3.4589, "step": 156 }, { "epoch": 0.2826282628262826, "grad_norm": 4.188577175140381, "learning_rate": 7.850000000000001e-05, "loss": 3.7084, "step": 157 }, { "epoch": 0.2844284428442844, "grad_norm": 4.088983058929443, "learning_rate": 7.900000000000001e-05, "loss": 3.8008, "step": 158 }, { "epoch": 0.28622862286228623, "grad_norm": 3.6383821964263916, "learning_rate": 7.950000000000001e-05, "loss": 3.9436, "step": 159 }, { "epoch": 0.28802880288028804, "grad_norm": 4.231886863708496, "learning_rate": 8e-05, "loss": 4.5073, "step": 160 }, { "epoch": 0.28982898289828984, "grad_norm": 4.518820762634277, "learning_rate": 8.05e-05, "loss": 3.5422, "step": 161 }, { "epoch": 0.29162916291629165, "grad_norm": 4.026054382324219, "learning_rate": 8.1e-05, "loss": 3.7513, "step": 162 }, { "epoch": 0.2934293429342934, "grad_norm": 3.7655909061431885, "learning_rate": 8.15e-05, "loss": 4.2454, "step": 163 }, { "epoch": 0.2952295229522952, "grad_norm": 3.6118693351745605, "learning_rate": 8.2e-05, "loss": 3.7299, "step": 164 }, { "epoch": 0.297029702970297, "grad_norm": 4.284143447875977, "learning_rate": 8.25e-05, "loss": 4.07, "step": 165 }, { "epoch": 0.2988298829882988, "grad_norm": 4.056666374206543, "learning_rate": 8.3e-05, "loss": 4.1954, "step": 166 }, { "epoch": 0.30063006300630063, "grad_norm": 4.1431565284729, "learning_rate": 8.35e-05, "loss": 4.252, "step": 167 }, { "epoch": 0.30243024302430244, "grad_norm": 4.241552829742432, "learning_rate": 8.4e-05, "loss": 4.1166, "step": 168 }, { "epoch": 0.30423042304230424, "grad_norm": 4.314858913421631, "learning_rate": 8.450000000000001e-05, "loss": 4.0205, "step": 169 }, { "epoch": 0.30603060306030605, "grad_norm": 4.8378400802612305, "learning_rate": 8.5e-05, "loss": 4.4915, "step": 170 }, { "epoch": 0.30783078307830786, "grad_norm": 3.758328437805176, "learning_rate": 8.55e-05, "loss": 4.2785, "step": 171 }, { "epoch": 0.3096309630963096, "grad_norm": 4.4342360496521, "learning_rate": 8.6e-05, "loss": 4.2285, "step": 172 }, { "epoch": 0.3114311431143114, "grad_norm": 4.5404052734375, "learning_rate": 8.65e-05, "loss": 4.3089, "step": 173 }, { "epoch": 0.3132313231323132, "grad_norm": 3.7609307765960693, "learning_rate": 8.7e-05, "loss": 4.3388, "step": 174 }, { "epoch": 0.31503150315031503, "grad_norm": 4.27951717376709, "learning_rate": 8.75e-05, "loss": 4.3054, "step": 175 }, { "epoch": 0.31683168316831684, "grad_norm": 4.713607311248779, "learning_rate": 8.800000000000001e-05, "loss": 4.5114, "step": 176 }, { "epoch": 0.31863186318631864, "grad_norm": 3.725269079208374, "learning_rate": 8.850000000000001e-05, "loss": 4.4626, "step": 177 }, { "epoch": 0.32043204320432045, "grad_norm": 4.011518478393555, "learning_rate": 8.900000000000001e-05, "loss": 4.4429, "step": 178 }, { "epoch": 0.32223222322232226, "grad_norm": 4.707291603088379, "learning_rate": 8.950000000000001e-05, "loss": 4.3184, "step": 179 }, { "epoch": 0.324032403240324, "grad_norm": 4.8839192390441895, "learning_rate": 9e-05, "loss": 4.9505, "step": 180 }, { "epoch": 0.3258325832583258, "grad_norm": 4.054986953735352, "learning_rate": 9.05e-05, "loss": 4.754, "step": 181 }, { "epoch": 0.3276327632763276, "grad_norm": 3.74249529838562, "learning_rate": 9.1e-05, "loss": 4.5569, "step": 182 }, { "epoch": 0.32943294329432943, "grad_norm": 4.3617939949035645, "learning_rate": 9.15e-05, "loss": 4.1692, "step": 183 }, { "epoch": 0.33123312331233123, "grad_norm": 4.2513346672058105, "learning_rate": 9.200000000000001e-05, "loss": 4.6149, "step": 184 }, { "epoch": 0.33303330333033304, "grad_norm": 4.553803443908691, "learning_rate": 9.250000000000001e-05, "loss": 4.5845, "step": 185 }, { "epoch": 0.33483348334833485, "grad_norm": 3.855337619781494, "learning_rate": 9.300000000000001e-05, "loss": 4.5672, "step": 186 }, { "epoch": 0.33663366336633666, "grad_norm": 4.010441303253174, "learning_rate": 9.350000000000001e-05, "loss": 5.1355, "step": 187 }, { "epoch": 0.3384338433843384, "grad_norm": 4.344321250915527, "learning_rate": 9.4e-05, "loss": 5.2234, "step": 188 }, { "epoch": 0.3402340234023402, "grad_norm": 3.883000373840332, "learning_rate": 9.449999999999999e-05, "loss": 4.6074, "step": 189 }, { "epoch": 0.342034203420342, "grad_norm": 4.245122909545898, "learning_rate": 9.5e-05, "loss": 4.7656, "step": 190 }, { "epoch": 0.3438343834383438, "grad_norm": 4.1966023445129395, "learning_rate": 9.55e-05, "loss": 5.1034, "step": 191 }, { "epoch": 0.34563456345634563, "grad_norm": 4.888998031616211, "learning_rate": 9.6e-05, "loss": 4.543, "step": 192 }, { "epoch": 0.34743474347434744, "grad_norm": 4.0577311515808105, "learning_rate": 9.65e-05, "loss": 5.0657, "step": 193 }, { "epoch": 0.34923492349234925, "grad_norm": 4.154134750366211, "learning_rate": 9.7e-05, "loss": 5.1474, "step": 194 }, { "epoch": 0.35103510351035105, "grad_norm": 4.5482282638549805, "learning_rate": 9.75e-05, "loss": 5.611, "step": 195 }, { "epoch": 0.35283528352835286, "grad_norm": 4.7754292488098145, "learning_rate": 9.8e-05, "loss": 5.8798, "step": 196 }, { "epoch": 0.3546354635463546, "grad_norm": 4.595111846923828, "learning_rate": 9.850000000000001e-05, "loss": 5.5665, "step": 197 }, { "epoch": 0.3564356435643564, "grad_norm": 4.434641361236572, "learning_rate": 9.900000000000001e-05, "loss": 5.017, "step": 198 }, { "epoch": 0.3582358235823582, "grad_norm": 4.628114223480225, "learning_rate": 9.95e-05, "loss": 5.4085, "step": 199 }, { "epoch": 0.36003600360036003, "grad_norm": 4.574793815612793, "learning_rate": 0.0001, "loss": 4.868, "step": 200 }, { "epoch": 0.36183618361836184, "grad_norm": 48.827030181884766, "learning_rate": 9.999804214640151e-05, "loss": 5.6442, "step": 201 }, { "epoch": 0.36363636363636365, "grad_norm": 7.539086818695068, "learning_rate": 9.999216873893363e-05, "loss": 2.9833, "step": 202 }, { "epoch": 0.36543654365436545, "grad_norm": 4.620749473571777, "learning_rate": 9.998238023756727e-05, "loss": 2.9572, "step": 203 }, { "epoch": 0.36723672367236726, "grad_norm": 3.8858468532562256, "learning_rate": 9.996867740888052e-05, "loss": 3.4909, "step": 204 }, { "epoch": 0.369036903690369, "grad_norm": 4.157265663146973, "learning_rate": 9.995106132599869e-05, "loss": 3.794, "step": 205 }, { "epoch": 0.3708370837083708, "grad_norm": 3.5816516876220703, "learning_rate": 9.99295333685102e-05, "loss": 3.6478, "step": 206 }, { "epoch": 0.3726372637263726, "grad_norm": 4.691586494445801, "learning_rate": 9.990409522235866e-05, "loss": 3.6508, "step": 207 }, { "epoch": 0.37443744374437443, "grad_norm": 4.206489562988281, "learning_rate": 9.987474887971067e-05, "loss": 4.0346, "step": 208 }, { "epoch": 0.37623762376237624, "grad_norm": 4.957794189453125, "learning_rate": 9.984149663879993e-05, "loss": 3.9599, "step": 209 }, { "epoch": 0.37803780378037805, "grad_norm": 3.705470561981201, "learning_rate": 9.980434110374724e-05, "loss": 3.7593, "step": 210 }, { "epoch": 0.37983798379837985, "grad_norm": 3.760009288787842, "learning_rate": 9.976328518435653e-05, "loss": 3.6834, "step": 211 }, { "epoch": 0.38163816381638166, "grad_norm": 4.849155902862549, "learning_rate": 9.971833209588696e-05, "loss": 3.7644, "step": 212 }, { "epoch": 0.38343834383438347, "grad_norm": 4.008197784423828, "learning_rate": 9.966948535880118e-05, "loss": 3.9239, "step": 213 }, { "epoch": 0.3852385238523852, "grad_norm": 4.340366840362549, "learning_rate": 9.961674879848957e-05, "loss": 3.8541, "step": 214 }, { "epoch": 0.387038703870387, "grad_norm": 4.3433332443237305, "learning_rate": 9.956012654497074e-05, "loss": 3.6372, "step": 215 }, { "epoch": 0.38883888388838883, "grad_norm": 3.838965654373169, "learning_rate": 9.949962303256796e-05, "loss": 4.1607, "step": 216 }, { "epoch": 0.39063906390639064, "grad_norm": 3.958833932876587, "learning_rate": 9.943524299956207e-05, "loss": 4.0357, "step": 217 }, { "epoch": 0.39243924392439244, "grad_norm": 4.2294840812683105, "learning_rate": 9.936699148782019e-05, "loss": 4.1384, "step": 218 }, { "epoch": 0.39423942394239425, "grad_norm": 4.255922317504883, "learning_rate": 9.929487384240103e-05, "loss": 4.3947, "step": 219 }, { "epoch": 0.39603960396039606, "grad_norm": 4.967748641967773, "learning_rate": 9.921889571113628e-05, "loss": 4.2571, "step": 220 }, { "epoch": 0.39783978397839787, "grad_norm": 5.7947845458984375, "learning_rate": 9.913906304418825e-05, "loss": 4.3985, "step": 221 }, { "epoch": 0.3996399639963996, "grad_norm": 4.166469097137451, "learning_rate": 9.90553820935839e-05, "loss": 4.3324, "step": 222 }, { "epoch": 0.4014401440144014, "grad_norm": 3.7093265056610107, "learning_rate": 9.896785941272523e-05, "loss": 4.4313, "step": 223 }, { "epoch": 0.40324032403240323, "grad_norm": 4.43189001083374, "learning_rate": 9.887650185587612e-05, "loss": 4.723, "step": 224 }, { "epoch": 0.40504050405040504, "grad_norm": 4.146303653717041, "learning_rate": 9.878131657762535e-05, "loss": 4.555, "step": 225 }, { "epoch": 0.40684068406840684, "grad_norm": 4.075449466705322, "learning_rate": 9.868231103232655e-05, "loss": 4.4108, "step": 226 }, { "epoch": 0.40864086408640865, "grad_norm": 4.104311943054199, "learning_rate": 9.857949297351422e-05, "loss": 4.867, "step": 227 }, { "epoch": 0.41044104410441046, "grad_norm": 3.903787612915039, "learning_rate": 9.847287045329666e-05, "loss": 4.3154, "step": 228 }, { "epoch": 0.41224122412241226, "grad_norm": 4.177591800689697, "learning_rate": 9.83624518217252e-05, "loss": 4.3139, "step": 229 }, { "epoch": 0.414041404140414, "grad_norm": 3.577852725982666, "learning_rate": 9.824824572614051e-05, "loss": 4.6006, "step": 230 }, { "epoch": 0.4158415841584158, "grad_norm": 4.238375186920166, "learning_rate": 9.813026111049515e-05, "loss": 4.1235, "step": 231 }, { "epoch": 0.41764176417641763, "grad_norm": 4.316523551940918, "learning_rate": 9.800850721465334e-05, "loss": 4.9625, "step": 232 }, { "epoch": 0.41944194419441944, "grad_norm": 3.8222544193267822, "learning_rate": 9.788299357366716e-05, "loss": 4.843, "step": 233 }, { "epoch": 0.42124212421242124, "grad_norm": 4.11749792098999, "learning_rate": 9.775373001702998e-05, "loss": 4.5882, "step": 234 }, { "epoch": 0.42304230423042305, "grad_norm": 3.6795084476470947, "learning_rate": 9.762072666790658e-05, "loss": 5.2389, "step": 235 }, { "epoch": 0.42484248424842486, "grad_norm": 3.996441602706909, "learning_rate": 9.748399394234038e-05, "loss": 4.6672, "step": 236 }, { "epoch": 0.42664266426642666, "grad_norm": 4.023096561431885, "learning_rate": 9.734354254843773e-05, "loss": 5.3549, "step": 237 }, { "epoch": 0.42844284428442847, "grad_norm": 4.174993515014648, "learning_rate": 9.71993834855293e-05, "loss": 5.1587, "step": 238 }, { "epoch": 0.4302430243024302, "grad_norm": 4.372593879699707, "learning_rate": 9.705152804330872e-05, "loss": 5.6517, "step": 239 }, { "epoch": 0.43204320432043203, "grad_norm": 4.535490989685059, "learning_rate": 9.689998780094837e-05, "loss": 5.4977, "step": 240 }, { "epoch": 0.43384338433843384, "grad_norm": 4.093436241149902, "learning_rate": 9.67447746261926e-05, "loss": 4.3227, "step": 241 }, { "epoch": 0.43564356435643564, "grad_norm": 4.311260223388672, "learning_rate": 9.658590067442829e-05, "loss": 5.137, "step": 242 }, { "epoch": 0.43744374437443745, "grad_norm": 4.544931411743164, "learning_rate": 9.6423378387733e-05, "loss": 5.3618, "step": 243 }, { "epoch": 0.43924392439243926, "grad_norm": 4.6238298416137695, "learning_rate": 9.625722049390046e-05, "loss": 5.7622, "step": 244 }, { "epoch": 0.44104410441044106, "grad_norm": 4.168641567230225, "learning_rate": 9.608744000544392e-05, "loss": 5.2432, "step": 245 }, { "epoch": 0.44284428442844287, "grad_norm": 4.634570598602295, "learning_rate": 9.591405021857695e-05, "loss": 5.8623, "step": 246 }, { "epoch": 0.4446444644464446, "grad_norm": 4.366397380828857, "learning_rate": 9.573706471217232e-05, "loss": 5.4993, "step": 247 }, { "epoch": 0.4464446444644464, "grad_norm": 4.067739963531494, "learning_rate": 9.55564973466984e-05, "loss": 5.6379, "step": 248 }, { "epoch": 0.44824482448244823, "grad_norm": 3.8812038898468018, "learning_rate": 9.537236226313389e-05, "loss": 5.4358, "step": 249 }, { "epoch": 0.45004500450045004, "grad_norm": 5.2032318115234375, "learning_rate": 9.51846738818602e-05, "loss": 5.2139, "step": 250 }, { "epoch": 0.45184518451845185, "grad_norm": 64.75902557373047, "learning_rate": 9.499344690153225e-05, "loss": 6.1023, "step": 251 }, { "epoch": 0.45364536453645365, "grad_norm": 35.30427932739258, "learning_rate": 9.479869629792728e-05, "loss": 3.4732, "step": 252 }, { "epoch": 0.45544554455445546, "grad_norm": 4.7051496505737305, "learning_rate": 9.460043732277213e-05, "loss": 2.73, "step": 253 }, { "epoch": 0.45724572457245727, "grad_norm": 3.9896416664123535, "learning_rate": 9.439868550254869e-05, "loss": 3.2622, "step": 254 }, { "epoch": 0.459045904590459, "grad_norm": 3.751338005065918, "learning_rate": 9.419345663727805e-05, "loss": 3.0939, "step": 255 }, { "epoch": 0.4608460846084608, "grad_norm": 3.8281116485595703, "learning_rate": 9.398476679928313e-05, "loss": 3.9917, "step": 256 }, { "epoch": 0.46264626462646263, "grad_norm": 3.47261905670166, "learning_rate": 9.377263233192992e-05, "loss": 3.471, "step": 257 }, { "epoch": 0.46444644464446444, "grad_norm": 3.8381025791168213, "learning_rate": 9.355706984834763e-05, "loss": 3.7771, "step": 258 }, { "epoch": 0.46624662466246625, "grad_norm": 4.145185470581055, "learning_rate": 9.333809623012763e-05, "loss": 3.7653, "step": 259 }, { "epoch": 0.46804680468046805, "grad_norm": 3.826444387435913, "learning_rate": 9.311572862600139e-05, "loss": 3.7425, "step": 260 }, { "epoch": 0.46984698469846986, "grad_norm": 3.433979034423828, "learning_rate": 9.288998445049746e-05, "loss": 3.4126, "step": 261 }, { "epoch": 0.47164716471647167, "grad_norm": 4.036341667175293, "learning_rate": 9.266088138257768e-05, "loss": 3.6221, "step": 262 }, { "epoch": 0.4734473447344735, "grad_norm": 3.754890203475952, "learning_rate": 9.24284373642527e-05, "loss": 4.0088, "step": 263 }, { "epoch": 0.4752475247524752, "grad_norm": 3.838566780090332, "learning_rate": 9.219267059917682e-05, "loss": 3.6828, "step": 264 }, { "epoch": 0.47704770477047703, "grad_norm": 3.6993019580841064, "learning_rate": 9.195359955122244e-05, "loss": 3.7106, "step": 265 }, { "epoch": 0.47884788478847884, "grad_norm": 3.721072196960449, "learning_rate": 9.171124294303399e-05, "loss": 3.9501, "step": 266 }, { "epoch": 0.48064806480648065, "grad_norm": 3.40067982673645, "learning_rate": 9.14656197545618e-05, "loss": 3.6647, "step": 267 }, { "epoch": 0.48244824482448245, "grad_norm": 4.225182056427002, "learning_rate": 9.121674922157558e-05, "loss": 4.2479, "step": 268 }, { "epoch": 0.48424842484248426, "grad_norm": 3.574556589126587, "learning_rate": 9.096465083415808e-05, "loss": 4.3452, "step": 269 }, { "epoch": 0.48604860486048607, "grad_norm": 3.840705394744873, "learning_rate": 9.070934433517873e-05, "loss": 4.7066, "step": 270 }, { "epoch": 0.4878487848784879, "grad_norm": 3.4904778003692627, "learning_rate": 9.045084971874738e-05, "loss": 3.7437, "step": 271 }, { "epoch": 0.4896489648964896, "grad_norm": 3.8224945068359375, "learning_rate": 9.018918722864867e-05, "loss": 4.1212, "step": 272 }, { "epoch": 0.49144914491449143, "grad_norm": 4.084190368652344, "learning_rate": 8.992437735675651e-05, "loss": 3.883, "step": 273 }, { "epoch": 0.49324932493249324, "grad_norm": 4.186155796051025, "learning_rate": 8.965644084142932e-05, "loss": 4.6328, "step": 274 }, { "epoch": 0.49504950495049505, "grad_norm": 4.2307047843933105, "learning_rate": 8.938539866588592e-05, "loss": 3.9238, "step": 275 }, { "epoch": 0.49684968496849685, "grad_norm": 4.097965717315674, "learning_rate": 8.911127205656226e-05, "loss": 4.4123, "step": 276 }, { "epoch": 0.49864986498649866, "grad_norm": 3.495042562484741, "learning_rate": 8.88340824814491e-05, "loss": 3.8635, "step": 277 }, { "epoch": 0.5004500450045004, "grad_norm": 3.689110517501831, "learning_rate": 8.855385164841072e-05, "loss": 4.6177, "step": 278 }, { "epoch": 0.5022502250225023, "grad_norm": 3.6329872608184814, "learning_rate": 8.82706015034849e-05, "loss": 4.2534, "step": 279 }, { "epoch": 0.504050405040504, "grad_norm": 3.8967461585998535, "learning_rate": 8.798435422916425e-05, "loss": 3.7907, "step": 280 }, { "epoch": 0.5058505850585059, "grad_norm": 4.086053848266602, "learning_rate": 8.769513224265904e-05, "loss": 4.5879, "step": 281 }, { "epoch": 0.5076507650765076, "grad_norm": 4.173221111297607, "learning_rate": 8.740295819414155e-05, "loss": 4.1702, "step": 282 }, { "epoch": 0.5094509450945095, "grad_norm": 3.921698570251465, "learning_rate": 8.710785496497225e-05, "loss": 4.2377, "step": 283 }, { "epoch": 0.5112511251125113, "grad_norm": 3.830862045288086, "learning_rate": 8.680984566590793e-05, "loss": 4.3651, "step": 284 }, { "epoch": 0.513051305130513, "grad_norm": 3.77140212059021, "learning_rate": 8.650895363529173e-05, "loss": 4.4687, "step": 285 }, { "epoch": 0.5148514851485149, "grad_norm": 3.5552985668182373, "learning_rate": 8.620520243722541e-05, "loss": 4.8082, "step": 286 }, { "epoch": 0.5166516651665166, "grad_norm": 3.6477136611938477, "learning_rate": 8.589861585972407e-05, "loss": 5.2562, "step": 287 }, { "epoch": 0.5184518451845185, "grad_norm": 4.382421493530273, "learning_rate": 8.558921791285304e-05, "loss": 5.0799, "step": 288 }, { "epoch": 0.5202520252025202, "grad_norm": 4.065561294555664, "learning_rate": 8.527703282684766e-05, "loss": 4.9109, "step": 289 }, { "epoch": 0.5220522052205221, "grad_norm": 3.5327258110046387, "learning_rate": 8.496208505021571e-05, "loss": 4.2811, "step": 290 }, { "epoch": 0.5238523852385238, "grad_norm": 4.304015636444092, "learning_rate": 8.464439924782271e-05, "loss": 5.0544, "step": 291 }, { "epoch": 0.5256525652565257, "grad_norm": 4.300441265106201, "learning_rate": 8.432400029896028e-05, "loss": 5.1385, "step": 292 }, { "epoch": 0.5274527452745275, "grad_norm": 3.970590114593506, "learning_rate": 8.400091329539784e-05, "loss": 5.5669, "step": 293 }, { "epoch": 0.5292529252925292, "grad_norm": 4.322851657867432, "learning_rate": 8.367516353941751e-05, "loss": 5.1609, "step": 294 }, { "epoch": 0.5310531053105311, "grad_norm": 4.039634704589844, "learning_rate": 8.334677654183254e-05, "loss": 5.3988, "step": 295 }, { "epoch": 0.5328532853285328, "grad_norm": 4.090119361877441, "learning_rate": 8.301577801998954e-05, "loss": 5.3261, "step": 296 }, { "epoch": 0.5346534653465347, "grad_norm": 4.731489181518555, "learning_rate": 8.26821938957544e-05, "loss": 5.4674, "step": 297 }, { "epoch": 0.5364536453645364, "grad_norm": 4.765017032623291, "learning_rate": 8.234605029348223e-05, "loss": 5.5113, "step": 298 }, { "epoch": 0.5382538253825383, "grad_norm": 4.620230197906494, "learning_rate": 8.200737353797149e-05, "loss": 6.6506, "step": 299 }, { "epoch": 0.54005400540054, "grad_norm": 4.256903648376465, "learning_rate": 8.166619015240236e-05, "loss": 5.401, "step": 300 }, { "epoch": 0.5418541854185418, "grad_norm": 72.8096694946289, "learning_rate": 8.132252685625958e-05, "loss": 5.5679, "step": 301 }, { "epoch": 0.5436543654365437, "grad_norm": 12.82461166381836, "learning_rate": 8.097641056324004e-05, "loss": 2.3965, "step": 302 }, { "epoch": 0.5454545454545454, "grad_norm": 3.8120839595794678, "learning_rate": 8.062786837914492e-05, "loss": 2.5916, "step": 303 }, { "epoch": 0.5472547254725473, "grad_norm": 4.205955505371094, "learning_rate": 8.027692759975695e-05, "loss": 3.1019, "step": 304 }, { "epoch": 0.549054905490549, "grad_norm": 3.6653831005096436, "learning_rate": 7.992361570870288e-05, "loss": 3.0688, "step": 305 }, { "epoch": 0.5508550855085509, "grad_norm": 3.9019699096679688, "learning_rate": 7.956796037530099e-05, "loss": 3.4956, "step": 306 }, { "epoch": 0.5526552655265526, "grad_norm": 4.44380521774292, "learning_rate": 7.920998945239427e-05, "loss": 3.8163, "step": 307 }, { "epoch": 0.5544554455445545, "grad_norm": 4.306615352630615, "learning_rate": 7.884973097416908e-05, "loss": 3.8254, "step": 308 }, { "epoch": 0.5562556255625563, "grad_norm": 3.5584664344787598, "learning_rate": 7.848721315395974e-05, "loss": 3.9131, "step": 309 }, { "epoch": 0.558055805580558, "grad_norm": 3.8743574619293213, "learning_rate": 7.812246438203904e-05, "loss": 3.7078, "step": 310 }, { "epoch": 0.5598559855985599, "grad_norm": 3.7516815662384033, "learning_rate": 7.775551322339476e-05, "loss": 3.8186, "step": 311 }, { "epoch": 0.5616561656165616, "grad_norm": 3.739532232284546, "learning_rate": 7.738638841549277e-05, "loss": 3.7782, "step": 312 }, { "epoch": 0.5634563456345635, "grad_norm": 3.378246784210205, "learning_rate": 7.701511886602643e-05, "loss": 3.703, "step": 313 }, { "epoch": 0.5652565256525652, "grad_norm": 3.18391752243042, "learning_rate": 7.664173365065265e-05, "loss": 3.7862, "step": 314 }, { "epoch": 0.5670567056705671, "grad_norm": 3.660829782485962, "learning_rate": 7.626626201071494e-05, "loss": 4.0942, "step": 315 }, { "epoch": 0.5688568856885688, "grad_norm": 3.49433970451355, "learning_rate": 7.588873335095337e-05, "loss": 4.4192, "step": 316 }, { "epoch": 0.5706570657065707, "grad_norm": 3.631074905395508, "learning_rate": 7.550917723720172e-05, "loss": 4.0662, "step": 317 }, { "epoch": 0.5724572457245725, "grad_norm": 3.1131577491760254, "learning_rate": 7.512762339407214e-05, "loss": 3.5052, "step": 318 }, { "epoch": 0.5742574257425742, "grad_norm": 3.5523805618286133, "learning_rate": 7.474410170262718e-05, "loss": 4.2908, "step": 319 }, { "epoch": 0.5760576057605761, "grad_norm": 3.9423253536224365, "learning_rate": 7.435864219803983e-05, "loss": 4.3511, "step": 320 }, { "epoch": 0.5778577857785778, "grad_norm": 3.6397900581359863, "learning_rate": 7.39712750672412e-05, "loss": 4.0423, "step": 321 }, { "epoch": 0.5796579657965797, "grad_norm": 3.341845750808716, "learning_rate": 7.358203064655653e-05, "loss": 4.0845, "step": 322 }, { "epoch": 0.5814581458145814, "grad_norm": 3.8859944343566895, "learning_rate": 7.31909394193294e-05, "loss": 3.6363, "step": 323 }, { "epoch": 0.5832583258325833, "grad_norm": 3.2730302810668945, "learning_rate": 7.27980320135345e-05, "loss": 3.9988, "step": 324 }, { "epoch": 0.585058505850585, "grad_norm": 3.4280858039855957, "learning_rate": 7.240333919937893e-05, "loss": 3.9773, "step": 325 }, { "epoch": 0.5868586858685868, "grad_norm": 3.675854444503784, "learning_rate": 7.200689188689256e-05, "loss": 4.4474, "step": 326 }, { "epoch": 0.5886588658865887, "grad_norm": 3.922302007675171, "learning_rate": 7.16087211235073e-05, "loss": 4.1046, "step": 327 }, { "epoch": 0.5904590459045904, "grad_norm": 4.065755367279053, "learning_rate": 7.120885809162561e-05, "loss": 4.6008, "step": 328 }, { "epoch": 0.5922592259225923, "grad_norm": 4.271775722503662, "learning_rate": 7.080733410617853e-05, "loss": 4.2479, "step": 329 }, { "epoch": 0.594059405940594, "grad_norm": 3.5132906436920166, "learning_rate": 7.040418061217325e-05, "loss": 4.0792, "step": 330 }, { "epoch": 0.5958595859585959, "grad_norm": 3.62770676612854, "learning_rate": 6.999942918223053e-05, "loss": 4.6398, "step": 331 }, { "epoch": 0.5976597659765976, "grad_norm": 3.656203508377075, "learning_rate": 6.959311151411212e-05, "loss": 4.2742, "step": 332 }, { "epoch": 0.5994599459945995, "grad_norm": 4.29753303527832, "learning_rate": 6.918525942823835e-05, "loss": 4.6216, "step": 333 }, { "epoch": 0.6012601260126013, "grad_norm": 3.8375144004821777, "learning_rate": 6.87759048651962e-05, "loss": 4.6474, "step": 334 }, { "epoch": 0.603060306030603, "grad_norm": 3.9374406337738037, "learning_rate": 6.836507988323784e-05, "loss": 4.8221, "step": 335 }, { "epoch": 0.6048604860486049, "grad_norm": 4.202131748199463, "learning_rate": 6.795281665577003e-05, "loss": 4.4779, "step": 336 }, { "epoch": 0.6066606660666066, "grad_norm": 4.333240985870361, "learning_rate": 6.753914746883451e-05, "loss": 5.2416, "step": 337 }, { "epoch": 0.6084608460846085, "grad_norm": 4.519970417022705, "learning_rate": 6.712410471857955e-05, "loss": 5.6601, "step": 338 }, { "epoch": 0.6102610261026102, "grad_norm": 4.426661968231201, "learning_rate": 6.67077209087228e-05, "loss": 4.9072, "step": 339 }, { "epoch": 0.6120612061206121, "grad_norm": 3.4359755516052246, "learning_rate": 6.629002864800589e-05, "loss": 4.8438, "step": 340 }, { "epoch": 0.6138613861386139, "grad_norm": 3.7532505989074707, "learning_rate": 6.587106064764064e-05, "loss": 5.4166, "step": 341 }, { "epoch": 0.6156615661566157, "grad_norm": 3.9281110763549805, "learning_rate": 6.545084971874738e-05, "loss": 4.7921, "step": 342 }, { "epoch": 0.6174617461746175, "grad_norm": 3.9203011989593506, "learning_rate": 6.502942876978524e-05, "loss": 4.6832, "step": 343 }, { "epoch": 0.6192619261926192, "grad_norm": 3.9752700328826904, "learning_rate": 6.460683080397508e-05, "loss": 5.1418, "step": 344 }, { "epoch": 0.6210621062106211, "grad_norm": 3.637481451034546, "learning_rate": 6.418308891671484e-05, "loss": 4.8411, "step": 345 }, { "epoch": 0.6228622862286228, "grad_norm": 3.9811997413635254, "learning_rate": 6.375823629298765e-05, "loss": 4.9951, "step": 346 }, { "epoch": 0.6246624662466247, "grad_norm": 4.1592559814453125, "learning_rate": 6.333230620476307e-05, "loss": 5.9485, "step": 347 }, { "epoch": 0.6264626462646264, "grad_norm": 3.6163744926452637, "learning_rate": 6.29053320083913e-05, "loss": 5.3969, "step": 348 }, { "epoch": 0.6282628262826283, "grad_norm": 4.366085052490234, "learning_rate": 6.247734714199105e-05, "loss": 5.3296, "step": 349 }, { "epoch": 0.6300630063006301, "grad_norm": 4.482414722442627, "learning_rate": 6.204838512283072e-05, "loss": 6.0348, "step": 350 }, { "epoch": 0.6318631863186318, "grad_norm": 38.78593826293945, "learning_rate": 6.161847954470365e-05, "loss": 4.4123, "step": 351 }, { "epoch": 0.6336633663366337, "grad_norm": 4.5870041847229, "learning_rate": 6.118766407529713e-05, "loss": 2.4915, "step": 352 }, { "epoch": 0.6354635463546354, "grad_norm": 3.5220015048980713, "learning_rate": 6.075597245355589e-05, "loss": 2.8723, "step": 353 }, { "epoch": 0.6372637263726373, "grad_norm": 3.2156431674957275, "learning_rate": 6.03234384870397e-05, "loss": 3.0837, "step": 354 }, { "epoch": 0.639063906390639, "grad_norm": 3.223064422607422, "learning_rate": 5.989009604927587e-05, "loss": 2.9846, "step": 355 }, { "epoch": 0.6408640864086409, "grad_norm": 3.5234880447387695, "learning_rate": 5.9455979077106463e-05, "loss": 3.4383, "step": 356 }, { "epoch": 0.6426642664266426, "grad_norm": 3.648319721221924, "learning_rate": 5.9021121568030514e-05, "loss": 3.2179, "step": 357 }, { "epoch": 0.6444644464446445, "grad_norm": 3.348188877105713, "learning_rate": 5.8585557577541595e-05, "loss": 3.2722, "step": 358 }, { "epoch": 0.6462646264626463, "grad_norm": 3.6280741691589355, "learning_rate": 5.814932121646074e-05, "loss": 3.6418, "step": 359 }, { "epoch": 0.648064806480648, "grad_norm": 3.3993427753448486, "learning_rate": 5.771244664826512e-05, "loss": 3.6959, "step": 360 }, { "epoch": 0.6498649864986499, "grad_norm": 3.4230000972747803, "learning_rate": 5.727496808641256e-05, "loss": 3.5139, "step": 361 }, { "epoch": 0.6516651665166516, "grad_norm": 3.6757190227508545, "learning_rate": 5.6836919791662136e-05, "loss": 4.1188, "step": 362 }, { "epoch": 0.6534653465346535, "grad_norm": 3.232292413711548, "learning_rate": 5.639833606939102e-05, "loss": 3.7585, "step": 363 }, { "epoch": 0.6552655265526552, "grad_norm": 3.3726987838745117, "learning_rate": 5.595925126690801e-05, "loss": 3.8637, "step": 364 }, { "epoch": 0.6570657065706571, "grad_norm": 3.570510149002075, "learning_rate": 5.55196997707635e-05, "loss": 3.6172, "step": 365 }, { "epoch": 0.6588658865886589, "grad_norm": 3.3099026679992676, "learning_rate": 5.507971600405663e-05, "loss": 4.0602, "step": 366 }, { "epoch": 0.6606660666066607, "grad_norm": 3.730677843093872, "learning_rate": 5.463933442373945e-05, "loss": 4.0647, "step": 367 }, { "epoch": 0.6624662466246625, "grad_norm": 3.8183600902557373, "learning_rate": 5.419858951791842e-05, "loss": 3.9943, "step": 368 }, { "epoch": 0.6642664266426642, "grad_norm": 3.7817885875701904, "learning_rate": 5.3757515803153546e-05, "loss": 3.6044, "step": 369 }, { "epoch": 0.6660666066606661, "grad_norm": 3.696652412414551, "learning_rate": 5.3316147821755205e-05, "loss": 3.8305, "step": 370 }, { "epoch": 0.6678667866786678, "grad_norm": 3.8938591480255127, "learning_rate": 5.2874520139079006e-05, "loss": 4.1762, "step": 371 }, { "epoch": 0.6696669666966697, "grad_norm": 3.8813014030456543, "learning_rate": 5.2432667340818906e-05, "loss": 4.0786, "step": 372 }, { "epoch": 0.6714671467146714, "grad_norm": 3.655909299850464, "learning_rate": 5.1990624030298506e-05, "loss": 3.9051, "step": 373 }, { "epoch": 0.6732673267326733, "grad_norm": 3.8451969623565674, "learning_rate": 5.154842482576127e-05, "loss": 4.1839, "step": 374 }, { "epoch": 0.6750675067506751, "grad_norm": 4.180874824523926, "learning_rate": 5.110610435765934e-05, "loss": 4.0388, "step": 375 }, { "epoch": 0.6768676867686768, "grad_norm": 3.5825304985046387, "learning_rate": 5.0663697265941545e-05, "loss": 4.1734, "step": 376 }, { "epoch": 0.6786678667866787, "grad_norm": 3.780687093734741, "learning_rate": 5.022123819734053e-05, "loss": 4.3508, "step": 377 }, { "epoch": 0.6804680468046804, "grad_norm": 3.889195203781128, "learning_rate": 4.977876180265948e-05, "loss": 4.1945, "step": 378 }, { "epoch": 0.6822682268226823, "grad_norm": 3.726132869720459, "learning_rate": 4.933630273405847e-05, "loss": 3.8745, "step": 379 }, { "epoch": 0.684068406840684, "grad_norm": 3.7415261268615723, "learning_rate": 4.889389564234066e-05, "loss": 4.2891, "step": 380 }, { "epoch": 0.6858685868586859, "grad_norm": 3.6238245964050293, "learning_rate": 4.845157517423874e-05, "loss": 4.3567, "step": 381 }, { "epoch": 0.6876687668766877, "grad_norm": 3.964704990386963, "learning_rate": 4.8009375969701505e-05, "loss": 4.5154, "step": 382 }, { "epoch": 0.6894689468946895, "grad_norm": 3.337338447570801, "learning_rate": 4.7567332659181106e-05, "loss": 4.5835, "step": 383 }, { "epoch": 0.6912691269126913, "grad_norm": 3.447648286819458, "learning_rate": 4.7125479860921e-05, "loss": 4.2734, "step": 384 }, { "epoch": 0.693069306930693, "grad_norm": 4.408335208892822, "learning_rate": 4.668385217824482e-05, "loss": 4.8601, "step": 385 }, { "epoch": 0.6948694869486949, "grad_norm": 3.683608293533325, "learning_rate": 4.6242484196846465e-05, "loss": 4.9472, "step": 386 }, { "epoch": 0.6966696669666966, "grad_norm": 4.226807117462158, "learning_rate": 4.5801410482081584e-05, "loss": 4.6788, "step": 387 }, { "epoch": 0.6984698469846985, "grad_norm": 3.806995391845703, "learning_rate": 4.536066557626056e-05, "loss": 4.9015, "step": 388 }, { "epoch": 0.7002700270027002, "grad_norm": 3.541186571121216, "learning_rate": 4.4920283995943383e-05, "loss": 5.1853, "step": 389 }, { "epoch": 0.7020702070207021, "grad_norm": 3.971658706665039, "learning_rate": 4.448030022923652e-05, "loss": 4.8798, "step": 390 }, { "epoch": 0.7038703870387039, "grad_norm": 3.6937930583953857, "learning_rate": 4.404074873309201e-05, "loss": 4.9251, "step": 391 }, { "epoch": 0.7056705670567057, "grad_norm": 4.2643723487854, "learning_rate": 4.3601663930608995e-05, "loss": 4.8249, "step": 392 }, { "epoch": 0.7074707470747075, "grad_norm": 3.8684353828430176, "learning_rate": 4.316308020833788e-05, "loss": 5.6238, "step": 393 }, { "epoch": 0.7092709270927092, "grad_norm": 3.8425145149230957, "learning_rate": 4.272503191358743e-05, "loss": 4.9036, "step": 394 }, { "epoch": 0.7110711071107111, "grad_norm": 3.797442674636841, "learning_rate": 4.228755335173488e-05, "loss": 5.0452, "step": 395 }, { "epoch": 0.7128712871287128, "grad_norm": 4.2270355224609375, "learning_rate": 4.1850678783539265e-05, "loss": 4.7575, "step": 396 }, { "epoch": 0.7146714671467147, "grad_norm": 4.077787399291992, "learning_rate": 4.141444242245841e-05, "loss": 5.9227, "step": 397 }, { "epoch": 0.7164716471647165, "grad_norm": 3.7013025283813477, "learning_rate": 4.0978878431969484e-05, "loss": 4.7173, "step": 398 }, { "epoch": 0.7182718271827183, "grad_norm": 4.04628849029541, "learning_rate": 4.054402092289354e-05, "loss": 5.4247, "step": 399 }, { "epoch": 0.7200720072007201, "grad_norm": 4.189887523651123, "learning_rate": 4.0109903950724134e-05, "loss": 5.3008, "step": 400 }, { "epoch": 0.7218721872187218, "grad_norm": 20.025436401367188, "learning_rate": 3.967656151296031e-05, "loss": 3.8243, "step": 401 }, { "epoch": 0.7236723672367237, "grad_norm": 4.733638286590576, "learning_rate": 3.924402754644412e-05, "loss": 2.4717, "step": 402 }, { "epoch": 0.7254725472547254, "grad_norm": 3.5588278770446777, "learning_rate": 3.881233592470287e-05, "loss": 2.7574, "step": 403 }, { "epoch": 0.7272727272727273, "grad_norm": 3.0511527061462402, "learning_rate": 3.8381520455296364e-05, "loss": 2.6524, "step": 404 }, { "epoch": 0.729072907290729, "grad_norm": 3.7463011741638184, "learning_rate": 3.7951614877169284e-05, "loss": 3.2022, "step": 405 }, { "epoch": 0.7308730873087309, "grad_norm": 3.974733829498291, "learning_rate": 3.7522652858008964e-05, "loss": 3.7213, "step": 406 }, { "epoch": 0.7326732673267327, "grad_norm": 3.5178704261779785, "learning_rate": 3.7094667991608704e-05, "loss": 3.7185, "step": 407 }, { "epoch": 0.7344734473447345, "grad_norm": 3.9185893535614014, "learning_rate": 3.6667693795236946e-05, "loss": 4.0272, "step": 408 }, { "epoch": 0.7362736273627363, "grad_norm": 3.0775508880615234, "learning_rate": 3.624176370701236e-05, "loss": 3.0848, "step": 409 }, { "epoch": 0.738073807380738, "grad_norm": 3.445986270904541, "learning_rate": 3.581691108328517e-05, "loss": 3.857, "step": 410 }, { "epoch": 0.7398739873987399, "grad_norm": 3.1240127086639404, "learning_rate": 3.5393169196024925e-05, "loss": 3.0593, "step": 411 }, { "epoch": 0.7416741674167416, "grad_norm": 3.4222195148468018, "learning_rate": 3.4970571230214774e-05, "loss": 3.551, "step": 412 }, { "epoch": 0.7434743474347435, "grad_norm": 3.3344242572784424, "learning_rate": 3.4549150281252636e-05, "loss": 3.6521, "step": 413 }, { "epoch": 0.7452745274527453, "grad_norm": 3.6145176887512207, "learning_rate": 3.412893935235936e-05, "loss": 3.7669, "step": 414 }, { "epoch": 0.7470747074707471, "grad_norm": 3.2903146743774414, "learning_rate": 3.370997135199413e-05, "loss": 3.333, "step": 415 }, { "epoch": 0.7488748874887489, "grad_norm": 3.5195930004119873, "learning_rate": 3.329227909127722e-05, "loss": 4.2719, "step": 416 }, { "epoch": 0.7506750675067507, "grad_norm": 3.5990970134735107, "learning_rate": 3.287589528142048e-05, "loss": 4.1684, "step": 417 }, { "epoch": 0.7524752475247525, "grad_norm": 3.387673854827881, "learning_rate": 3.24608525311655e-05, "loss": 3.8335, "step": 418 }, { "epoch": 0.7542754275427542, "grad_norm": 3.6439712047576904, "learning_rate": 3.204718334422998e-05, "loss": 4.3415, "step": 419 }, { "epoch": 0.7560756075607561, "grad_norm": 3.3546173572540283, "learning_rate": 3.1634920116762176e-05, "loss": 3.8877, "step": 420 }, { "epoch": 0.7578757875787578, "grad_norm": 3.6753296852111816, "learning_rate": 3.122409513480381e-05, "loss": 4.1525, "step": 421 }, { "epoch": 0.7596759675967597, "grad_norm": 3.644063711166382, "learning_rate": 3.081474057176164e-05, "loss": 3.8885, "step": 422 }, { "epoch": 0.7614761476147615, "grad_norm": 4.089780807495117, "learning_rate": 3.040688848588788e-05, "loss": 4.0093, "step": 423 }, { "epoch": 0.7632763276327633, "grad_norm": 3.1043710708618164, "learning_rate": 3.000057081776947e-05, "loss": 4.0178, "step": 424 }, { "epoch": 0.7650765076507651, "grad_norm": 3.322037935256958, "learning_rate": 2.959581938782675e-05, "loss": 4.1183, "step": 425 }, { "epoch": 0.7668766876687669, "grad_norm": 3.9151599407196045, "learning_rate": 2.9192665893821475e-05, "loss": 4.2036, "step": 426 }, { "epoch": 0.7686768676867687, "grad_norm": 3.6815481185913086, "learning_rate": 2.8791141908374397e-05, "loss": 4.7545, "step": 427 }, { "epoch": 0.7704770477047704, "grad_norm": 4.053499698638916, "learning_rate": 2.839127887649271e-05, "loss": 4.0245, "step": 428 }, { "epoch": 0.7722772277227723, "grad_norm": 3.4955430030822754, "learning_rate": 2.7993108113107447e-05, "loss": 3.901, "step": 429 }, { "epoch": 0.774077407740774, "grad_norm": 3.2984628677368164, "learning_rate": 2.7596660800621078e-05, "loss": 3.9142, "step": 430 }, { "epoch": 0.7758775877587759, "grad_norm": 3.7901782989501953, "learning_rate": 2.7201967986465515e-05, "loss": 3.669, "step": 431 }, { "epoch": 0.7776777677767777, "grad_norm": 3.3658523559570312, "learning_rate": 2.6809060580670607e-05, "loss": 4.3362, "step": 432 }, { "epoch": 0.7794779477947795, "grad_norm": 3.361583948135376, "learning_rate": 2.6417969353443483e-05, "loss": 4.4502, "step": 433 }, { "epoch": 0.7812781278127813, "grad_norm": 3.271467924118042, "learning_rate": 2.6028724932758813e-05, "loss": 4.1924, "step": 434 }, { "epoch": 0.783078307830783, "grad_norm": 3.368018627166748, "learning_rate": 2.5641357801960187e-05, "loss": 4.2437, "step": 435 }, { "epoch": 0.7848784878487849, "grad_norm": 4.052670955657959, "learning_rate": 2.525589829737284e-05, "loss": 4.7636, "step": 436 }, { "epoch": 0.7866786678667866, "grad_norm": 3.8365402221679688, "learning_rate": 2.487237660592789e-05, "loss": 4.259, "step": 437 }, { "epoch": 0.7884788478847885, "grad_norm": 3.631394386291504, "learning_rate": 2.44908227627983e-05, "loss": 4.7527, "step": 438 }, { "epoch": 0.7902790279027903, "grad_norm": 3.6414878368377686, "learning_rate": 2.4111266649046647e-05, "loss": 4.5161, "step": 439 }, { "epoch": 0.7920792079207921, "grad_norm": 3.9235427379608154, "learning_rate": 2.373373798928507e-05, "loss": 4.7452, "step": 440 }, { "epoch": 0.7938793879387939, "grad_norm": 3.502267599105835, "learning_rate": 2.335826634934737e-05, "loss": 4.2153, "step": 441 }, { "epoch": 0.7956795679567957, "grad_norm": 3.719147205352783, "learning_rate": 2.298488113397359e-05, "loss": 4.7187, "step": 442 }, { "epoch": 0.7974797479747975, "grad_norm": 4.110799789428711, "learning_rate": 2.2613611584507228e-05, "loss": 4.898, "step": 443 }, { "epoch": 0.7992799279927992, "grad_norm": 3.9746625423431396, "learning_rate": 2.2244486776605243e-05, "loss": 5.2031, "step": 444 }, { "epoch": 0.8010801080108011, "grad_norm": 3.9334423542022705, "learning_rate": 2.187753561796097e-05, "loss": 5.5832, "step": 445 }, { "epoch": 0.8028802880288028, "grad_norm": 3.596407413482666, "learning_rate": 2.1512786846040255e-05, "loss": 5.0518, "step": 446 }, { "epoch": 0.8046804680468047, "grad_norm": 3.7824950218200684, "learning_rate": 2.1150269025830925e-05, "loss": 5.0857, "step": 447 }, { "epoch": 0.8064806480648065, "grad_norm": 3.96897292137146, "learning_rate": 2.0790010547605743e-05, "loss": 5.4674, "step": 448 }, { "epoch": 0.8082808280828083, "grad_norm": 3.868983030319214, "learning_rate": 2.0432039624699016e-05, "loss": 5.3334, "step": 449 }, { "epoch": 0.8100810081008101, "grad_norm": 4.298758029937744, "learning_rate": 2.0076384291297134e-05, "loss": 5.33, "step": 450 }, { "epoch": 0.8118811881188119, "grad_norm": 14.23192024230957, "learning_rate": 1.972307240024307e-05, "loss": 3.0892, "step": 451 }, { "epoch": 0.8136813681368137, "grad_norm": 4.406993865966797, "learning_rate": 1.9372131620855095e-05, "loss": 2.5208, "step": 452 }, { "epoch": 0.8154815481548154, "grad_norm": 5.394495487213135, "learning_rate": 1.9023589436759954e-05, "loss": 2.8551, "step": 453 }, { "epoch": 0.8172817281728173, "grad_norm": 3.0106537342071533, "learning_rate": 1.867747314374041e-05, "loss": 3.0363, "step": 454 }, { "epoch": 0.819081908190819, "grad_norm": 2.969430685043335, "learning_rate": 1.8333809847597642e-05, "loss": 2.9957, "step": 455 }, { "epoch": 0.8208820882088209, "grad_norm": 3.101269245147705, "learning_rate": 1.7992626462028513e-05, "loss": 3.0854, "step": 456 }, { "epoch": 0.8226822682268227, "grad_norm": 3.125643730163574, "learning_rate": 1.765394970651777e-05, "loss": 3.2697, "step": 457 }, { "epoch": 0.8244824482448245, "grad_norm": 3.0227460861206055, "learning_rate": 1.7317806104245597e-05, "loss": 3.4779, "step": 458 }, { "epoch": 0.8262826282628263, "grad_norm": 3.203869342803955, "learning_rate": 1.6984221980010455e-05, "loss": 3.7277, "step": 459 }, { "epoch": 0.828082808280828, "grad_norm": 3.483610153198242, "learning_rate": 1.665322345816746e-05, "loss": 3.8941, "step": 460 }, { "epoch": 0.8298829882988299, "grad_norm": 3.1454741954803467, "learning_rate": 1.6324836460582498e-05, "loss": 3.883, "step": 461 }, { "epoch": 0.8316831683168316, "grad_norm": 3.6109461784362793, "learning_rate": 1.599908670460216e-05, "loss": 3.2933, "step": 462 }, { "epoch": 0.8334833483348335, "grad_norm": 3.4250762462615967, "learning_rate": 1.5675999701039735e-05, "loss": 3.628, "step": 463 }, { "epoch": 0.8352835283528353, "grad_norm": 3.624478578567505, "learning_rate": 1.535560075217731e-05, "loss": 4.2137, "step": 464 }, { "epoch": 0.8370837083708371, "grad_norm": 3.289508819580078, "learning_rate": 1.5037914949784299e-05, "loss": 3.5101, "step": 465 }, { "epoch": 0.8388838883888389, "grad_norm": 3.1444578170776367, "learning_rate": 1.4722967173152353e-05, "loss": 4.0805, "step": 466 }, { "epoch": 0.8406840684068407, "grad_norm": 2.974825143814087, "learning_rate": 1.441078208714698e-05, "loss": 3.5221, "step": 467 }, { "epoch": 0.8424842484248425, "grad_norm": 3.4931812286376953, "learning_rate": 1.4101384140275947e-05, "loss": 3.6931, "step": 468 }, { "epoch": 0.8442844284428442, "grad_norm": 3.261122703552246, "learning_rate": 1.3794797562774598e-05, "loss": 3.769, "step": 469 }, { "epoch": 0.8460846084608461, "grad_norm": 3.2930214405059814, "learning_rate": 1.3491046364708293e-05, "loss": 3.6359, "step": 470 }, { "epoch": 0.8478847884788479, "grad_norm": 3.166865587234497, "learning_rate": 1.319015433409208e-05, "loss": 4.2264, "step": 471 }, { "epoch": 0.8496849684968497, "grad_norm": 3.187593936920166, "learning_rate": 1.2892145035027758e-05, "loss": 4.2423, "step": 472 }, { "epoch": 0.8514851485148515, "grad_norm": 2.961714506149292, "learning_rate": 1.2597041805858468e-05, "loss": 3.7844, "step": 473 }, { "epoch": 0.8532853285328533, "grad_norm": 3.1579463481903076, "learning_rate": 1.2304867757340972e-05, "loss": 3.9704, "step": 474 }, { "epoch": 0.8550855085508551, "grad_norm": 3.4631705284118652, "learning_rate": 1.2015645770835764e-05, "loss": 3.943, "step": 475 }, { "epoch": 0.8568856885688569, "grad_norm": 2.9861361980438232, "learning_rate": 1.1729398496515126e-05, "loss": 3.6184, "step": 476 }, { "epoch": 0.8586858685868587, "grad_norm": 3.6547658443450928, "learning_rate": 1.1446148351589297e-05, "loss": 4.8532, "step": 477 }, { "epoch": 0.8604860486048604, "grad_norm": 3.7734615802764893, "learning_rate": 1.1165917518550911e-05, "loss": 4.3348, "step": 478 }, { "epoch": 0.8622862286228623, "grad_norm": 3.309903621673584, "learning_rate": 1.0888727943437748e-05, "loss": 4.2671, "step": 479 }, { "epoch": 0.8640864086408641, "grad_norm": 3.503133535385132, "learning_rate": 1.0614601334114099e-05, "loss": 4.4343, "step": 480 }, { "epoch": 0.8658865886588659, "grad_norm": 3.389491558074951, "learning_rate": 1.0343559158570704e-05, "loss": 4.7319, "step": 481 }, { "epoch": 0.8676867686768677, "grad_norm": 3.3840572834014893, "learning_rate": 1.0075622643243499e-05, "loss": 4.5601, "step": 482 }, { "epoch": 0.8694869486948695, "grad_norm": 3.25216007232666, "learning_rate": 9.810812771351335e-06, "loss": 4.2966, "step": 483 }, { "epoch": 0.8712871287128713, "grad_norm": 3.635775089263916, "learning_rate": 9.549150281252633e-06, "loss": 4.7205, "step": 484 }, { "epoch": 0.873087308730873, "grad_norm": 3.3356218338012695, "learning_rate": 9.290655664821296e-06, "loss": 4.9521, "step": 485 }, { "epoch": 0.8748874887488749, "grad_norm": 3.6561996936798096, "learning_rate": 9.035349165841923e-06, "loss": 4.7275, "step": 486 }, { "epoch": 0.8766876687668766, "grad_norm": 3.564394235610962, "learning_rate": 8.783250778424428e-06, "loss": 4.1435, "step": 487 }, { "epoch": 0.8784878487848785, "grad_norm": 3.7181074619293213, "learning_rate": 8.534380245438211e-06, "loss": 5.1263, "step": 488 }, { "epoch": 0.8802880288028803, "grad_norm": 3.6209585666656494, "learning_rate": 8.288757056966018e-06, "loss": 5.0345, "step": 489 }, { "epoch": 0.8820882088208821, "grad_norm": 4.174393653869629, "learning_rate": 8.046400448777574e-06, "loss": 5.2958, "step": 490 }, { "epoch": 0.8838883888388839, "grad_norm": 3.6721696853637695, "learning_rate": 7.807329400823177e-06, "loss": 5.1729, "step": 491 }, { "epoch": 0.8856885688568857, "grad_norm": 3.6310222148895264, "learning_rate": 7.571562635747304e-06, "loss": 4.2984, "step": 492 }, { "epoch": 0.8874887488748875, "grad_norm": 3.5798494815826416, "learning_rate": 7.3391186174223245e-06, "loss": 4.7273, "step": 493 }, { "epoch": 0.8892889288928892, "grad_norm": 3.331014633178711, "learning_rate": 7.1100155495025464e-06, "loss": 5.1268, "step": 494 }, { "epoch": 0.8910891089108911, "grad_norm": 3.6740384101867676, "learning_rate": 6.8842713739986075e-06, "loss": 5.3065, "step": 495 }, { "epoch": 0.8928892889288929, "grad_norm": 3.8056387901306152, "learning_rate": 6.661903769872363e-06, "loss": 5.405, "step": 496 }, { "epoch": 0.8946894689468947, "grad_norm": 3.7739763259887695, "learning_rate": 6.442930151652371e-06, "loss": 5.9443, "step": 497 }, { "epoch": 0.8964896489648965, "grad_norm": 3.922481060028076, "learning_rate": 6.227367668070083e-06, "loss": 5.1865, "step": 498 }, { "epoch": 0.8982898289828983, "grad_norm": 3.815195083618164, "learning_rate": 6.015233200716874e-06, "loss": 5.4361, "step": 499 }, { "epoch": 0.9000900090009001, "grad_norm": 4.479306697845459, "learning_rate": 5.806543362721945e-06, "loss": 6.0537, "step": 500 }, { "epoch": 0.9018901890189019, "grad_norm": 25.004121780395508, "learning_rate": 5.601314497451316e-06, "loss": 3.6237, "step": 501 }, { "epoch": 0.9036903690369037, "grad_norm": 3.7974557876586914, "learning_rate": 5.399562677227876e-06, "loss": 2.1349, "step": 502 }, { "epoch": 0.9054905490549054, "grad_norm": 3.4542431831359863, "learning_rate": 5.201303702072724e-06, "loss": 2.8379, "step": 503 }, { "epoch": 0.9072907290729073, "grad_norm": 3.14579176902771, "learning_rate": 5.006553098467764e-06, "loss": 2.9554, "step": 504 }, { "epoch": 0.9090909090909091, "grad_norm": 2.649793863296509, "learning_rate": 4.8153261181398125e-06, "loss": 2.2224, "step": 505 }, { "epoch": 0.9108910891089109, "grad_norm": 3.07043719291687, "learning_rate": 4.627637736866119e-06, "loss": 3.5249, "step": 506 }, { "epoch": 0.9126912691269127, "grad_norm": 3.646077871322632, "learning_rate": 4.443502653301601e-06, "loss": 3.5183, "step": 507 }, { "epoch": 0.9144914491449145, "grad_norm": 3.1195287704467773, "learning_rate": 4.262935287827696e-06, "loss": 3.507, "step": 508 }, { "epoch": 0.9162916291629163, "grad_norm": 3.2357397079467773, "learning_rate": 4.085949781423048e-06, "loss": 3.3765, "step": 509 }, { "epoch": 0.918091809180918, "grad_norm": 3.1325178146362305, "learning_rate": 3.912559994556086e-06, "loss": 3.5724, "step": 510 }, { "epoch": 0.9198919891989199, "grad_norm": 3.3491556644439697, "learning_rate": 3.7427795060995306e-06, "loss": 3.8159, "step": 511 }, { "epoch": 0.9216921692169217, "grad_norm": 3.5712788105010986, "learning_rate": 3.5766216122670003e-06, "loss": 3.6621, "step": 512 }, { "epoch": 0.9234923492349235, "grad_norm": 3.1630241870880127, "learning_rate": 3.4140993255717126e-06, "loss": 3.6589, "step": 513 }, { "epoch": 0.9252925292529253, "grad_norm": 3.174769401550293, "learning_rate": 3.2552253738074077e-06, "loss": 3.8955, "step": 514 }, { "epoch": 0.9270927092709271, "grad_norm": 3.325157880783081, "learning_rate": 3.1000121990516274e-06, "loss": 3.8899, "step": 515 }, { "epoch": 0.9288928892889289, "grad_norm": 3.770094394683838, "learning_rate": 2.9484719566912735e-06, "loss": 4.0021, "step": 516 }, { "epoch": 0.9306930693069307, "grad_norm": 3.7571139335632324, "learning_rate": 2.800616514470694e-06, "loss": 4.1601, "step": 517 }, { "epoch": 0.9324932493249325, "grad_norm": 3.390287160873413, "learning_rate": 2.656457451562283e-06, "loss": 4.0102, "step": 518 }, { "epoch": 0.9342934293429342, "grad_norm": 3.436528444290161, "learning_rate": 2.5160060576596366e-06, "loss": 4.2173, "step": 519 }, { "epoch": 0.9360936093609361, "grad_norm": 2.9943151473999023, "learning_rate": 2.3792733320934346e-06, "loss": 3.7667, "step": 520 }, { "epoch": 0.9378937893789379, "grad_norm": 3.1079506874084473, "learning_rate": 2.2462699829700286e-06, "loss": 3.9543, "step": 521 }, { "epoch": 0.9396939693969397, "grad_norm": 3.511355400085449, "learning_rate": 2.1170064263328514e-06, "loss": 4.3681, "step": 522 }, { "epoch": 0.9414941494149415, "grad_norm": 3.2926292419433594, "learning_rate": 1.991492785346677e-06, "loss": 4.2362, "step": 523 }, { "epoch": 0.9432943294329433, "grad_norm": 3.607163667678833, "learning_rate": 1.8697388895048518e-06, "loss": 4.1022, "step": 524 }, { "epoch": 0.9450945094509451, "grad_norm": 3.624814510345459, "learning_rate": 1.7517542738595071e-06, "loss": 3.7475, "step": 525 }, { "epoch": 0.946894689468947, "grad_norm": 3.673523426055908, "learning_rate": 1.637548178274806e-06, "loss": 3.8728, "step": 526 }, { "epoch": 0.9486948694869487, "grad_norm": 3.480508804321289, "learning_rate": 1.5271295467033653e-06, "loss": 4.1873, "step": 527 }, { "epoch": 0.9504950495049505, "grad_norm": 3.1145169734954834, "learning_rate": 1.42050702648579e-06, "loss": 4.2106, "step": 528 }, { "epoch": 0.9522952295229523, "grad_norm": 3.012479782104492, "learning_rate": 1.3176889676734693e-06, "loss": 4.4033, "step": 529 }, { "epoch": 0.9540954095409541, "grad_norm": 3.321237564086914, "learning_rate": 1.2186834223746612e-06, "loss": 3.8453, "step": 530 }, { "epoch": 0.9558955895589559, "grad_norm": 3.554750680923462, "learning_rate": 1.1234981441239022e-06, "loss": 4.2771, "step": 531 }, { "epoch": 0.9576957695769577, "grad_norm": 3.595700979232788, "learning_rate": 1.0321405872747603e-06, "loss": 4.7381, "step": 532 }, { "epoch": 0.9594959495949595, "grad_norm": 3.4314756393432617, "learning_rate": 9.446179064161009e-07, "loss": 4.7025, "step": 533 }, { "epoch": 0.9612961296129613, "grad_norm": 3.4042766094207764, "learning_rate": 8.609369558117531e-07, "loss": 4.5694, "step": 534 }, { "epoch": 0.963096309630963, "grad_norm": 3.677279472351074, "learning_rate": 7.81104288863721e-07, "loss": 4.4108, "step": 535 }, { "epoch": 0.9648964896489649, "grad_norm": 4.0047712326049805, "learning_rate": 7.051261575989787e-07, "loss": 5.2683, "step": 536 }, { "epoch": 0.9666966696669667, "grad_norm": 3.8668036460876465, "learning_rate": 6.330085121798279e-07, "loss": 5.0821, "step": 537 }, { "epoch": 0.9684968496849685, "grad_norm": 3.524517297744751, "learning_rate": 5.647570004379432e-07, "loss": 4.7731, "step": 538 }, { "epoch": 0.9702970297029703, "grad_norm": 4.2224955558776855, "learning_rate": 5.003769674320369e-07, "loss": 5.4911, "step": 539 }, { "epoch": 0.9720972097209721, "grad_norm": 3.456015110015869, "learning_rate": 4.3987345502927157e-07, "loss": 4.9279, "step": 540 }, { "epoch": 0.9738973897389739, "grad_norm": 4.04163122177124, "learning_rate": 3.832512015104317e-07, "loss": 4.8574, "step": 541 }, { "epoch": 0.9756975697569757, "grad_norm": 3.488652467727661, "learning_rate": 3.3051464119883136e-07, "loss": 4.5105, "step": 542 }, { "epoch": 0.9774977497749775, "grad_norm": 3.6072845458984375, "learning_rate": 2.816679041130477e-07, "loss": 5.0124, "step": 543 }, { "epoch": 0.9792979297929792, "grad_norm": 3.613863229751587, "learning_rate": 2.3671481564347397e-07, "loss": 4.7951, "step": 544 }, { "epoch": 0.9810981098109811, "grad_norm": 3.676605701446533, "learning_rate": 1.9565889625275946e-07, "loss": 5.7295, "step": 545 }, { "epoch": 0.9828982898289829, "grad_norm": 3.8148982524871826, "learning_rate": 1.5850336120007414e-07, "loss": 5.0033, "step": 546 }, { "epoch": 0.9846984698469847, "grad_norm": 3.664438486099243, "learning_rate": 1.252511202893436e-07, "loss": 4.9703, "step": 547 }, { "epoch": 0.9864986498649865, "grad_norm": 3.756563663482666, "learning_rate": 9.590477764135353e-08, "loss": 5.3585, "step": 548 }, { "epoch": 0.9882988298829883, "grad_norm": 3.9945781230926514, "learning_rate": 7.046663148979616e-08, "loss": 5.4977, "step": 549 }, { "epoch": 0.9900990099009901, "grad_norm": 4.365446090698242, "learning_rate": 4.89386740013198e-08, "loss": 5.6052, "step": 550 }, { "epoch": 0.991899189918992, "grad_norm": 3.4431326389312744, "learning_rate": 3.1322591119481436e-08, "loss": 2.9021, "step": 551 }, { "epoch": 0.9936993699369937, "grad_norm": 3.6631813049316406, "learning_rate": 1.7619762432730113e-08, "loss": 4.0126, "step": 552 }, { "epoch": 0.9954995499549955, "grad_norm": 3.126596212387085, "learning_rate": 7.831261066371109e-09, "loss": 3.7145, "step": 553 }, { "epoch": 0.9972997299729973, "grad_norm": 3.9898624420166016, "learning_rate": 1.9578535984998705e-09, "loss": 4.723, "step": 554 }, { "epoch": 0.9990999099909991, "grad_norm": 3.7214019298553467, "learning_rate": 0.0, "loss": 5.0888, "step": 555 } ], "logging_steps": 1, "max_steps": 555, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 239, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.5203413975262822e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }