|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.06291899179806, |
|
"eval_steps": 42, |
|
"global_step": 420, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00014980712332871428, |
|
"eval_loss": 1.160882592201233, |
|
"eval_runtime": 311.16, |
|
"eval_samples_per_second": 36.133, |
|
"eval_steps_per_second": 4.519, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00044942136998614283, |
|
"grad_norm": 0.12129199504852295, |
|
"learning_rate": 1.5e-05, |
|
"loss": 1.1193, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0008988427399722857, |
|
"grad_norm": 0.11971050500869751, |
|
"learning_rate": 3e-05, |
|
"loss": 1.103, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0013482641099584285, |
|
"grad_norm": 0.10862895101308823, |
|
"learning_rate": 4.5e-05, |
|
"loss": 1.0659, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0017976854799445713, |
|
"grad_norm": 0.11977853626012802, |
|
"learning_rate": 4.9997944716957985e-05, |
|
"loss": 1.0914, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0022471068499307144, |
|
"grad_norm": 0.12554779648780823, |
|
"learning_rate": 4.99871554050172e-05, |
|
"loss": 1.1725, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.002696528219916857, |
|
"grad_norm": 0.13287784159183502, |
|
"learning_rate": 4.996712222958461e-05, |
|
"loss": 1.1145, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.003145949589903, |
|
"grad_norm": 0.1167389452457428, |
|
"learning_rate": 4.993785260182552e-05, |
|
"loss": 1.0956, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0035953709598891426, |
|
"grad_norm": 0.11548212915658951, |
|
"learning_rate": 4.989935734988098e-05, |
|
"loss": 1.12, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.004044792329875285, |
|
"grad_norm": 0.12251826375722885, |
|
"learning_rate": 4.9851650714862006e-05, |
|
"loss": 1.0895, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.004494213699861429, |
|
"grad_norm": 0.12304849177598953, |
|
"learning_rate": 4.979475034558115e-05, |
|
"loss": 1.1302, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.004943635069847571, |
|
"grad_norm": 0.12594439089298248, |
|
"learning_rate": 4.9728677292023405e-05, |
|
"loss": 1.0825, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.005393056439833714, |
|
"grad_norm": 0.11802031844854355, |
|
"learning_rate": 4.965345599755887e-05, |
|
"loss": 1.0949, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.005842477809819857, |
|
"grad_norm": 0.12052260339260101, |
|
"learning_rate": 4.95691142899001e-05, |
|
"loss": 1.2106, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.006291899179806, |
|
"grad_norm": 0.14151117205619812, |
|
"learning_rate": 4.9475683370807326e-05, |
|
"loss": 1.0855, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.006291899179806, |
|
"eval_loss": 1.1297273635864258, |
|
"eval_runtime": 313.2801, |
|
"eval_samples_per_second": 35.888, |
|
"eval_steps_per_second": 4.488, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.006741320549792143, |
|
"grad_norm": 0.13759616017341614, |
|
"learning_rate": 4.937319780454559e-05, |
|
"loss": 1.1133, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.007190741919778285, |
|
"grad_norm": 0.14887671172618866, |
|
"learning_rate": 4.926169550509787e-05, |
|
"loss": 1.1339, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.007640163289764428, |
|
"grad_norm": 0.13235358893871307, |
|
"learning_rate": 4.914121772213898e-05, |
|
"loss": 1.0057, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.00808958465975057, |
|
"grad_norm": 0.14951473474502563, |
|
"learning_rate": 4.9011809025775486e-05, |
|
"loss": 1.0919, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.008539006029736713, |
|
"grad_norm": 0.1409119963645935, |
|
"learning_rate": 4.887351729005726e-05, |
|
"loss": 1.143, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.008988427399722857, |
|
"grad_norm": 0.13198496401309967, |
|
"learning_rate": 4.8726393675266716e-05, |
|
"loss": 1.0853, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.009437848769709, |
|
"grad_norm": 0.15319645404815674, |
|
"learning_rate": 4.8570492608992325e-05, |
|
"loss": 1.0534, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.009887270139695143, |
|
"grad_norm": 0.1532219797372818, |
|
"learning_rate": 4.8405871765993433e-05, |
|
"loss": 1.1016, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.010336691509681285, |
|
"grad_norm": 0.14198264479637146, |
|
"learning_rate": 4.82325920468638e-05, |
|
"loss": 1.029, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.010786112879667428, |
|
"grad_norm": 0.12415461987257004, |
|
"learning_rate": 4.805071755550177e-05, |
|
"loss": 1.1565, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.01123553424965357, |
|
"grad_norm": 0.1350441873073578, |
|
"learning_rate": 4.7860315575395316e-05, |
|
"loss": 1.1276, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.011684955619639713, |
|
"grad_norm": 0.12075574696063995, |
|
"learning_rate": 4.766145654473095e-05, |
|
"loss": 1.0753, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.012134376989625857, |
|
"grad_norm": 0.11550460755825043, |
|
"learning_rate": 4.745421403033548e-05, |
|
"loss": 1.0078, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.012583798359612, |
|
"grad_norm": 0.12455905228853226, |
|
"learning_rate": 4.72386647004603e-05, |
|
"loss": 1.0373, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.012583798359612, |
|
"eval_loss": 1.1157194375991821, |
|
"eval_runtime": 312.8234, |
|
"eval_samples_per_second": 35.94, |
|
"eval_steps_per_second": 4.495, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.013033219729598143, |
|
"grad_norm": 0.12993893027305603, |
|
"learning_rate": 4.701488829641845e-05, |
|
"loss": 1.1313, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.013482641099584285, |
|
"grad_norm": 0.18705669045448303, |
|
"learning_rate": 4.678296760308474e-05, |
|
"loss": 1.2752, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.013932062469570428, |
|
"grad_norm": 0.14135660231113434, |
|
"learning_rate": 4.6542988418269876e-05, |
|
"loss": 1.1628, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.01438148383955657, |
|
"grad_norm": 0.11707276850938797, |
|
"learning_rate": 4.629503952098011e-05, |
|
"loss": 1.0801, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.014830905209542713, |
|
"grad_norm": 0.14167702198028564, |
|
"learning_rate": 4.6039212638573833e-05, |
|
"loss": 1.0452, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.015280326579528856, |
|
"grad_norm": 0.15983764827251434, |
|
"learning_rate": 4.5775602412827604e-05, |
|
"loss": 1.1844, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.015729747949515, |
|
"grad_norm": 0.12750358879566193, |
|
"learning_rate": 4.55043063649239e-05, |
|
"loss": 1.0877, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.01617916931950114, |
|
"grad_norm": 0.1487520933151245, |
|
"learning_rate": 4.522542485937369e-05, |
|
"loss": 1.1373, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.016628590689487285, |
|
"grad_norm": 0.16089919209480286, |
|
"learning_rate": 4.493906106688712e-05, |
|
"loss": 1.1806, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.017078012059473426, |
|
"grad_norm": 0.158865287899971, |
|
"learning_rate": 4.4645320926206064e-05, |
|
"loss": 1.0689, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.01752743342945957, |
|
"grad_norm": 0.16464003920555115, |
|
"learning_rate": 4.434431310491267e-05, |
|
"loss": 1.1319, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.017976854799445715, |
|
"grad_norm": 0.12468679249286652, |
|
"learning_rate": 4.4036148959228365e-05, |
|
"loss": 1.0237, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.018426276169431856, |
|
"grad_norm": 0.14475342631340027, |
|
"learning_rate": 4.372094249281821e-05, |
|
"loss": 1.1352, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.018875697539418, |
|
"grad_norm": 0.1386341154575348, |
|
"learning_rate": 4.3398810314615876e-05, |
|
"loss": 1.177, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.018875697539418, |
|
"eval_loss": 1.1068843603134155, |
|
"eval_runtime": 312.5635, |
|
"eval_samples_per_second": 35.97, |
|
"eval_steps_per_second": 4.498, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.01932511890940414, |
|
"grad_norm": 0.13838274776935577, |
|
"learning_rate": 4.306987159568479e-05, |
|
"loss": 0.9996, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.019774540279390285, |
|
"grad_norm": 0.14386090636253357, |
|
"learning_rate": 4.273424802513145e-05, |
|
"loss": 0.9947, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.020223961649376426, |
|
"grad_norm": 0.14473505318164825, |
|
"learning_rate": 4.239206376508717e-05, |
|
"loss": 1.1779, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.02067338301936257, |
|
"grad_norm": 0.15975573658943176, |
|
"learning_rate": 4.204344540477499e-05, |
|
"loss": 1.2125, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.021122804389348715, |
|
"grad_norm": 0.12576735019683838, |
|
"learning_rate": 4.16885219136787e-05, |
|
"loss": 1.1312, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.021572225759334856, |
|
"grad_norm": 0.16916967928409576, |
|
"learning_rate": 4.132742459383122e-05, |
|
"loss": 1.0823, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.022021647129321, |
|
"grad_norm": 0.17954471707344055, |
|
"learning_rate": 4.096028703124014e-05, |
|
"loss": 1.0728, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.02247106849930714, |
|
"grad_norm": 0.11972087621688843, |
|
"learning_rate": 4.058724504646834e-05, |
|
"loss": 1.0886, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.022920489869293285, |
|
"grad_norm": 0.14640676975250244, |
|
"learning_rate": 4.0208436644387834e-05, |
|
"loss": 1.0472, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.023369911239279426, |
|
"grad_norm": 0.13775284588336945, |
|
"learning_rate": 3.982400196312564e-05, |
|
"loss": 1.0984, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.02381933260926557, |
|
"grad_norm": 0.14184071123600006, |
|
"learning_rate": 3.943408322222049e-05, |
|
"loss": 1.1431, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.024268753979251715, |
|
"grad_norm": 0.1685701608657837, |
|
"learning_rate": 3.903882467000937e-05, |
|
"loss": 1.1531, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.024718175349237856, |
|
"grad_norm": 0.1468999981880188, |
|
"learning_rate": 3.8638372530263715e-05, |
|
"loss": 1.1069, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.025167596719224, |
|
"grad_norm": 0.13278649747371674, |
|
"learning_rate": 3.823287494809469e-05, |
|
"loss": 1.1124, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.025167596719224, |
|
"eval_loss": 1.1012423038482666, |
|
"eval_runtime": 312.1008, |
|
"eval_samples_per_second": 36.024, |
|
"eval_steps_per_second": 4.505, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.02561701808921014, |
|
"grad_norm": 0.180728018283844, |
|
"learning_rate": 3.782248193514766e-05, |
|
"loss": 1.0867, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.026066439459196285, |
|
"grad_norm": 0.15069565176963806, |
|
"learning_rate": 3.740734531410626e-05, |
|
"loss": 1.0624, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.026515860829182426, |
|
"grad_norm": 0.13892242312431335, |
|
"learning_rate": 3.698761866252635e-05, |
|
"loss": 1.0351, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.02696528219916857, |
|
"grad_norm": 0.1399199515581131, |
|
"learning_rate": 3.656345725602089e-05, |
|
"loss": 1.1609, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.027414703569154715, |
|
"grad_norm": 0.14930486679077148, |
|
"learning_rate": 3.6135018010816477e-05, |
|
"loss": 1.1117, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.027864124939140856, |
|
"grad_norm": 0.15556196868419647, |
|
"learning_rate": 3.570245942570315e-05, |
|
"loss": 1.1169, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.028313546309127, |
|
"grad_norm": 0.17272590100765228, |
|
"learning_rate": 3.526594152339845e-05, |
|
"loss": 1.115, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.02876296767911314, |
|
"grad_norm": 0.17533355951309204, |
|
"learning_rate": 3.4825625791348096e-05, |
|
"loss": 1.1298, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.029212389049099285, |
|
"grad_norm": 0.14778710901737213, |
|
"learning_rate": 3.438167512198436e-05, |
|
"loss": 1.1183, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.029661810419085426, |
|
"grad_norm": 0.14693984389305115, |
|
"learning_rate": 3.393425375246503e-05, |
|
"loss": 1.0647, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.03011123178907157, |
|
"grad_norm": 0.14994005858898163, |
|
"learning_rate": 3.348352720391469e-05, |
|
"loss": 1.0008, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.03056065315905771, |
|
"grad_norm": 0.1611510068178177, |
|
"learning_rate": 3.3029662220191144e-05, |
|
"loss": 1.094, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.031010074529043856, |
|
"grad_norm": 0.19615799188613892, |
|
"learning_rate": 3.2572826706199305e-05, |
|
"loss": 1.051, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.03145949589903, |
|
"grad_norm": 0.15789468586444855, |
|
"learning_rate": 3.211318966577581e-05, |
|
"loss": 1.0302, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.03145949589903, |
|
"eval_loss": 1.0971506834030151, |
|
"eval_runtime": 312.684, |
|
"eval_samples_per_second": 35.956, |
|
"eval_steps_per_second": 4.497, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.031908917269016145, |
|
"grad_norm": 0.15568239986896515, |
|
"learning_rate": 3.165092113916688e-05, |
|
"loss": 1.12, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.03235833863900228, |
|
"grad_norm": 0.16250278055667877, |
|
"learning_rate": 3.118619214012286e-05, |
|
"loss": 1.0651, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.032807760008988426, |
|
"grad_norm": 0.16043943166732788, |
|
"learning_rate": 3.071917459263264e-05, |
|
"loss": 1.1254, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.03325718137897457, |
|
"grad_norm": 0.19213679432868958, |
|
"learning_rate": 3.0250041267321232e-05, |
|
"loss": 1.1082, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.033706602748960715, |
|
"grad_norm": 0.17149952054023743, |
|
"learning_rate": 2.9778965717534313e-05, |
|
"loss": 1.0345, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.03415602411894685, |
|
"grad_norm": 0.19161058962345123, |
|
"learning_rate": 2.9306122215132976e-05, |
|
"loss": 1.2631, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.034605445488933, |
|
"grad_norm": 0.12983594834804535, |
|
"learning_rate": 2.8831685686022897e-05, |
|
"loss": 1.0125, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.03505486685891914, |
|
"grad_norm": 0.1923818439245224, |
|
"learning_rate": 2.8355831645441388e-05, |
|
"loss": 1.1152, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.035504288228905286, |
|
"grad_norm": 0.17539535462856293, |
|
"learning_rate": 2.787873613302649e-05, |
|
"loss": 1.0698, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.03595370959889143, |
|
"grad_norm": 0.20096057653427124, |
|
"learning_rate": 2.7400575647692046e-05, |
|
"loss": 1.0742, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.03640313096887757, |
|
"grad_norm": 0.1583949774503708, |
|
"learning_rate": 2.692152708233292e-05, |
|
"loss": 1.0255, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.03685255233886371, |
|
"grad_norm": 0.13673090934753418, |
|
"learning_rate": 2.6441767658384366e-05, |
|
"loss": 1.1167, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.037301973708849856, |
|
"grad_norm": 0.15132947266101837, |
|
"learning_rate": 2.596147486025996e-05, |
|
"loss": 1.053, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.037751395078836, |
|
"grad_norm": 0.1649513989686966, |
|
"learning_rate": 2.5480826369692178e-05, |
|
"loss": 1.1041, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.037751395078836, |
|
"eval_loss": 1.0942788124084473, |
|
"eval_runtime": 312.947, |
|
"eval_samples_per_second": 35.926, |
|
"eval_steps_per_second": 4.493, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.038200816448822145, |
|
"grad_norm": 0.19113826751708984, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.0117, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.03865023781880828, |
|
"grad_norm": 0.17212168872356415, |
|
"learning_rate": 2.4519173630307825e-05, |
|
"loss": 1.136, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.039099659188794426, |
|
"grad_norm": 0.16413848102092743, |
|
"learning_rate": 2.403852513974004e-05, |
|
"loss": 1.0806, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.03954908055878057, |
|
"grad_norm": 0.15564818680286407, |
|
"learning_rate": 2.3558232341615643e-05, |
|
"loss": 1.0233, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.039998501928766715, |
|
"grad_norm": 0.15014681220054626, |
|
"learning_rate": 2.3078472917667092e-05, |
|
"loss": 1.09, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.04044792329875285, |
|
"grad_norm": 0.1784486323595047, |
|
"learning_rate": 2.2599424352307957e-05, |
|
"loss": 1.0867, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.040897344668739, |
|
"grad_norm": 0.1629609763622284, |
|
"learning_rate": 2.212126386697352e-05, |
|
"loss": 1.0916, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.04134676603872514, |
|
"grad_norm": 0.1595894992351532, |
|
"learning_rate": 2.164416835455862e-05, |
|
"loss": 1.0537, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.041796187408711286, |
|
"grad_norm": 0.16272102296352386, |
|
"learning_rate": 2.11683143139771e-05, |
|
"loss": 1.1907, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.04224560877869743, |
|
"grad_norm": 0.15896858274936676, |
|
"learning_rate": 2.069387778486703e-05, |
|
"loss": 1.0492, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.04269503014868357, |
|
"grad_norm": 0.18164744973182678, |
|
"learning_rate": 2.02210342824657e-05, |
|
"loss": 1.064, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.04314445151866971, |
|
"grad_norm": 0.17921298742294312, |
|
"learning_rate": 1.9749958732678767e-05, |
|
"loss": 1.1456, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.043593872888655856, |
|
"grad_norm": 0.14521931111812592, |
|
"learning_rate": 1.928082540736737e-05, |
|
"loss": 1.1073, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.044043294258642, |
|
"grad_norm": 0.15362174808979034, |
|
"learning_rate": 1.8813807859877147e-05, |
|
"loss": 1.0171, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.044043294258642, |
|
"eval_loss": 1.0923734903335571, |
|
"eval_runtime": 312.595, |
|
"eval_samples_per_second": 35.967, |
|
"eval_steps_per_second": 4.498, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.044492715628628145, |
|
"grad_norm": 0.18710507452487946, |
|
"learning_rate": 1.8349078860833123e-05, |
|
"loss": 1.0978, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.04494213699861428, |
|
"grad_norm": 0.1569843292236328, |
|
"learning_rate": 1.7886810334224192e-05, |
|
"loss": 1.0783, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.045391558368600426, |
|
"grad_norm": 0.14208896458148956, |
|
"learning_rate": 1.74271732938007e-05, |
|
"loss": 1.0512, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.04584097973858657, |
|
"grad_norm": 0.16557104885578156, |
|
"learning_rate": 1.6970337779808862e-05, |
|
"loss": 1.143, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.046290401108572715, |
|
"grad_norm": 0.17679089307785034, |
|
"learning_rate": 1.6516472796085315e-05, |
|
"loss": 1.0428, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.04673982247855885, |
|
"grad_norm": 0.21449750661849976, |
|
"learning_rate": 1.6065746247534984e-05, |
|
"loss": 1.0541, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.047189243848545, |
|
"grad_norm": 0.19140197336673737, |
|
"learning_rate": 1.561832487801565e-05, |
|
"loss": 1.0943, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.04763866521853114, |
|
"grad_norm": 0.21146325767040253, |
|
"learning_rate": 1.5174374208651912e-05, |
|
"loss": 1.1286, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.048088086588517286, |
|
"grad_norm": 0.16576792299747467, |
|
"learning_rate": 1.4734058476601553e-05, |
|
"loss": 1.1543, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.04853750795850343, |
|
"grad_norm": 0.17730730772018433, |
|
"learning_rate": 1.4297540574296869e-05, |
|
"loss": 1.0809, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.04898692932848957, |
|
"grad_norm": 0.16303078830242157, |
|
"learning_rate": 1.386498198918352e-05, |
|
"loss": 1.0781, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.04943635069847571, |
|
"grad_norm": 0.15803277492523193, |
|
"learning_rate": 1.3436542743979125e-05, |
|
"loss": 1.1138, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.049885772068461856, |
|
"grad_norm": 0.161447212100029, |
|
"learning_rate": 1.3012381337473656e-05, |
|
"loss": 1.0087, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.050335193438448, |
|
"grad_norm": 0.14916126430034637, |
|
"learning_rate": 1.2592654685893757e-05, |
|
"loss": 1.0916, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.050335193438448, |
|
"eval_loss": 1.0906805992126465, |
|
"eval_runtime": 312.9768, |
|
"eval_samples_per_second": 35.923, |
|
"eval_steps_per_second": 4.492, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.05078461480843414, |
|
"grad_norm": 0.18624208867549896, |
|
"learning_rate": 1.217751806485235e-05, |
|
"loss": 1.1368, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.05123403617842028, |
|
"grad_norm": 0.18036052584648132, |
|
"learning_rate": 1.1767125051905315e-05, |
|
"loss": 1.0682, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.051683457548406427, |
|
"grad_norm": 0.17344728112220764, |
|
"learning_rate": 1.1361627469736285e-05, |
|
"loss": 1.0299, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.05213287891839257, |
|
"grad_norm": 0.17225950956344604, |
|
"learning_rate": 1.096117532999063e-05, |
|
"loss": 1.0217, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.052582300288378715, |
|
"grad_norm": 0.1770411878824234, |
|
"learning_rate": 1.0565916777779519e-05, |
|
"loss": 1.1027, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.05303172165836485, |
|
"grad_norm": 0.15190331637859344, |
|
"learning_rate": 1.0175998036874356e-05, |
|
"loss": 1.0684, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.053481143028351, |
|
"grad_norm": 0.20575201511383057, |
|
"learning_rate": 9.791563355612172e-06, |
|
"loss": 1.0359, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.05393056439833714, |
|
"grad_norm": 0.2163867950439453, |
|
"learning_rate": 9.412754953531663e-06, |
|
"loss": 1.0943, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.054379985768323286, |
|
"grad_norm": 0.1523335576057434, |
|
"learning_rate": 9.039712968759864e-06, |
|
"loss": 1.1357, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.05482940713830943, |
|
"grad_norm": 0.1748288869857788, |
|
"learning_rate": 8.672575406168782e-06, |
|
"loss": 1.0745, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.05527882850829557, |
|
"grad_norm": 0.16783830523490906, |
|
"learning_rate": 8.3114780863213e-06, |
|
"loss": 1.0432, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.05572824987828171, |
|
"grad_norm": 0.17801252007484436, |
|
"learning_rate": 7.956554595225016e-06, |
|
"loss": 1.0649, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.056177671248267856, |
|
"grad_norm": 0.17944374680519104, |
|
"learning_rate": 7.607936234912841e-06, |
|
"loss": 1.0862, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.056627092618254, |
|
"grad_norm": 0.19096186757087708, |
|
"learning_rate": 7.265751974868554e-06, |
|
"loss": 1.0477, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.056627092618254, |
|
"eval_loss": 1.0896837711334229, |
|
"eval_runtime": 312.799, |
|
"eval_samples_per_second": 35.943, |
|
"eval_steps_per_second": 4.495, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.05707651398824014, |
|
"grad_norm": 0.17599642276763916, |
|
"learning_rate": 6.930128404315214e-06, |
|
"loss": 1.1339, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.05752593535822628, |
|
"grad_norm": 0.16360458731651306, |
|
"learning_rate": 6.601189685384126e-06, |
|
"loss": 1.1844, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.05797535672821243, |
|
"grad_norm": 0.1602054089307785, |
|
"learning_rate": 6.279057507181796e-06, |
|
"loss": 1.0835, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.05842477809819857, |
|
"grad_norm": 0.33288466930389404, |
|
"learning_rate": 5.9638510407716394e-06, |
|
"loss": 1.0366, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.058874199468184715, |
|
"grad_norm": 0.17551077902317047, |
|
"learning_rate": 5.655686895087329e-06, |
|
"loss": 1.0361, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.05932362083817085, |
|
"grad_norm": 0.19817480444908142, |
|
"learning_rate": 5.354679073793942e-06, |
|
"loss": 1.1113, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.059773042208157, |
|
"grad_norm": 0.18397966027259827, |
|
"learning_rate": 5.060938933112891e-06, |
|
"loss": 1.0809, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.06022246357814314, |
|
"grad_norm": 0.18848171830177307, |
|
"learning_rate": 4.7745751406263165e-06, |
|
"loss": 1.0232, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.060671884948129286, |
|
"grad_norm": 0.18873141705989838, |
|
"learning_rate": 4.495693635076101e-06, |
|
"loss": 1.0633, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.06112130631811542, |
|
"grad_norm": 0.17433975636959076, |
|
"learning_rate": 4.224397587172402e-06, |
|
"loss": 1.0064, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.06157072768810157, |
|
"grad_norm": 0.2025955468416214, |
|
"learning_rate": 3.9607873614261715e-06, |
|
"loss": 1.1142, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.06202014905808771, |
|
"grad_norm": 0.1579374223947525, |
|
"learning_rate": 3.7049604790198976e-06, |
|
"loss": 1.0722, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.062469570428073856, |
|
"grad_norm": 0.15210741758346558, |
|
"learning_rate": 3.4570115817301243e-06, |
|
"loss": 1.011, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.06291899179806, |
|
"grad_norm": 0.21344028413295746, |
|
"learning_rate": 3.217032396915265e-06, |
|
"loss": 1.1333, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.06291899179806, |
|
"eval_loss": 1.0891081094741821, |
|
"eval_runtime": 312.8277, |
|
"eval_samples_per_second": 35.94, |
|
"eval_steps_per_second": 4.494, |
|
"step": 420 |
|
} |
|
], |
|
"logging_steps": 3, |
|
"max_steps": 500, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 42, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.3537807518872371e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|