|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 1861, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0005373455131649651, |
|
"grad_norm": 23.072626458069166, |
|
"learning_rate": 5.3475935828877005e-08, |
|
"loss": 1.329, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0026867275658248252, |
|
"grad_norm": 23.380544253320735, |
|
"learning_rate": 2.6737967914438503e-07, |
|
"loss": 1.3238, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0053734551316496505, |
|
"grad_norm": 14.233532908496148, |
|
"learning_rate": 5.347593582887701e-07, |
|
"loss": 1.2647, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.008060182697474477, |
|
"grad_norm": 11.791057381257374, |
|
"learning_rate": 8.021390374331551e-07, |
|
"loss": 1.1476, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.010746910263299301, |
|
"grad_norm": 7.369104708436159, |
|
"learning_rate": 1.0695187165775401e-06, |
|
"loss": 1.0118, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.013433637829124127, |
|
"grad_norm": 3.790654764997102, |
|
"learning_rate": 1.3368983957219254e-06, |
|
"loss": 0.9345, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.016120365394948953, |
|
"grad_norm": 3.447275461870041, |
|
"learning_rate": 1.6042780748663103e-06, |
|
"loss": 0.8918, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.018807092960773777, |
|
"grad_norm": 2.9186352407291847, |
|
"learning_rate": 1.8716577540106954e-06, |
|
"loss": 0.8518, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.021493820526598602, |
|
"grad_norm": 2.9408795060935695, |
|
"learning_rate": 2.1390374331550802e-06, |
|
"loss": 0.844, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.02418054809242343, |
|
"grad_norm": 2.8790720251278565, |
|
"learning_rate": 2.4064171122994653e-06, |
|
"loss": 0.8318, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.026867275658248254, |
|
"grad_norm": 2.963731808583039, |
|
"learning_rate": 2.673796791443851e-06, |
|
"loss": 0.8099, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02955400322407308, |
|
"grad_norm": 2.8588219846284297, |
|
"learning_rate": 2.9411764705882355e-06, |
|
"loss": 0.8054, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.032240730789897906, |
|
"grad_norm": 3.105229156653952, |
|
"learning_rate": 3.2085561497326205e-06, |
|
"loss": 0.7913, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.03492745835572273, |
|
"grad_norm": 3.1413414626518943, |
|
"learning_rate": 3.4759358288770056e-06, |
|
"loss": 0.7867, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.037614185921547555, |
|
"grad_norm": 3.132841956959321, |
|
"learning_rate": 3.7433155080213907e-06, |
|
"loss": 0.7893, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.04030091348737238, |
|
"grad_norm": 3.0623425732500142, |
|
"learning_rate": 4.010695187165775e-06, |
|
"loss": 0.7717, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.042987641053197204, |
|
"grad_norm": 9.233814569420407, |
|
"learning_rate": 4.2780748663101604e-06, |
|
"loss": 0.7715, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.04567436861902203, |
|
"grad_norm": 3.3097521575259368, |
|
"learning_rate": 4.5454545454545455e-06, |
|
"loss": 0.761, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.04836109618484686, |
|
"grad_norm": 3.0349283160634597, |
|
"learning_rate": 4.812834224598931e-06, |
|
"loss": 0.7621, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.051047823750671684, |
|
"grad_norm": 3.1760527715984153, |
|
"learning_rate": 5.0802139037433165e-06, |
|
"loss": 0.7591, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.05373455131649651, |
|
"grad_norm": 2.928755685153135, |
|
"learning_rate": 5.347593582887702e-06, |
|
"loss": 0.7433, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.05642127888232133, |
|
"grad_norm": 3.0202006224727174, |
|
"learning_rate": 5.614973262032086e-06, |
|
"loss": 0.7314, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.05910800644814616, |
|
"grad_norm": 3.1295278900578682, |
|
"learning_rate": 5.882352941176471e-06, |
|
"loss": 0.7483, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.06179473401397098, |
|
"grad_norm": 3.0863071896945167, |
|
"learning_rate": 6.149732620320856e-06, |
|
"loss": 0.7276, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.06448146157979581, |
|
"grad_norm": 3.0744127119070526, |
|
"learning_rate": 6.417112299465241e-06, |
|
"loss": 0.7183, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.06716818914562063, |
|
"grad_norm": 3.22926416708401, |
|
"learning_rate": 6.684491978609626e-06, |
|
"loss": 0.7242, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.06985491671144546, |
|
"grad_norm": 2.9238184305563033, |
|
"learning_rate": 6.951871657754011e-06, |
|
"loss": 0.7194, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.07254164427727028, |
|
"grad_norm": 2.907399933470749, |
|
"learning_rate": 7.219251336898396e-06, |
|
"loss": 0.7204, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.07522837184309511, |
|
"grad_norm": 2.8941698993873137, |
|
"learning_rate": 7.486631016042781e-06, |
|
"loss": 0.6986, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.07791509940891994, |
|
"grad_norm": 2.899562495835514, |
|
"learning_rate": 7.754010695187166e-06, |
|
"loss": 0.7021, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.08060182697474476, |
|
"grad_norm": 2.8740571869341687, |
|
"learning_rate": 8.02139037433155e-06, |
|
"loss": 0.7096, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.08328855454056959, |
|
"grad_norm": 2.912059141124366, |
|
"learning_rate": 8.288770053475937e-06, |
|
"loss": 0.7012, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.08597528210639441, |
|
"grad_norm": 2.933852566630193, |
|
"learning_rate": 8.556149732620321e-06, |
|
"loss": 0.6909, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.08866200967221924, |
|
"grad_norm": 2.954361982135904, |
|
"learning_rate": 8.823529411764707e-06, |
|
"loss": 0.6947, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.09134873723804406, |
|
"grad_norm": 2.914429965048127, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 0.7064, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.09403546480386889, |
|
"grad_norm": 2.824165175450085, |
|
"learning_rate": 9.358288770053477e-06, |
|
"loss": 0.6962, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.09672219236969372, |
|
"grad_norm": 2.7629699563001724, |
|
"learning_rate": 9.625668449197861e-06, |
|
"loss": 0.7093, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.09940891993551854, |
|
"grad_norm": 2.9700788007957484, |
|
"learning_rate": 9.893048128342247e-06, |
|
"loss": 0.7001, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.10209564750134337, |
|
"grad_norm": 2.720187963907155, |
|
"learning_rate": 9.999920755303033e-06, |
|
"loss": 0.6972, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.10478237506716818, |
|
"grad_norm": 2.579060861199631, |
|
"learning_rate": 9.999436491251425e-06, |
|
"loss": 0.7011, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.10746910263299302, |
|
"grad_norm": 2.650831290761427, |
|
"learning_rate": 9.998512030567253e-06, |
|
"loss": 0.7045, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.11015583019881783, |
|
"grad_norm": 3.3197406459470673, |
|
"learning_rate": 9.99714745464859e-06, |
|
"loss": 0.6992, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.11284255776464266, |
|
"grad_norm": 3.0709444563569126, |
|
"learning_rate": 9.995342883645325e-06, |
|
"loss": 0.708, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.1155292853304675, |
|
"grad_norm": 2.815374931010047, |
|
"learning_rate": 9.993098476448576e-06, |
|
"loss": 0.7011, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.11821601289629231, |
|
"grad_norm": 2.9613320208688005, |
|
"learning_rate": 9.990414430676716e-06, |
|
"loss": 0.6821, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.12090274046211714, |
|
"grad_norm": 2.690351994232397, |
|
"learning_rate": 9.987290982657961e-06, |
|
"loss": 0.6848, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.12358946802794196, |
|
"grad_norm": 3.5601408393747445, |
|
"learning_rate": 9.983728407409565e-06, |
|
"loss": 0.6832, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.1262761955937668, |
|
"grad_norm": 2.4701820214975014, |
|
"learning_rate": 9.979727018613607e-06, |
|
"loss": 0.6934, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.12896292315959162, |
|
"grad_norm": 2.4626384928709983, |
|
"learning_rate": 9.975287168589369e-06, |
|
"loss": 0.6785, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.13164965072541646, |
|
"grad_norm": 2.588440598744194, |
|
"learning_rate": 9.970409248262314e-06, |
|
"loss": 0.6978, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.13433637829124126, |
|
"grad_norm": 2.6880919710650413, |
|
"learning_rate": 9.965093687129669e-06, |
|
"loss": 0.6808, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.1370231058570661, |
|
"grad_norm": 2.9087730515287955, |
|
"learning_rate": 9.959340953222602e-06, |
|
"loss": 0.6758, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.13970983342289092, |
|
"grad_norm": 2.6772738824086075, |
|
"learning_rate": 9.953151553065019e-06, |
|
"loss": 0.6698, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.14239656098871575, |
|
"grad_norm": 2.669218732871304, |
|
"learning_rate": 9.94652603162896e-06, |
|
"loss": 0.6939, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.14508328855454056, |
|
"grad_norm": 2.4902462796339457, |
|
"learning_rate": 9.939464972286618e-06, |
|
"loss": 0.6702, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.1477700161203654, |
|
"grad_norm": 2.4766182391586975, |
|
"learning_rate": 9.931968996758972e-06, |
|
"loss": 0.6697, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.15045674368619022, |
|
"grad_norm": 2.5255485155595228, |
|
"learning_rate": 9.924038765061042e-06, |
|
"loss": 0.6665, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.15314347125201505, |
|
"grad_norm": 2.5656888228811794, |
|
"learning_rate": 9.915674975443779e-06, |
|
"loss": 0.6565, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.15583019881783988, |
|
"grad_norm": 2.825801360492887, |
|
"learning_rate": 9.906878364332586e-06, |
|
"loss": 0.6642, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.15851692638366469, |
|
"grad_norm": 2.6497249215881253, |
|
"learning_rate": 9.897649706262474e-06, |
|
"loss": 0.6624, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.16120365394948952, |
|
"grad_norm": 2.6523672302785575, |
|
"learning_rate": 9.88798981380986e-06, |
|
"loss": 0.6659, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.16389038151531435, |
|
"grad_norm": 2.6257923259309495, |
|
"learning_rate": 9.877899537521028e-06, |
|
"loss": 0.6533, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.16657710908113918, |
|
"grad_norm": 2.465767578464491, |
|
"learning_rate": 9.867379765837237e-06, |
|
"loss": 0.6518, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.169263836646964, |
|
"grad_norm": 2.556282218650964, |
|
"learning_rate": 9.85643142501649e-06, |
|
"loss": 0.6559, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.17195056421278881, |
|
"grad_norm": 2.6286087389340227, |
|
"learning_rate": 9.845055479051986e-06, |
|
"loss": 0.6469, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.17463729177861365, |
|
"grad_norm": 2.419535618698552, |
|
"learning_rate": 9.833252929587231e-06, |
|
"loss": 0.6631, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.17732401934443848, |
|
"grad_norm": 2.4542391410254956, |
|
"learning_rate": 9.821024815827854e-06, |
|
"loss": 0.6591, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.1800107469102633, |
|
"grad_norm": 2.4908873755968997, |
|
"learning_rate": 9.808372214450093e-06, |
|
"loss": 0.6658, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.1826974744760881, |
|
"grad_norm": 2.3733848426387176, |
|
"learning_rate": 9.795296239506011e-06, |
|
"loss": 0.662, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.18538420204191294, |
|
"grad_norm": 2.865296798119886, |
|
"learning_rate": 9.781798042325392e-06, |
|
"loss": 0.6526, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.18807092960773777, |
|
"grad_norm": 2.487445934484789, |
|
"learning_rate": 9.767878811414373e-06, |
|
"loss": 0.6525, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1907576571735626, |
|
"grad_norm": 2.388699669948045, |
|
"learning_rate": 9.753539772350792e-06, |
|
"loss": 0.647, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.19344438473938744, |
|
"grad_norm": 2.445554737991476, |
|
"learning_rate": 9.738782187676282e-06, |
|
"loss": 0.6425, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.19613111230521224, |
|
"grad_norm": 2.435104584478995, |
|
"learning_rate": 9.723607356785103e-06, |
|
"loss": 0.6443, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.19881783987103707, |
|
"grad_norm": 2.479158175926532, |
|
"learning_rate": 9.70801661580973e-06, |
|
"loss": 0.6575, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.2015045674368619, |
|
"grad_norm": 2.637755494092567, |
|
"learning_rate": 9.692011337503212e-06, |
|
"loss": 0.6421, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.20419129500268673, |
|
"grad_norm": 2.520737872507786, |
|
"learning_rate": 9.675592931118293e-06, |
|
"loss": 0.6358, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.20687802256851157, |
|
"grad_norm": 2.3907603747141675, |
|
"learning_rate": 9.658762842283343e-06, |
|
"loss": 0.6285, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.20956475013433637, |
|
"grad_norm": 2.3651549722988023, |
|
"learning_rate": 9.641522552875055e-06, |
|
"loss": 0.6119, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.2122514777001612, |
|
"grad_norm": 2.5208115737245604, |
|
"learning_rate": 9.62387358088798e-06, |
|
"loss": 0.6333, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.21493820526598603, |
|
"grad_norm": 2.6819070319357037, |
|
"learning_rate": 9.605817480300863e-06, |
|
"loss": 0.6205, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.21762493283181086, |
|
"grad_norm": 2.356493009815594, |
|
"learning_rate": 9.587355840939813e-06, |
|
"loss": 0.64, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.22031166039763567, |
|
"grad_norm": 2.405371194871644, |
|
"learning_rate": 9.568490288338324e-06, |
|
"loss": 0.6245, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.2229983879634605, |
|
"grad_norm": 2.6012482105829706, |
|
"learning_rate": 9.549222483594154e-06, |
|
"loss": 0.6451, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.22568511552928533, |
|
"grad_norm": 2.3956259391703196, |
|
"learning_rate": 9.529554123223053e-06, |
|
"loss": 0.6223, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.22837184309511016, |
|
"grad_norm": 2.398583860789219, |
|
"learning_rate": 9.5094869390094e-06, |
|
"loss": 0.6427, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.231058570660935, |
|
"grad_norm": 2.5232663534614863, |
|
"learning_rate": 9.48902269785371e-06, |
|
"loss": 0.6264, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.2337452982267598, |
|
"grad_norm": 2.4227785562286033, |
|
"learning_rate": 9.468163201617063e-06, |
|
"loss": 0.6172, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.23643202579258463, |
|
"grad_norm": 2.4898638907719586, |
|
"learning_rate": 9.446910286962453e-06, |
|
"loss": 0.6186, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.23911875335840946, |
|
"grad_norm": 2.3600937624812524, |
|
"learning_rate": 9.425265825193077e-06, |
|
"loss": 0.6009, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.2418054809242343, |
|
"grad_norm": 2.479426617963906, |
|
"learning_rate": 9.403231722087554e-06, |
|
"loss": 0.6373, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.24449220849005912, |
|
"grad_norm": 2.4357419818938117, |
|
"learning_rate": 9.380809917732132e-06, |
|
"loss": 0.6076, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.24717893605588392, |
|
"grad_norm": 2.47734706096724, |
|
"learning_rate": 9.358002386349862e-06, |
|
"loss": 0.6052, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.24986566362170876, |
|
"grad_norm": 2.545179042557431, |
|
"learning_rate": 9.334811136126778e-06, |
|
"loss": 0.6067, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.2525523911875336, |
|
"grad_norm": 2.576572600102202, |
|
"learning_rate": 9.31123820903506e-06, |
|
"loss": 0.6035, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.2552391187533584, |
|
"grad_norm": 2.4612256004699145, |
|
"learning_rate": 9.287285680653254e-06, |
|
"loss": 0.606, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.25792584631918325, |
|
"grad_norm": 2.495237864289767, |
|
"learning_rate": 9.262955659983522e-06, |
|
"loss": 0.5942, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.2606125738850081, |
|
"grad_norm": 2.65866167773882, |
|
"learning_rate": 9.238250289265921e-06, |
|
"loss": 0.599, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.2632993014508329, |
|
"grad_norm": 2.473870495050297, |
|
"learning_rate": 9.21317174378982e-06, |
|
"loss": 0.612, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.2659860290166577, |
|
"grad_norm": 2.3022323203953, |
|
"learning_rate": 9.187722231702326e-06, |
|
"loss": 0.5794, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.2686727565824825, |
|
"grad_norm": 2.3360448253147643, |
|
"learning_rate": 9.161903993813892e-06, |
|
"loss": 0.5821, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.27135948414830735, |
|
"grad_norm": 2.4400501057412454, |
|
"learning_rate": 9.135719303400995e-06, |
|
"loss": 0.5969, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.2740462117141322, |
|
"grad_norm": 2.324402047254847, |
|
"learning_rate": 9.10917046600598e-06, |
|
"loss": 0.5914, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.276732939279957, |
|
"grad_norm": 2.4328149953899296, |
|
"learning_rate": 9.082259819234063e-06, |
|
"loss": 0.5871, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.27941966684578184, |
|
"grad_norm": 2.326659710915374, |
|
"learning_rate": 9.054989732547507e-06, |
|
"loss": 0.5874, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.2821063944116067, |
|
"grad_norm": 2.4409777215420942, |
|
"learning_rate": 9.027362607056986e-06, |
|
"loss": 0.5964, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.2847931219774315, |
|
"grad_norm": 2.412955363836415, |
|
"learning_rate": 8.999380875310176e-06, |
|
"loss": 0.5877, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.28747984954325634, |
|
"grad_norm": 2.8194077899428516, |
|
"learning_rate": 8.971047001077561e-06, |
|
"loss": 0.5883, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.2901665771090811, |
|
"grad_norm": 2.469304660089161, |
|
"learning_rate": 8.942363479135516e-06, |
|
"loss": 0.5801, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.29285330467490595, |
|
"grad_norm": 2.3564533376477192, |
|
"learning_rate": 8.913332835046629e-06, |
|
"loss": 0.576, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.2955400322407308, |
|
"grad_norm": 2.4491464187209067, |
|
"learning_rate": 8.883957624937333e-06, |
|
"loss": 0.5792, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.2982267598065556, |
|
"grad_norm": 2.3159629602291276, |
|
"learning_rate": 8.854240435272842e-06, |
|
"loss": 0.5617, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.30091348737238044, |
|
"grad_norm": 2.5317785480271584, |
|
"learning_rate": 8.824183882629411e-06, |
|
"loss": 0.5669, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.30360021493820527, |
|
"grad_norm": 2.3852078136194135, |
|
"learning_rate": 8.793790613463956e-06, |
|
"loss": 0.5697, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.3062869425040301, |
|
"grad_norm": 2.4921927583022168, |
|
"learning_rate": 8.763063303881021e-06, |
|
"loss": 0.5796, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.30897367006985493, |
|
"grad_norm": 9.308745310175, |
|
"learning_rate": 8.73200465939717e-06, |
|
"loss": 0.5735, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.31166039763567976, |
|
"grad_norm": 2.4517036683344458, |
|
"learning_rate": 8.700617414702746e-06, |
|
"loss": 0.5671, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.31434712520150454, |
|
"grad_norm": 2.2949061651822635, |
|
"learning_rate": 8.668904333421098e-06, |
|
"loss": 0.5696, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.31703385276732937, |
|
"grad_norm": 2.2916177476898096, |
|
"learning_rate": 8.636868207865244e-06, |
|
"loss": 0.5756, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.3197205803331542, |
|
"grad_norm": 2.5518854172821297, |
|
"learning_rate": 8.604511858792006e-06, |
|
"loss": 0.5622, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.32240730789897903, |
|
"grad_norm": 2.4475160043733797, |
|
"learning_rate": 8.571838135153645e-06, |
|
"loss": 0.5525, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.32509403546480387, |
|
"grad_norm": 2.3488225651794497, |
|
"learning_rate": 8.538849913847019e-06, |
|
"loss": 0.569, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.3277807630306287, |
|
"grad_norm": 2.353785375547004, |
|
"learning_rate": 8.505550099460264e-06, |
|
"loss": 0.5575, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.33046749059645353, |
|
"grad_norm": 2.5058455114214166, |
|
"learning_rate": 8.471941624017058e-06, |
|
"loss": 0.5497, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.33315421816227836, |
|
"grad_norm": 2.4037582668437674, |
|
"learning_rate": 8.43802744671845e-06, |
|
"loss": 0.5592, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.3358409457281032, |
|
"grad_norm": 2.3703138788584206, |
|
"learning_rate": 8.403810553682307e-06, |
|
"loss": 0.55, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.338527673293928, |
|
"grad_norm": 2.4422133082637605, |
|
"learning_rate": 8.369293957680397e-06, |
|
"loss": 0.5451, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.3412144008597528, |
|
"grad_norm": 2.372256863029062, |
|
"learning_rate": 8.334480697873101e-06, |
|
"loss": 0.5616, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.34390112842557763, |
|
"grad_norm": 2.30267735280899, |
|
"learning_rate": 8.299373839541829e-06, |
|
"loss": 0.547, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.34658785599140246, |
|
"grad_norm": 2.312324577995851, |
|
"learning_rate": 8.26397647381912e-06, |
|
"loss": 0.5406, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.3492745835572273, |
|
"grad_norm": 2.4374915255270175, |
|
"learning_rate": 8.228291717416472e-06, |
|
"loss": 0.5388, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.3519613111230521, |
|
"grad_norm": 2.3871133470130674, |
|
"learning_rate": 8.192322712349917e-06, |
|
"loss": 0.5432, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.35464803868887695, |
|
"grad_norm": 2.519969391683205, |
|
"learning_rate": 8.15607262566337e-06, |
|
"loss": 0.5348, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.3573347662547018, |
|
"grad_norm": 2.4866718815954623, |
|
"learning_rate": 8.119544649149762e-06, |
|
"loss": 0.5567, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.3600214938205266, |
|
"grad_norm": 2.401099368264956, |
|
"learning_rate": 8.08274199907003e-06, |
|
"loss": 0.5409, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.36270822138635145, |
|
"grad_norm": 2.320605485178993, |
|
"learning_rate": 8.0456679158699e-06, |
|
"loss": 0.5283, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.3653949489521762, |
|
"grad_norm": 2.3005736041616642, |
|
"learning_rate": 8.008325663894586e-06, |
|
"loss": 0.5368, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.36808167651800106, |
|
"grad_norm": 2.305647834819274, |
|
"learning_rate": 7.970718531101365e-06, |
|
"loss": 0.5398, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.3707684040838259, |
|
"grad_norm": 2.3268228923372263, |
|
"learning_rate": 7.932849828770062e-06, |
|
"loss": 0.5348, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.3734551316496507, |
|
"grad_norm": 2.2813438553059053, |
|
"learning_rate": 7.89472289121151e-06, |
|
"loss": 0.5602, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.37614185921547555, |
|
"grad_norm": 2.2064861994852816, |
|
"learning_rate": 7.856341075473963e-06, |
|
"loss": 0.5227, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.3788285867813004, |
|
"grad_norm": 2.385886752618868, |
|
"learning_rate": 7.817707761047498e-06, |
|
"loss": 0.5292, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.3815153143471252, |
|
"grad_norm": 2.3717077533920867, |
|
"learning_rate": 7.77882634956647e-06, |
|
"loss": 0.5332, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.38420204191295004, |
|
"grad_norm": 2.389159454524817, |
|
"learning_rate": 7.739700264509993e-06, |
|
"loss": 0.5236, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.3868887694787749, |
|
"grad_norm": 2.314814719990629, |
|
"learning_rate": 7.700332950900504e-06, |
|
"loss": 0.5217, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.38957549704459965, |
|
"grad_norm": 2.207891685389126, |
|
"learning_rate": 7.660727875000432e-06, |
|
"loss": 0.5078, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.3922622246104245, |
|
"grad_norm": 2.3264819673930432, |
|
"learning_rate": 7.6208885240069995e-06, |
|
"loss": 0.5267, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.3949489521762493, |
|
"grad_norm": 2.288819911752258, |
|
"learning_rate": 7.5808184057451765e-06, |
|
"loss": 0.5067, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.39763567974207414, |
|
"grad_norm": 2.3653593555899857, |
|
"learning_rate": 7.540521048358814e-06, |
|
"loss": 0.5344, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.400322407307899, |
|
"grad_norm": 2.4648359924535743, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.5227, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.4030091348737238, |
|
"grad_norm": 2.409473996165666, |
|
"learning_rate": 7.459258828516645e-06, |
|
"loss": 0.5096, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.40569586243954864, |
|
"grad_norm": 2.3668323179746236, |
|
"learning_rate": 7.418301121138335e-06, |
|
"loss": 0.5171, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.40838259000537347, |
|
"grad_norm": 2.4782926067309825, |
|
"learning_rate": 7.3771304841604764e-06, |
|
"loss": 0.51, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.4110693175711983, |
|
"grad_norm": 2.267794928141394, |
|
"learning_rate": 7.335750542626772e-06, |
|
"loss": 0.5245, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.41375604513702313, |
|
"grad_norm": 2.318079683575523, |
|
"learning_rate": 7.294164940010031e-06, |
|
"loss": 0.5124, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.4164427727028479, |
|
"grad_norm": 2.1909268149999988, |
|
"learning_rate": 7.2523773378913655e-06, |
|
"loss": 0.5168, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.41912950026867274, |
|
"grad_norm": 2.2456539972858773, |
|
"learning_rate": 7.210391415637797e-06, |
|
"loss": 0.4946, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.42181622783449757, |
|
"grad_norm": 2.4974603199079133, |
|
"learning_rate": 7.168210870078277e-06, |
|
"loss": 0.5064, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.4245029554003224, |
|
"grad_norm": 2.334444583921718, |
|
"learning_rate": 7.125839415178204e-06, |
|
"loss": 0.5116, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.42718968296614723, |
|
"grad_norm": 2.397330148256394, |
|
"learning_rate": 7.083280781712394e-06, |
|
"loss": 0.5081, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.42987641053197206, |
|
"grad_norm": 2.5416270893996273, |
|
"learning_rate": 7.0405387169365965e-06, |
|
"loss": 0.499, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.4325631380977969, |
|
"grad_norm": 2.3363347223952657, |
|
"learning_rate": 6.9976169842575526e-06, |
|
"loss": 0.5049, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.4352498656636217, |
|
"grad_norm": 2.296216857638828, |
|
"learning_rate": 6.9545193629016215e-06, |
|
"loss": 0.5168, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.43793659322944656, |
|
"grad_norm": 2.4900891751826237, |
|
"learning_rate": 6.911249647582036e-06, |
|
"loss": 0.4939, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.44062332079527133, |
|
"grad_norm": 2.3708479871646797, |
|
"learning_rate": 6.867811648164769e-06, |
|
"loss": 0.4968, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.44331004836109617, |
|
"grad_norm": 2.2942739105778007, |
|
"learning_rate": 6.824209189333082e-06, |
|
"loss": 0.4892, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.445996775926921, |
|
"grad_norm": 2.298491349290504, |
|
"learning_rate": 6.780446110250766e-06, |
|
"loss": 0.5064, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.4486835034927458, |
|
"grad_norm": 2.3571392373027624, |
|
"learning_rate": 6.736526264224101e-06, |
|
"loss": 0.5005, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.45137023105857066, |
|
"grad_norm": 2.459885767558715, |
|
"learning_rate": 6.692453518362587e-06, |
|
"loss": 0.4951, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.4540569586243955, |
|
"grad_norm": 2.411449421942393, |
|
"learning_rate": 6.648231753238431e-06, |
|
"loss": 0.4828, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.4567436861902203, |
|
"grad_norm": 2.404363576863236, |
|
"learning_rate": 6.603864862544879e-06, |
|
"loss": 0.493, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.45943041375604515, |
|
"grad_norm": 2.2737487815721478, |
|
"learning_rate": 6.5593567527533715e-06, |
|
"loss": 0.4744, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.46211714132187, |
|
"grad_norm": 2.3678978985955186, |
|
"learning_rate": 6.514711342769588e-06, |
|
"loss": 0.5012, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.4648038688876948, |
|
"grad_norm": 2.356303245264604, |
|
"learning_rate": 6.469932563588386e-06, |
|
"loss": 0.4973, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.4674905964535196, |
|
"grad_norm": 2.7020262831984527, |
|
"learning_rate": 6.425024357947677e-06, |
|
"loss": 0.4918, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.4701773240193444, |
|
"grad_norm": 2.3340485214143745, |
|
"learning_rate": 6.3799906799812805e-06, |
|
"loss": 0.4982, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.47286405158516925, |
|
"grad_norm": 2.4191389152178373, |
|
"learning_rate": 6.334835494870759e-06, |
|
"loss": 0.4751, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.4755507791509941, |
|
"grad_norm": 2.272119239990629, |
|
"learning_rate": 6.289562778496285e-06, |
|
"loss": 0.4872, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.4782375067168189, |
|
"grad_norm": 2.300477262266863, |
|
"learning_rate": 6.244176517086573e-06, |
|
"loss": 0.4821, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.48092423428264375, |
|
"grad_norm": 2.446204046472319, |
|
"learning_rate": 6.1986807068678926e-06, |
|
"loss": 0.4823, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.4836109618484686, |
|
"grad_norm": 2.290872783259082, |
|
"learning_rate": 6.153079353712201e-06, |
|
"loss": 0.4683, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.4862976894142934, |
|
"grad_norm": 2.311264668993125, |
|
"learning_rate": 6.107376472784438e-06, |
|
"loss": 0.4759, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.48898441698011824, |
|
"grad_norm": 2.1766053144740805, |
|
"learning_rate": 6.061576088188981e-06, |
|
"loss": 0.4541, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.491671144545943, |
|
"grad_norm": 2.2863598186899563, |
|
"learning_rate": 6.015682232615336e-06, |
|
"loss": 0.4751, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.49435787211176785, |
|
"grad_norm": 2.3157344313618453, |
|
"learning_rate": 5.969698946983055e-06, |
|
"loss": 0.4579, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.4970445996775927, |
|
"grad_norm": 2.4556667211388308, |
|
"learning_rate": 5.923630280085948e-06, |
|
"loss": 0.4667, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.4997313272434175, |
|
"grad_norm": 2.2693647056660575, |
|
"learning_rate": 5.877480288235569e-06, |
|
"loss": 0.4642, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.5024180548092424, |
|
"grad_norm": 2.299235513307737, |
|
"learning_rate": 5.831253034904083e-06, |
|
"loss": 0.4625, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.5051047823750672, |
|
"grad_norm": 2.2594758653976603, |
|
"learning_rate": 5.7849525903664636e-06, |
|
"loss": 0.4741, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.507791509940892, |
|
"grad_norm": 2.346200404836091, |
|
"learning_rate": 5.738583031342123e-06, |
|
"loss": 0.4533, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.5104782375067168, |
|
"grad_norm": 2.198398726512657, |
|
"learning_rate": 5.692148440635946e-06, |
|
"loss": 0.4561, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.5131649650725416, |
|
"grad_norm": 2.3681691462263466, |
|
"learning_rate": 5.645652906778808e-06, |
|
"loss": 0.4621, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.5158516926383665, |
|
"grad_norm": 2.344944047074419, |
|
"learning_rate": 5.599100523667586e-06, |
|
"loss": 0.4599, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.5185384202041913, |
|
"grad_norm": 2.2092289217419085, |
|
"learning_rate": 5.552495390204691e-06, |
|
"loss": 0.4722, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.5212251477700162, |
|
"grad_norm": 2.3297576895231606, |
|
"learning_rate": 5.505841609937162e-06, |
|
"loss": 0.4532, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.5239118753358409, |
|
"grad_norm": 2.267324308428592, |
|
"learning_rate": 5.4591432906953515e-06, |
|
"loss": 0.4527, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.5265986029016658, |
|
"grad_norm": 2.4550058840280884, |
|
"learning_rate": 5.412404544231235e-06, |
|
"loss": 0.4562, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.5292853304674906, |
|
"grad_norm": 2.268535616599798, |
|
"learning_rate": 5.365629485856381e-06, |
|
"loss": 0.4454, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.5319720580333154, |
|
"grad_norm": 2.598323262451839, |
|
"learning_rate": 5.318822234079584e-06, |
|
"loss": 0.4614, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.5346587855991403, |
|
"grad_norm": 6.381482924500483, |
|
"learning_rate": 5.271986910244254e-06, |
|
"loss": 0.4366, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.537345513164965, |
|
"grad_norm": 2.359236695134924, |
|
"learning_rate": 5.225127638165514e-06, |
|
"loss": 0.4535, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.5400322407307899, |
|
"grad_norm": 2.268448533208311, |
|
"learning_rate": 5.178248543767122e-06, |
|
"loss": 0.4342, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.5427189682966147, |
|
"grad_norm": 2.358403761646346, |
|
"learning_rate": 5.1313537547181716e-06, |
|
"loss": 0.4405, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.5454056958624396, |
|
"grad_norm": 2.2710106949262805, |
|
"learning_rate": 5.084447400069656e-06, |
|
"loss": 0.4419, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.5480924234282644, |
|
"grad_norm": 2.368278941060436, |
|
"learning_rate": 5.037533609890917e-06, |
|
"loss": 0.4364, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.5507791509940893, |
|
"grad_norm": 2.284175829190933, |
|
"learning_rate": 4.990616514905982e-06, |
|
"loss": 0.4452, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.553465878559914, |
|
"grad_norm": 2.41532321709073, |
|
"learning_rate": 4.943700246129871e-06, |
|
"loss": 0.436, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.5561526061257388, |
|
"grad_norm": 2.3211025544242787, |
|
"learning_rate": 4.896788934504853e-06, |
|
"loss": 0.4467, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.5588393336915637, |
|
"grad_norm": 2.347444291429596, |
|
"learning_rate": 4.849886710536725e-06, |
|
"loss": 0.4288, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.5615260612573885, |
|
"grad_norm": 2.4127644606277014, |
|
"learning_rate": 4.802997703931124e-06, |
|
"loss": 0.4355, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.5642127888232134, |
|
"grad_norm": 2.4325621488004776, |
|
"learning_rate": 4.7561260432299015e-06, |
|
"loss": 0.4551, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.5668995163890381, |
|
"grad_norm": 2.242665015732091, |
|
"learning_rate": 4.7092758554476215e-06, |
|
"loss": 0.4364, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.569586243954863, |
|
"grad_norm": 2.4289606347980692, |
|
"learning_rate": 4.662451265708174e-06, |
|
"loss": 0.4316, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.5722729715206878, |
|
"grad_norm": 2.2828904861500257, |
|
"learning_rate": 4.6156563968815575e-06, |
|
"loss": 0.4298, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.5749596990865127, |
|
"grad_norm": 2.320278023559682, |
|
"learning_rate": 4.568895369220868e-06, |
|
"loss": 0.4332, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.5776464266523375, |
|
"grad_norm": 2.228672536344676, |
|
"learning_rate": 4.52217229999951e-06, |
|
"loss": 0.4291, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.5803331542181622, |
|
"grad_norm": 2.2818442860821664, |
|
"learning_rate": 4.47549130314868e-06, |
|
"loss": 0.4254, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.5830198817839871, |
|
"grad_norm": 2.218722467460458, |
|
"learning_rate": 4.428856488895128e-06, |
|
"loss": 0.441, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.5857066093498119, |
|
"grad_norm": 2.2852430401184907, |
|
"learning_rate": 4.382271963399268e-06, |
|
"loss": 0.4201, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.5883933369156368, |
|
"grad_norm": 2.2551743568535416, |
|
"learning_rate": 4.33574182839362e-06, |
|
"loss": 0.4284, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.5910800644814616, |
|
"grad_norm": 2.161938070818427, |
|
"learning_rate": 4.28927018082167e-06, |
|
"loss": 0.4274, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.5937667920472864, |
|
"grad_norm": 2.4061344797719753, |
|
"learning_rate": 4.2428611124771184e-06, |
|
"loss": 0.4141, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.5964535196131112, |
|
"grad_norm": 2.325443484824538, |
|
"learning_rate": 4.19651870964362e-06, |
|
"loss": 0.4286, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.5991402471789361, |
|
"grad_norm": 2.3440359985504133, |
|
"learning_rate": 4.150247052734979e-06, |
|
"loss": 0.4302, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.6018269747447609, |
|
"grad_norm": 2.4615226071875567, |
|
"learning_rate": 4.104050215935875e-06, |
|
"loss": 0.4111, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.6045137023105857, |
|
"grad_norm": 2.4429242324006983, |
|
"learning_rate": 4.0579322668431295e-06, |
|
"loss": 0.4193, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.6072004298764105, |
|
"grad_norm": 2.3015769237979677, |
|
"learning_rate": 4.011897266107567e-06, |
|
"loss": 0.42, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.6098871574422353, |
|
"grad_norm": 2.1719410742796, |
|
"learning_rate": 3.965949267076465e-06, |
|
"loss": 0.4027, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.6125738850080602, |
|
"grad_norm": 2.192156669745234, |
|
"learning_rate": 3.9200923154366685e-06, |
|
"loss": 0.423, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.615260612573885, |
|
"grad_norm": 2.2811472307718414, |
|
"learning_rate": 3.874330448858369e-06, |
|
"loss": 0.4142, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.6179473401397099, |
|
"grad_norm": 2.116499798174077, |
|
"learning_rate": 3.8286676966395895e-06, |
|
"loss": 0.4107, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.6206340677055346, |
|
"grad_norm": 2.244666553521093, |
|
"learning_rate": 3.7831080793514065e-06, |
|
"loss": 0.4017, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.6233207952713595, |
|
"grad_norm": 2.4675679743826295, |
|
"learning_rate": 3.7376556084839465e-06, |
|
"loss": 0.4059, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.6260075228371843, |
|
"grad_norm": 2.443117409755618, |
|
"learning_rate": 3.692314286093167e-06, |
|
"loss": 0.4162, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.6286942504030091, |
|
"grad_norm": 2.2029356738105363, |
|
"learning_rate": 3.647088104448494e-06, |
|
"loss": 0.4045, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.631380977968834, |
|
"grad_norm": 2.2250689532361045, |
|
"learning_rate": 3.601981045681292e-06, |
|
"loss": 0.4035, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.6340677055346587, |
|
"grad_norm": 2.3017262536884395, |
|
"learning_rate": 3.556997081434248e-06, |
|
"loss": 0.4043, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.6367544331004836, |
|
"grad_norm": 2.242688700397434, |
|
"learning_rate": 3.5121401725116653e-06, |
|
"loss": 0.405, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.6394411606663084, |
|
"grad_norm": 2.0746209840311747, |
|
"learning_rate": 3.4674142685307264e-06, |
|
"loss": 0.4086, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.6421278882321333, |
|
"grad_norm": 2.2459358965322993, |
|
"learning_rate": 3.4228233075737225e-06, |
|
"loss": 0.4121, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.6448146157979581, |
|
"grad_norm": 2.1930816457650235, |
|
"learning_rate": 3.3783712158413163e-06, |
|
"loss": 0.395, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.647501343363783, |
|
"grad_norm": 2.0405182523892145, |
|
"learning_rate": 3.3340619073068347e-06, |
|
"loss": 0.3865, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.6501880709296077, |
|
"grad_norm": 2.280921131796164, |
|
"learning_rate": 3.289899283371657e-06, |
|
"loss": 0.3905, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.6528747984954326, |
|
"grad_norm": 2.297801391720408, |
|
"learning_rate": 3.2458872325216893e-06, |
|
"loss": 0.3992, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.6555615260612574, |
|
"grad_norm": 2.2255966269222363, |
|
"learning_rate": 3.202029629984991e-06, |
|
"loss": 0.4012, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.6582482536270822, |
|
"grad_norm": 2.159969955127635, |
|
"learning_rate": 3.158330337390565e-06, |
|
"loss": 0.4008, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.6609349811929071, |
|
"grad_norm": 2.2717513219011636, |
|
"learning_rate": 3.1147932024283424e-06, |
|
"loss": 0.3915, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.6636217087587318, |
|
"grad_norm": 2.2223909448906247, |
|
"learning_rate": 3.071422058510394e-06, |
|
"loss": 0.3918, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.6663084363245567, |
|
"grad_norm": 2.3054337829096294, |
|
"learning_rate": 3.0282207244334084e-06, |
|
"loss": 0.4009, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.6689951638903815, |
|
"grad_norm": 2.2290900583766677, |
|
"learning_rate": 2.9851930040424383e-06, |
|
"loss": 0.4068, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.6716818914562064, |
|
"grad_norm": 2.1757281052534303, |
|
"learning_rate": 2.9423426858959892e-06, |
|
"loss": 0.3882, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.6743686190220312, |
|
"grad_norm": 2.1885464432673314, |
|
"learning_rate": 2.8996735429324256e-06, |
|
"loss": 0.389, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.677055346587856, |
|
"grad_norm": 2.0896855258083464, |
|
"learning_rate": 2.8571893321377773e-06, |
|
"loss": 0.3856, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.6797420741536808, |
|
"grad_norm": 2.313353541234272, |
|
"learning_rate": 2.8148937942149347e-06, |
|
"loss": 0.392, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.6824288017195056, |
|
"grad_norm": 2.3338524034127346, |
|
"learning_rate": 2.7727906532542783e-06, |
|
"loss": 0.3869, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.6851155292853305, |
|
"grad_norm": 2.084691037407018, |
|
"learning_rate": 2.7308836164057913e-06, |
|
"loss": 0.3826, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.6878022568511553, |
|
"grad_norm": 2.324796342452183, |
|
"learning_rate": 2.6891763735526223e-06, |
|
"loss": 0.3871, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.6904889844169801, |
|
"grad_norm": 2.160101574703143, |
|
"learning_rate": 2.6476725969862227e-06, |
|
"loss": 0.3728, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.6931757119828049, |
|
"grad_norm": 2.067721186262042, |
|
"learning_rate": 2.6063759410829813e-06, |
|
"loss": 0.3811, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.6958624395486298, |
|
"grad_norm": 2.4251135654528646, |
|
"learning_rate": 2.565290041982471e-06, |
|
"loss": 0.3884, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.6985491671144546, |
|
"grad_norm": 2.1893196908992123, |
|
"learning_rate": 2.524418517267283e-06, |
|
"loss": 0.3915, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.7012358946802795, |
|
"grad_norm": 2.1923809598940602, |
|
"learning_rate": 2.4837649656445117e-06, |
|
"loss": 0.3853, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.7039226222461042, |
|
"grad_norm": 2.2626797878823712, |
|
"learning_rate": 2.4433329666288774e-06, |
|
"loss": 0.3771, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.706609349811929, |
|
"grad_norm": 2.208206845682125, |
|
"learning_rate": 2.4031260802275623e-06, |
|
"loss": 0.3689, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.7092960773777539, |
|
"grad_norm": 2.18574695336372, |
|
"learning_rate": 2.3631478466267498e-06, |
|
"loss": 0.3728, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.7119828049435787, |
|
"grad_norm": 2.207451564888781, |
|
"learning_rate": 2.323401785879921e-06, |
|
"loss": 0.3758, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.7146695325094036, |
|
"grad_norm": 2.168655924857549, |
|
"learning_rate": 2.283891397597908e-06, |
|
"loss": 0.3672, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.7173562600752283, |
|
"grad_norm": 2.255193862061058, |
|
"learning_rate": 2.2446201606407537e-06, |
|
"loss": 0.3794, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.7200429876410532, |
|
"grad_norm": 2.2636128693117166, |
|
"learning_rate": 2.205591532811416e-06, |
|
"loss": 0.3742, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.722729715206878, |
|
"grad_norm": 2.1328893959069117, |
|
"learning_rate": 2.166808950551296e-06, |
|
"loss": 0.3778, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.7254164427727029, |
|
"grad_norm": 2.1602209025177705, |
|
"learning_rate": 2.128275828637664e-06, |
|
"loss": 0.3788, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.7281031703385277, |
|
"grad_norm": 2.28424029482692, |
|
"learning_rate": 2.089995559883004e-06, |
|
"loss": 0.371, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.7307898979043524, |
|
"grad_norm": 2.15439710916817, |
|
"learning_rate": 2.0519715148362585e-06, |
|
"loss": 0.3724, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.7334766254701773, |
|
"grad_norm": 2.1370354226282355, |
|
"learning_rate": 2.0142070414860704e-06, |
|
"loss": 0.3551, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.7361633530360021, |
|
"grad_norm": 2.1423156686501836, |
|
"learning_rate": 1.976705464965985e-06, |
|
"loss": 0.3817, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.738850080601827, |
|
"grad_norm": 2.1767372641376976, |
|
"learning_rate": 1.9394700872616856e-06, |
|
"loss": 0.3692, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.7415368081676518, |
|
"grad_norm": 2.1148705992310552, |
|
"learning_rate": 1.902504186920245e-06, |
|
"loss": 0.3701, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.7442235357334767, |
|
"grad_norm": 2.127192293566224, |
|
"learning_rate": 1.8658110187614538e-06, |
|
"loss": 0.3716, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.7469102632993014, |
|
"grad_norm": 2.1527774862898394, |
|
"learning_rate": 1.8293938135912475e-06, |
|
"loss": 0.3663, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.7495969908651263, |
|
"grad_norm": 2.3762653649020553, |
|
"learning_rate": 1.793255777917217e-06, |
|
"loss": 0.356, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.7522837184309511, |
|
"grad_norm": 2.0672220526784444, |
|
"learning_rate": 1.7574000936662928e-06, |
|
"loss": 0.3666, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.7549704459967759, |
|
"grad_norm": 2.131662760991029, |
|
"learning_rate": 1.7218299179045789e-06, |
|
"loss": 0.3777, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.7576571735626008, |
|
"grad_norm": 2.3447659932151423, |
|
"learning_rate": 1.6865483825593643e-06, |
|
"loss": 0.3718, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.7603439011284255, |
|
"grad_norm": 2.393514042224357, |
|
"learning_rate": 1.6515585941433694e-06, |
|
"loss": 0.3642, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.7630306286942504, |
|
"grad_norm": 2.3166220749115194, |
|
"learning_rate": 1.6168636334812126e-06, |
|
"loss": 0.3712, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.7657173562600752, |
|
"grad_norm": 2.1832693900107745, |
|
"learning_rate": 1.5824665554381579e-06, |
|
"loss": 0.3658, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.7684040838259001, |
|
"grad_norm": 2.0779160960958594, |
|
"learning_rate": 1.5483703886511191e-06, |
|
"loss": 0.3615, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.7710908113917249, |
|
"grad_norm": 2.098593414629146, |
|
"learning_rate": 1.5145781352620054e-06, |
|
"loss": 0.353, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.7737775389575497, |
|
"grad_norm": 2.1861074294439224, |
|
"learning_rate": 1.481092770653374e-06, |
|
"loss": 0.3599, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.7764642665233745, |
|
"grad_norm": 2.2448007733369923, |
|
"learning_rate": 1.4479172431864647e-06, |
|
"loss": 0.3548, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.7791509940891993, |
|
"grad_norm": 2.307957775652367, |
|
"learning_rate": 1.4150544739415755e-06, |
|
"loss": 0.3549, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.7818377216550242, |
|
"grad_norm": 2.1872047936836903, |
|
"learning_rate": 1.382507356460891e-06, |
|
"loss": 0.3533, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.784524449220849, |
|
"grad_norm": 2.1557683435766304, |
|
"learning_rate": 1.3502787564936875e-06, |
|
"loss": 0.3678, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.7872111767866738, |
|
"grad_norm": 2.088183401090359, |
|
"learning_rate": 1.3183715117440143e-06, |
|
"loss": 0.3452, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.7898979043524986, |
|
"grad_norm": 2.1559096637691195, |
|
"learning_rate": 1.2867884316208345e-06, |
|
"loss": 0.3605, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.7925846319183235, |
|
"grad_norm": 1.8737663503332451, |
|
"learning_rate": 1.255532296990662e-06, |
|
"loss": 0.3514, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.7952713594841483, |
|
"grad_norm": 2.196223751578153, |
|
"learning_rate": 1.2246058599327021e-06, |
|
"loss": 0.3511, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.7979580870499732, |
|
"grad_norm": 2.0822224737026294, |
|
"learning_rate": 1.194011843496537e-06, |
|
"loss": 0.3653, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.800644814615798, |
|
"grad_norm": 2.2216664507383457, |
|
"learning_rate": 1.163752941462362e-06, |
|
"loss": 0.3491, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.8033315421816228, |
|
"grad_norm": 2.259893004106475, |
|
"learning_rate": 1.1338318181038037e-06, |
|
"loss": 0.3495, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.8060182697474476, |
|
"grad_norm": 2.1666650973142687, |
|
"learning_rate": 1.1042511079533275e-06, |
|
"loss": 0.3611, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.8087049973132724, |
|
"grad_norm": 2.2659364719554898, |
|
"learning_rate": 1.0750134155702674e-06, |
|
"loss": 0.3626, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.8113917248790973, |
|
"grad_norm": 2.108890981344199, |
|
"learning_rate": 1.046121315311508e-06, |
|
"loss": 0.3522, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.814078452444922, |
|
"grad_norm": 1.9553754507307322, |
|
"learning_rate": 1.017577351104801e-06, |
|
"loss": 0.3491, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.8167651800107469, |
|
"grad_norm": 2.042354110902273, |
|
"learning_rate": 9.893840362247809e-07, |
|
"loss": 0.3489, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.8194519075765717, |
|
"grad_norm": 2.1701934651234613, |
|
"learning_rate": 9.615438530716753e-07, |
|
"loss": 0.3556, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.8221386351423966, |
|
"grad_norm": 2.298551493046633, |
|
"learning_rate": 9.340592529527237e-07, |
|
"loss": 0.3533, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.8248253627082214, |
|
"grad_norm": 2.0661919944679656, |
|
"learning_rate": 9.069326558663488e-07, |
|
"loss": 0.3504, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.8275120902740463, |
|
"grad_norm": 2.238182054477219, |
|
"learning_rate": 8.801664502890722e-07, |
|
"loss": 0.3583, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.830198817839871, |
|
"grad_norm": 2.206073109503085, |
|
"learning_rate": 8.537629929652186e-07, |
|
"loss": 0.3493, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.8328855454056958, |
|
"grad_norm": 2.112920269926838, |
|
"learning_rate": 8.277246086993962e-07, |
|
"loss": 0.3485, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.8355722729715207, |
|
"grad_norm": 2.1091946693423824, |
|
"learning_rate": 8.02053590151805e-07, |
|
"loss": 0.3602, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.8382590005373455, |
|
"grad_norm": 2.0674911667909592, |
|
"learning_rate": 7.767521976363735e-07, |
|
"loss": 0.344, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.8409457281031704, |
|
"grad_norm": 2.094630956022192, |
|
"learning_rate": 7.518226589217286e-07, |
|
"loss": 0.3555, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.8436324556689951, |
|
"grad_norm": 1.9871731688863412, |
|
"learning_rate": 7.27267169035053e-07, |
|
"loss": 0.3446, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.84631918323482, |
|
"grad_norm": 1.9968082439337067, |
|
"learning_rate": 7.030878900688115e-07, |
|
"loss": 0.3453, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.8490059108006448, |
|
"grad_norm": 2.0534061294688435, |
|
"learning_rate": 6.792869509903777e-07, |
|
"loss": 0.3471, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.8516926383664697, |
|
"grad_norm": 2.06210125589499, |
|
"learning_rate": 6.558664474545817e-07, |
|
"loss": 0.3395, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.8543793659322945, |
|
"grad_norm": 1.9567313896707874, |
|
"learning_rate": 6.328284416191893e-07, |
|
"loss": 0.3401, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.8570660934981192, |
|
"grad_norm": 2.060413092979503, |
|
"learning_rate": 6.101749619633346e-07, |
|
"loss": 0.3394, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.8597528210639441, |
|
"grad_norm": 2.1887059434963603, |
|
"learning_rate": 5.879080031089047e-07, |
|
"loss": 0.3543, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.8624395486297689, |
|
"grad_norm": 1.9845082909122282, |
|
"learning_rate": 5.660295256449233e-07, |
|
"loss": 0.3389, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.8651262761955938, |
|
"grad_norm": 2.3356240543268414, |
|
"learning_rate": 5.445414559549167e-07, |
|
"loss": 0.3403, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.8678130037614186, |
|
"grad_norm": 2.2880237510434123, |
|
"learning_rate": 5.234456860473042e-07, |
|
"loss": 0.333, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.8704997313272435, |
|
"grad_norm": 1.987792837037979, |
|
"learning_rate": 5.027440733887973e-07, |
|
"loss": 0.3413, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.8731864588930682, |
|
"grad_norm": 2.145351742840335, |
|
"learning_rate": 4.824384407408622e-07, |
|
"loss": 0.3419, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.8758731864588931, |
|
"grad_norm": 2.0454741447896034, |
|
"learning_rate": 4.625305759992205e-07, |
|
"loss": 0.3446, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.8785599140247179, |
|
"grad_norm": 2.3653559873352754, |
|
"learning_rate": 4.4302223203642803e-07, |
|
"loss": 0.3456, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.8812466415905427, |
|
"grad_norm": 1.9465586799987966, |
|
"learning_rate": 4.2391512654753443e-07, |
|
"loss": 0.3401, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.8839333691563676, |
|
"grad_norm": 2.0641483788200166, |
|
"learning_rate": 4.05210941898847e-07, |
|
"loss": 0.3444, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.8866200967221923, |
|
"grad_norm": 1.9634013461126347, |
|
"learning_rate": 3.8691132497979064e-07, |
|
"loss": 0.3262, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.8893068242880172, |
|
"grad_norm": 1.9771681166485953, |
|
"learning_rate": 3.6901788705790753e-07, |
|
"loss": 0.3339, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.891993551853842, |
|
"grad_norm": 2.0517920199325363, |
|
"learning_rate": 3.5153220363698225e-07, |
|
"loss": 0.3505, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.8946802794196669, |
|
"grad_norm": 1.9243621922759124, |
|
"learning_rate": 3.344558143183246e-07, |
|
"loss": 0.342, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.8973670069854917, |
|
"grad_norm": 1.9232396735880564, |
|
"learning_rate": 3.1779022266520245e-07, |
|
"loss": 0.347, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.9000537345513165, |
|
"grad_norm": 2.0661817058830305, |
|
"learning_rate": 3.015368960704584e-07, |
|
"loss": 0.3293, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.9027404621171413, |
|
"grad_norm": 2.1100704132509094, |
|
"learning_rate": 2.856972656273066e-07, |
|
"loss": 0.3331, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.9054271896829661, |
|
"grad_norm": 2.0718216201058444, |
|
"learning_rate": 2.7027272600332443e-07, |
|
"loss": 0.3422, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.908113917248791, |
|
"grad_norm": 2.1084735618419246, |
|
"learning_rate": 2.5526463531765467e-07, |
|
"loss": 0.3453, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.9108006448146158, |
|
"grad_norm": 1.8503057057006569, |
|
"learning_rate": 2.4067431502142414e-07, |
|
"loss": 0.3369, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.9134873723804406, |
|
"grad_norm": 2.2410702930225437, |
|
"learning_rate": 2.2650304978138916e-07, |
|
"loss": 0.3422, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.9161740999462654, |
|
"grad_norm": 1.988010259435427, |
|
"learning_rate": 2.1275208736682262e-07, |
|
"loss": 0.3273, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.9188608275120903, |
|
"grad_norm": 2.162508541847929, |
|
"learning_rate": 1.9942263853964917e-07, |
|
"loss": 0.348, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.9215475550779151, |
|
"grad_norm": 2.173499113968412, |
|
"learning_rate": 1.8651587694783924e-07, |
|
"loss": 0.3366, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.92423428264374, |
|
"grad_norm": 1.9783121068716565, |
|
"learning_rate": 1.7403293902206851e-07, |
|
"loss": 0.3282, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.9269210102095647, |
|
"grad_norm": 1.8953184142427286, |
|
"learning_rate": 1.6197492387565629e-07, |
|
"loss": 0.3346, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.9296077377753896, |
|
"grad_norm": 2.0149463904024474, |
|
"learning_rate": 1.503428932077916e-07, |
|
"loss": 0.3457, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.9322944653412144, |
|
"grad_norm": 2.0287919597007447, |
|
"learning_rate": 1.3913787121004717e-07, |
|
"loss": 0.3301, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.9349811929070392, |
|
"grad_norm": 2.049509308073804, |
|
"learning_rate": 1.2836084447620466e-07, |
|
"loss": 0.3483, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.9376679204728641, |
|
"grad_norm": 2.0687381489986714, |
|
"learning_rate": 1.180127619153837e-07, |
|
"loss": 0.3296, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.9403546480386888, |
|
"grad_norm": 2.129632171046833, |
|
"learning_rate": 1.0809453466849029e-07, |
|
"loss": 0.3416, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.9430413756045137, |
|
"grad_norm": 1.940321140330739, |
|
"learning_rate": 9.860703602799281e-08, |
|
"loss": 0.344, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.9457281031703385, |
|
"grad_norm": 2.2300828043810434, |
|
"learning_rate": 8.955110136102952e-08, |
|
"loss": 0.3332, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.9484148307361634, |
|
"grad_norm": 2.039077625849431, |
|
"learning_rate": 8.092752803585513e-08, |
|
"loss": 0.3398, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.9511015583019882, |
|
"grad_norm": 1.9309466845356567, |
|
"learning_rate": 7.273707535162988e-08, |
|
"loss": 0.3345, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.9537882858678131, |
|
"grad_norm": 2.085038229808163, |
|
"learning_rate": 6.498046447156958e-08, |
|
"loss": 0.3413, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.9564750134336378, |
|
"grad_norm": 2.073268982177981, |
|
"learning_rate": 5.7658378359443104e-08, |
|
"loss": 0.3389, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.9591617409994626, |
|
"grad_norm": 2.1191914780783185, |
|
"learning_rate": 5.077146171943936e-08, |
|
"loss": 0.3301, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.9618484685652875, |
|
"grad_norm": 2.0695293690655494, |
|
"learning_rate": 4.432032093940219e-08, |
|
"loss": 0.3387, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.9645351961311123, |
|
"grad_norm": 2.181866235925443, |
|
"learning_rate": 3.8305524037438035e-08, |
|
"loss": 0.3327, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.9672219236969372, |
|
"grad_norm": 2.0829327323896742, |
|
"learning_rate": 3.27276006119015e-08, |
|
"loss": 0.3269, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.9699086512627619, |
|
"grad_norm": 2.0402878486865084, |
|
"learning_rate": 2.7587041794766012e-08, |
|
"loss": 0.3428, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.9725953788285868, |
|
"grad_norm": 2.1613154681131386, |
|
"learning_rate": 2.2884300208378395e-08, |
|
"loss": 0.3389, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.9752821063944116, |
|
"grad_norm": 2.190787226227149, |
|
"learning_rate": 1.8619789925608534e-08, |
|
"loss": 0.3487, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.9779688339602365, |
|
"grad_norm": 1.9274721887684194, |
|
"learning_rate": 1.4793886433387417e-08, |
|
"loss": 0.3248, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.9806555615260613, |
|
"grad_norm": 2.0401602016541296, |
|
"learning_rate": 1.1406926599646373e-08, |
|
"loss": 0.3378, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.983342289091886, |
|
"grad_norm": 1.9245890826575305, |
|
"learning_rate": 8.459208643659122e-09, |
|
"loss": 0.3301, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.9860290166577109, |
|
"grad_norm": 2.145271841454969, |
|
"learning_rate": 5.950992109779452e-09, |
|
"loss": 0.3337, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.9887157442235357, |
|
"grad_norm": 2.102840117283832, |
|
"learning_rate": 3.88249784459227e-09, |
|
"loss": 0.3436, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.9914024717893606, |
|
"grad_norm": 1.8642178563070946, |
|
"learning_rate": 2.2539079774658303e-09, |
|
"loss": 0.3346, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.9940891993551854, |
|
"grad_norm": 2.0754130978586462, |
|
"learning_rate": 1.0653659045156695e-09, |
|
"loss": 0.323, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.9967759269210102, |
|
"grad_norm": 2.1177194452551293, |
|
"learning_rate": 3.1697627597970794e-10, |
|
"loss": 0.3273, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.999462654486835, |
|
"grad_norm": 2.036609175873343, |
|
"learning_rate": 8.804987003951937e-12, |
|
"loss": 0.3344, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_runtime": 3.3729, |
|
"eval_samples_per_second": 2.965, |
|
"eval_steps_per_second": 0.889, |
|
"step": 1861 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 1861, |
|
"total_flos": 194827769610240.0, |
|
"train_loss": 0.5009698675307838, |
|
"train_runtime": 16706.09, |
|
"train_samples_per_second": 1.782, |
|
"train_steps_per_second": 0.111 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1861, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 194827769610240.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|