|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 6181, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00016178611875101117, |
|
"grad_norm": 2.989531072795082, |
|
"learning_rate": 1.6155088852988694e-08, |
|
"loss": 1.6714, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0008089305937550558, |
|
"grad_norm": 3.0740133028448042, |
|
"learning_rate": 8.077544426494346e-08, |
|
"loss": 1.6455, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0016178611875101116, |
|
"grad_norm": 3.05751861374171, |
|
"learning_rate": 1.6155088852988693e-07, |
|
"loss": 1.6578, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0024267917812651673, |
|
"grad_norm": 3.05989958650721, |
|
"learning_rate": 2.4232633279483037e-07, |
|
"loss": 1.6633, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.003235722375020223, |
|
"grad_norm": 3.024001026912918, |
|
"learning_rate": 3.2310177705977386e-07, |
|
"loss": 1.6386, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.004044652968775279, |
|
"grad_norm": 3.1122902778444006, |
|
"learning_rate": 4.038772213247173e-07, |
|
"loss": 1.6687, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.004853583562530335, |
|
"grad_norm": 2.8890112766655296, |
|
"learning_rate": 4.846526655896607e-07, |
|
"loss": 1.6414, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0056625141562853904, |
|
"grad_norm": 2.6523925300683246, |
|
"learning_rate": 5.654281098546043e-07, |
|
"loss": 1.59, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.006471444750040446, |
|
"grad_norm": 2.603827366657395, |
|
"learning_rate": 6.462035541195477e-07, |
|
"loss": 1.6336, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.007280375343795502, |
|
"grad_norm": 2.577297564219146, |
|
"learning_rate": 7.269789983844912e-07, |
|
"loss": 1.6402, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.008089305937550558, |
|
"grad_norm": 2.439515573088333, |
|
"learning_rate": 8.077544426494346e-07, |
|
"loss": 1.6193, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.008898236531305614, |
|
"grad_norm": 2.4333132929087764, |
|
"learning_rate": 8.885298869143781e-07, |
|
"loss": 1.6122, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.00970716712506067, |
|
"grad_norm": 2.1113677966611464, |
|
"learning_rate": 9.693053311793215e-07, |
|
"loss": 1.6185, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.010516097718815726, |
|
"grad_norm": 1.8886265008348362, |
|
"learning_rate": 1.0500807754442651e-06, |
|
"loss": 1.5725, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.011325028312570781, |
|
"grad_norm": 1.8864578193982546, |
|
"learning_rate": 1.1308562197092086e-06, |
|
"loss": 1.5734, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.012133958906325838, |
|
"grad_norm": 1.6120301708706668, |
|
"learning_rate": 1.211631663974152e-06, |
|
"loss": 1.544, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.012942889500080892, |
|
"grad_norm": 1.5440235686664168, |
|
"learning_rate": 1.2924071082390954e-06, |
|
"loss": 1.5381, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.013751820093835949, |
|
"grad_norm": 1.3513181383313533, |
|
"learning_rate": 1.3731825525040387e-06, |
|
"loss": 1.5246, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.014560750687591004, |
|
"grad_norm": 1.4327471253077833, |
|
"learning_rate": 1.4539579967689823e-06, |
|
"loss": 1.4991, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01536968128134606, |
|
"grad_norm": 1.2301698886041794, |
|
"learning_rate": 1.5347334410339258e-06, |
|
"loss": 1.4618, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.016178611875101116, |
|
"grad_norm": 1.162576921517532, |
|
"learning_rate": 1.6155088852988692e-06, |
|
"loss": 1.4728, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01698754246885617, |
|
"grad_norm": 1.0168699965353747, |
|
"learning_rate": 1.6962843295638126e-06, |
|
"loss": 1.46, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.01779647306261123, |
|
"grad_norm": 1.0395803267996115, |
|
"learning_rate": 1.7770597738287563e-06, |
|
"loss": 1.4594, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.018605403656366284, |
|
"grad_norm": 0.9917353590665066, |
|
"learning_rate": 1.8578352180936995e-06, |
|
"loss": 1.4397, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.01941433425012134, |
|
"grad_norm": 0.9523144102295756, |
|
"learning_rate": 1.938610662358643e-06, |
|
"loss": 1.4234, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.020223264843876397, |
|
"grad_norm": 0.9258023739630329, |
|
"learning_rate": 2.0193861066235864e-06, |
|
"loss": 1.4226, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.021032195437631452, |
|
"grad_norm": 0.9483758496421177, |
|
"learning_rate": 2.1001615508885302e-06, |
|
"loss": 1.3762, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.021841126031386507, |
|
"grad_norm": 0.9530895080844857, |
|
"learning_rate": 2.1809369951534733e-06, |
|
"loss": 1.354, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.022650056625141562, |
|
"grad_norm": 0.9891655010686481, |
|
"learning_rate": 2.261712439418417e-06, |
|
"loss": 1.3418, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02345898721889662, |
|
"grad_norm": 1.0770341521239468, |
|
"learning_rate": 2.34248788368336e-06, |
|
"loss": 1.3239, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.024267917812651675, |
|
"grad_norm": 1.1321453887674335, |
|
"learning_rate": 2.423263327948304e-06, |
|
"loss": 1.2818, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02507684840640673, |
|
"grad_norm": 1.116259450785514, |
|
"learning_rate": 2.5040387722132474e-06, |
|
"loss": 1.2499, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.025885779000161785, |
|
"grad_norm": 1.1299351548572238, |
|
"learning_rate": 2.584814216478191e-06, |
|
"loss": 1.2403, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.026694709593916843, |
|
"grad_norm": 1.1335363316132652, |
|
"learning_rate": 2.6655896607431343e-06, |
|
"loss": 1.214, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.027503640187671898, |
|
"grad_norm": 1.1757124749289842, |
|
"learning_rate": 2.7463651050080773e-06, |
|
"loss": 1.163, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.028312570781426953, |
|
"grad_norm": 1.2651810945082007, |
|
"learning_rate": 2.827140549273021e-06, |
|
"loss": 1.1479, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.029121501375182008, |
|
"grad_norm": 1.2624388452227249, |
|
"learning_rate": 2.9079159935379646e-06, |
|
"loss": 1.0948, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.029930431968937066, |
|
"grad_norm": 1.2111655416186609, |
|
"learning_rate": 2.988691437802908e-06, |
|
"loss": 1.0487, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.03073936256269212, |
|
"grad_norm": 0.8837452655216318, |
|
"learning_rate": 3.0694668820678515e-06, |
|
"loss": 1.0465, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03154829315644718, |
|
"grad_norm": 0.8221498455856253, |
|
"learning_rate": 3.1502423263327954e-06, |
|
"loss": 1.0297, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.03235722375020223, |
|
"grad_norm": 0.8439527908802734, |
|
"learning_rate": 3.2310177705977384e-06, |
|
"loss": 1.0015, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.03316615434395729, |
|
"grad_norm": 0.7410910924751422, |
|
"learning_rate": 3.311793214862682e-06, |
|
"loss": 1.0173, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.03397508493771234, |
|
"grad_norm": 0.6701077230967789, |
|
"learning_rate": 3.3925686591276253e-06, |
|
"loss": 0.999, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.0347840155314674, |
|
"grad_norm": 0.6355481671837402, |
|
"learning_rate": 3.473344103392569e-06, |
|
"loss": 0.9725, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.03559294612522246, |
|
"grad_norm": 0.5875107208972431, |
|
"learning_rate": 3.5541195476575126e-06, |
|
"loss": 0.9867, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.03640187671897751, |
|
"grad_norm": 0.5994183367628626, |
|
"learning_rate": 3.6348949919224556e-06, |
|
"loss": 0.9943, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.03721080731273257, |
|
"grad_norm": 0.6182303752035966, |
|
"learning_rate": 3.715670436187399e-06, |
|
"loss": 0.9985, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.038019737906487626, |
|
"grad_norm": 0.5701097398554537, |
|
"learning_rate": 3.796445880452343e-06, |
|
"loss": 0.9736, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.03882866850024268, |
|
"grad_norm": 0.5584494985009502, |
|
"learning_rate": 3.877221324717286e-06, |
|
"loss": 0.9875, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.039637599093997736, |
|
"grad_norm": 0.5782425116149067, |
|
"learning_rate": 3.95799676898223e-06, |
|
"loss": 0.9701, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.040446529687752794, |
|
"grad_norm": 0.6299578118181779, |
|
"learning_rate": 4.038772213247173e-06, |
|
"loss": 0.9671, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.041255460281507846, |
|
"grad_norm": 0.5812417163133361, |
|
"learning_rate": 4.119547657512117e-06, |
|
"loss": 0.9579, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.042064390875262904, |
|
"grad_norm": 0.5781954014931824, |
|
"learning_rate": 4.2003231017770605e-06, |
|
"loss": 0.9685, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.042873321469017955, |
|
"grad_norm": 0.608052455969015, |
|
"learning_rate": 4.2810985460420035e-06, |
|
"loss": 0.9701, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.043682252062773014, |
|
"grad_norm": 0.5966466875482974, |
|
"learning_rate": 4.3618739903069465e-06, |
|
"loss": 0.9702, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.04449118265652807, |
|
"grad_norm": 0.5596371503375434, |
|
"learning_rate": 4.44264943457189e-06, |
|
"loss": 0.9563, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.045300113250283124, |
|
"grad_norm": 0.5994461579702871, |
|
"learning_rate": 4.523424878836834e-06, |
|
"loss": 0.9578, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.04610904384403818, |
|
"grad_norm": 0.5595808732955236, |
|
"learning_rate": 4.604200323101777e-06, |
|
"loss": 0.9607, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.04691797443779324, |
|
"grad_norm": 0.669537499667742, |
|
"learning_rate": 4.68497576736672e-06, |
|
"loss": 0.9476, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.04772690503154829, |
|
"grad_norm": 0.5511679026608909, |
|
"learning_rate": 4.765751211631664e-06, |
|
"loss": 0.9464, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.04853583562530335, |
|
"grad_norm": 0.5939912843219857, |
|
"learning_rate": 4.846526655896608e-06, |
|
"loss": 0.9429, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.0493447662190584, |
|
"grad_norm": 0.5798565127301802, |
|
"learning_rate": 4.927302100161551e-06, |
|
"loss": 0.9485, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.05015369681281346, |
|
"grad_norm": 0.5756338877596261, |
|
"learning_rate": 5.008077544426495e-06, |
|
"loss": 0.9572, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.05096262740656852, |
|
"grad_norm": 0.5842007801259245, |
|
"learning_rate": 5.088852988691439e-06, |
|
"loss": 0.9219, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.05177155800032357, |
|
"grad_norm": 0.5909234064830663, |
|
"learning_rate": 5.169628432956382e-06, |
|
"loss": 0.9463, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.05258048859407863, |
|
"grad_norm": 0.5964209125322439, |
|
"learning_rate": 5.250403877221325e-06, |
|
"loss": 0.9137, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.05338941918783369, |
|
"grad_norm": 0.6196287292433933, |
|
"learning_rate": 5.331179321486269e-06, |
|
"loss": 0.9275, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.05419834978158874, |
|
"grad_norm": 0.6885711787979655, |
|
"learning_rate": 5.411954765751212e-06, |
|
"loss": 0.9308, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.055007280375343796, |
|
"grad_norm": 0.5882034901608182, |
|
"learning_rate": 5.492730210016155e-06, |
|
"loss": 0.9226, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.055816210969098855, |
|
"grad_norm": 0.6167809439808443, |
|
"learning_rate": 5.573505654281099e-06, |
|
"loss": 0.9329, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.056625141562853906, |
|
"grad_norm": 0.6473624549151995, |
|
"learning_rate": 5.654281098546042e-06, |
|
"loss": 0.9342, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.057434072156608965, |
|
"grad_norm": 0.6268324107561812, |
|
"learning_rate": 5.735056542810986e-06, |
|
"loss": 0.9444, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.058243002750364016, |
|
"grad_norm": 0.6081881922413113, |
|
"learning_rate": 5.815831987075929e-06, |
|
"loss": 0.9266, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.059051933344119074, |
|
"grad_norm": 0.623515938924003, |
|
"learning_rate": 5.896607431340873e-06, |
|
"loss": 0.927, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.05986086393787413, |
|
"grad_norm": 0.595326688964164, |
|
"learning_rate": 5.977382875605816e-06, |
|
"loss": 0.941, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.060669794531629184, |
|
"grad_norm": 0.7035346939039838, |
|
"learning_rate": 6.058158319870759e-06, |
|
"loss": 0.9243, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.06147872512538424, |
|
"grad_norm": 0.6486917875962197, |
|
"learning_rate": 6.138933764135703e-06, |
|
"loss": 0.945, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.0622876557191393, |
|
"grad_norm": 0.6272725751694639, |
|
"learning_rate": 6.219709208400647e-06, |
|
"loss": 0.9333, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.06309658631289436, |
|
"grad_norm": 0.6847897267332435, |
|
"learning_rate": 6.300484652665591e-06, |
|
"loss": 0.9201, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.0639055169066494, |
|
"grad_norm": 0.6063331771589294, |
|
"learning_rate": 6.381260096930534e-06, |
|
"loss": 0.9301, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.06471444750040446, |
|
"grad_norm": 0.6535104175430165, |
|
"learning_rate": 6.462035541195477e-06, |
|
"loss": 0.9268, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.06552337809415952, |
|
"grad_norm": 0.6298942805679355, |
|
"learning_rate": 6.542810985460421e-06, |
|
"loss": 0.9126, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.06633230868791458, |
|
"grad_norm": 0.687088041257841, |
|
"learning_rate": 6.623586429725364e-06, |
|
"loss": 0.8997, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.06714123928166964, |
|
"grad_norm": 0.6809022568244187, |
|
"learning_rate": 6.7043618739903075e-06, |
|
"loss": 0.9221, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.06795016987542468, |
|
"grad_norm": 0.6595953972527364, |
|
"learning_rate": 6.7851373182552505e-06, |
|
"loss": 0.9351, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.06875910046917974, |
|
"grad_norm": 0.6526975586915988, |
|
"learning_rate": 6.865912762520195e-06, |
|
"loss": 0.9046, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.0695680310629348, |
|
"grad_norm": 0.6747286601047916, |
|
"learning_rate": 6.946688206785138e-06, |
|
"loss": 0.9384, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.07037696165668986, |
|
"grad_norm": 0.705527948534626, |
|
"learning_rate": 7.027463651050081e-06, |
|
"loss": 0.9171, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.07118589225044492, |
|
"grad_norm": 0.6636402242054834, |
|
"learning_rate": 7.108239095315025e-06, |
|
"loss": 0.9311, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.07199482284419997, |
|
"grad_norm": 0.6340851719954015, |
|
"learning_rate": 7.189014539579968e-06, |
|
"loss": 0.9194, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.07280375343795502, |
|
"grad_norm": 0.7041525654790047, |
|
"learning_rate": 7.269789983844911e-06, |
|
"loss": 0.906, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.07361268403171008, |
|
"grad_norm": 1.0272938347497549, |
|
"learning_rate": 7.350565428109855e-06, |
|
"loss": 0.9156, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.07442161462546514, |
|
"grad_norm": 0.7023049681523049, |
|
"learning_rate": 7.431340872374798e-06, |
|
"loss": 0.8947, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.0752305452192202, |
|
"grad_norm": 0.6957484415499231, |
|
"learning_rate": 7.512116316639743e-06, |
|
"loss": 0.9088, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.07603947581297525, |
|
"grad_norm": 0.6699201382717078, |
|
"learning_rate": 7.592891760904686e-06, |
|
"loss": 0.9086, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.0768484064067303, |
|
"grad_norm": 0.7101796332013423, |
|
"learning_rate": 7.673667205169629e-06, |
|
"loss": 0.9037, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.07765733700048535, |
|
"grad_norm": 0.7497394327378185, |
|
"learning_rate": 7.754442649434572e-06, |
|
"loss": 0.929, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.07846626759424041, |
|
"grad_norm": 0.7422612177167345, |
|
"learning_rate": 7.835218093699516e-06, |
|
"loss": 0.8977, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.07927519818799547, |
|
"grad_norm": 0.7103824932867557, |
|
"learning_rate": 7.91599353796446e-06, |
|
"loss": 0.9151, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.08008412878175053, |
|
"grad_norm": 0.7605669414288259, |
|
"learning_rate": 7.996768982229403e-06, |
|
"loss": 0.9234, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.08089305937550559, |
|
"grad_norm": 0.7765227746326537, |
|
"learning_rate": 8.077544426494346e-06, |
|
"loss": 0.9317, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.08170198996926063, |
|
"grad_norm": 0.7114224103403521, |
|
"learning_rate": 8.15831987075929e-06, |
|
"loss": 0.8977, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.08251092056301569, |
|
"grad_norm": 0.709653797696482, |
|
"learning_rate": 8.239095315024233e-06, |
|
"loss": 0.9048, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.08331985115677075, |
|
"grad_norm": 0.6694669312530299, |
|
"learning_rate": 8.319870759289176e-06, |
|
"loss": 0.9079, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.08412878175052581, |
|
"grad_norm": 0.6918229992578886, |
|
"learning_rate": 8.400646203554121e-06, |
|
"loss": 0.9161, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.08493771234428087, |
|
"grad_norm": 0.826690733304416, |
|
"learning_rate": 8.481421647819064e-06, |
|
"loss": 0.9179, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.08574664293803591, |
|
"grad_norm": 0.8502072042848492, |
|
"learning_rate": 8.562197092084007e-06, |
|
"loss": 0.8922, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.08655557353179097, |
|
"grad_norm": 0.7888391565118393, |
|
"learning_rate": 8.64297253634895e-06, |
|
"loss": 0.9087, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.08736450412554603, |
|
"grad_norm": 0.7612155866818295, |
|
"learning_rate": 8.723747980613893e-06, |
|
"loss": 0.9067, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.08817343471930109, |
|
"grad_norm": 0.7340496045333647, |
|
"learning_rate": 8.804523424878838e-06, |
|
"loss": 0.9122, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.08898236531305614, |
|
"grad_norm": 0.6807951345118415, |
|
"learning_rate": 8.88529886914378e-06, |
|
"loss": 0.9209, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.08979129590681119, |
|
"grad_norm": 0.7182636447883889, |
|
"learning_rate": 8.966074313408725e-06, |
|
"loss": 0.9036, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.09060022650056625, |
|
"grad_norm": 0.6906380745424776, |
|
"learning_rate": 9.046849757673668e-06, |
|
"loss": 0.8775, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.0914091570943213, |
|
"grad_norm": 0.7205121865769839, |
|
"learning_rate": 9.127625201938612e-06, |
|
"loss": 0.8799, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.09221808768807636, |
|
"grad_norm": 0.699683431919587, |
|
"learning_rate": 9.208400646203555e-06, |
|
"loss": 0.8776, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.09302701828183142, |
|
"grad_norm": 0.7092973559328101, |
|
"learning_rate": 9.289176090468498e-06, |
|
"loss": 0.8976, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.09383594887558648, |
|
"grad_norm": 0.7013577925584726, |
|
"learning_rate": 9.36995153473344e-06, |
|
"loss": 0.8915, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.09464487946934153, |
|
"grad_norm": 0.7360966395662988, |
|
"learning_rate": 9.450726978998385e-06, |
|
"loss": 0.9049, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.09545381006309658, |
|
"grad_norm": 0.7793159637124253, |
|
"learning_rate": 9.531502423263328e-06, |
|
"loss": 0.8924, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.09626274065685164, |
|
"grad_norm": 0.7373798826049844, |
|
"learning_rate": 9.612277867528273e-06, |
|
"loss": 0.9125, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.0970716712506067, |
|
"grad_norm": 0.7155304623776628, |
|
"learning_rate": 9.693053311793216e-06, |
|
"loss": 0.8922, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.09788060184436176, |
|
"grad_norm": 0.722457498504299, |
|
"learning_rate": 9.773828756058159e-06, |
|
"loss": 0.9133, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.0986895324381168, |
|
"grad_norm": 0.7006683969526154, |
|
"learning_rate": 9.854604200323102e-06, |
|
"loss": 0.8794, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.09949846303187186, |
|
"grad_norm": 0.7854584772109673, |
|
"learning_rate": 9.935379644588045e-06, |
|
"loss": 0.9029, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.10030739362562692, |
|
"grad_norm": 0.7970045944993353, |
|
"learning_rate": 9.999999202413539e-06, |
|
"loss": 0.9009, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.10111632421938198, |
|
"grad_norm": 0.8001952905335415, |
|
"learning_rate": 9.999971286914108e-06, |
|
"loss": 0.8689, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.10192525481313704, |
|
"grad_norm": 0.7947330109524953, |
|
"learning_rate": 9.999903492346063e-06, |
|
"loss": 0.8836, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.1027341854068921, |
|
"grad_norm": 0.6517743774036119, |
|
"learning_rate": 9.999795819250126e-06, |
|
"loss": 0.8937, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.10354311600064714, |
|
"grad_norm": 0.7153316955972873, |
|
"learning_rate": 9.99964826848508e-06, |
|
"loss": 0.8952, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.1043520465944022, |
|
"grad_norm": 0.7470689793235693, |
|
"learning_rate": 9.99946084122777e-06, |
|
"loss": 0.905, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.10516097718815726, |
|
"grad_norm": 0.8088466120200852, |
|
"learning_rate": 9.99923353897309e-06, |
|
"loss": 0.8982, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.10596990778191231, |
|
"grad_norm": 0.6912247019886972, |
|
"learning_rate": 9.998966363533972e-06, |
|
"loss": 0.8895, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.10677883837566737, |
|
"grad_norm": 0.7942586835594388, |
|
"learning_rate": 9.998659317041367e-06, |
|
"loss": 0.8913, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.10758776896942242, |
|
"grad_norm": 0.7841096349195548, |
|
"learning_rate": 9.998312401944237e-06, |
|
"loss": 0.8774, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.10839669956317748, |
|
"grad_norm": 0.7485526011092998, |
|
"learning_rate": 9.997925621009527e-06, |
|
"loss": 0.8955, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.10920563015693253, |
|
"grad_norm": 0.7271239921076256, |
|
"learning_rate": 9.997498977322146e-06, |
|
"loss": 0.8871, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.11001456075068759, |
|
"grad_norm": 0.7148944860864371, |
|
"learning_rate": 9.997032474284949e-06, |
|
"loss": 0.9008, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.11082349134444265, |
|
"grad_norm": 0.7486275146950349, |
|
"learning_rate": 9.996526115618694e-06, |
|
"loss": 0.8786, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.11163242193819771, |
|
"grad_norm": 0.7878464787195636, |
|
"learning_rate": 9.995979905362028e-06, |
|
"loss": 0.8805, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.11244135253195275, |
|
"grad_norm": 0.7844738669203999, |
|
"learning_rate": 9.995393847871446e-06, |
|
"loss": 0.8768, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.11325028312570781, |
|
"grad_norm": 0.7008587939833357, |
|
"learning_rate": 9.994767947821261e-06, |
|
"loss": 0.8848, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.11405921371946287, |
|
"grad_norm": 0.7176705233353148, |
|
"learning_rate": 9.994102210203567e-06, |
|
"loss": 0.9161, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.11486814431321793, |
|
"grad_norm": 0.7984767869907693, |
|
"learning_rate": 9.993396640328191e-06, |
|
"loss": 0.8984, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.11567707490697299, |
|
"grad_norm": 0.7660329081563221, |
|
"learning_rate": 9.992651243822658e-06, |
|
"loss": 0.8918, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.11648600550072803, |
|
"grad_norm": 0.7692169317757526, |
|
"learning_rate": 9.991866026632146e-06, |
|
"loss": 0.8845, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.11729493609448309, |
|
"grad_norm": 0.7728932225344195, |
|
"learning_rate": 9.991040995019441e-06, |
|
"loss": 0.8944, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.11810386668823815, |
|
"grad_norm": 0.7481133899054797, |
|
"learning_rate": 9.990176155564874e-06, |
|
"loss": 0.8782, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.11891279728199321, |
|
"grad_norm": 0.7221002980302099, |
|
"learning_rate": 9.989271515166287e-06, |
|
"loss": 0.8762, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.11972172787574827, |
|
"grad_norm": 0.7251326764742602, |
|
"learning_rate": 9.988327081038962e-06, |
|
"loss": 0.865, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.12053065846950331, |
|
"grad_norm": 0.70263798156615, |
|
"learning_rate": 9.987342860715575e-06, |
|
"loss": 0.8885, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.12133958906325837, |
|
"grad_norm": 0.7527476269587685, |
|
"learning_rate": 9.986318862046129e-06, |
|
"loss": 0.9032, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.12214851965701343, |
|
"grad_norm": 0.7398355773627006, |
|
"learning_rate": 9.98525509319789e-06, |
|
"loss": 0.8788, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.12295745025076849, |
|
"grad_norm": 0.7459365223897065, |
|
"learning_rate": 9.984151562655333e-06, |
|
"loss": 0.8754, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.12376638084452354, |
|
"grad_norm": 0.7806602396452277, |
|
"learning_rate": 9.983008279220062e-06, |
|
"loss": 0.8716, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.1245753114382786, |
|
"grad_norm": 0.7479182177455358, |
|
"learning_rate": 9.981825252010743e-06, |
|
"loss": 0.8849, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.12538424203203366, |
|
"grad_norm": 0.7644325690144114, |
|
"learning_rate": 9.980602490463037e-06, |
|
"loss": 0.8742, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.12619317262578872, |
|
"grad_norm": 0.8269622902407507, |
|
"learning_rate": 9.979340004329516e-06, |
|
"loss": 0.8747, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.12700210321954375, |
|
"grad_norm": 0.709710236101346, |
|
"learning_rate": 9.978037803679595e-06, |
|
"loss": 0.8929, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.1278110338132988, |
|
"grad_norm": 0.7487057568451865, |
|
"learning_rate": 9.97669589889944e-06, |
|
"loss": 0.8944, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.12861996440705387, |
|
"grad_norm": 0.8332024529186928, |
|
"learning_rate": 9.975314300691898e-06, |
|
"loss": 0.8803, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.12942889500080892, |
|
"grad_norm": 0.7617229913207306, |
|
"learning_rate": 9.973893020076402e-06, |
|
"loss": 0.8764, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.13023782559456398, |
|
"grad_norm": 0.7990666084729844, |
|
"learning_rate": 9.972432068388885e-06, |
|
"loss": 0.8841, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.13104675618831904, |
|
"grad_norm": 0.8144767375507436, |
|
"learning_rate": 9.970931457281693e-06, |
|
"loss": 0.8751, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.1318556867820741, |
|
"grad_norm": 0.8471484705899229, |
|
"learning_rate": 9.96939119872349e-06, |
|
"loss": 0.878, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.13266461737582916, |
|
"grad_norm": 0.7419857259389397, |
|
"learning_rate": 9.96781130499916e-06, |
|
"loss": 0.8593, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.13347354796958422, |
|
"grad_norm": 0.6966615337982099, |
|
"learning_rate": 9.966191788709716e-06, |
|
"loss": 0.8754, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.13428247856333927, |
|
"grad_norm": 0.7552286075445729, |
|
"learning_rate": 9.96453266277219e-06, |
|
"loss": 0.8729, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.13509140915709433, |
|
"grad_norm": 0.8299494943538954, |
|
"learning_rate": 9.96283394041954e-06, |
|
"loss": 0.881, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.13590033975084936, |
|
"grad_norm": 0.734889508057368, |
|
"learning_rate": 9.961095635200536e-06, |
|
"loss": 0.8642, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.13670927034460442, |
|
"grad_norm": 0.7083874584173164, |
|
"learning_rate": 9.959317760979654e-06, |
|
"loss": 0.87, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.13751820093835948, |
|
"grad_norm": 0.6795303692054026, |
|
"learning_rate": 9.957500331936971e-06, |
|
"loss": 0.8916, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.13832713153211454, |
|
"grad_norm": 0.7711107263091647, |
|
"learning_rate": 9.955643362568048e-06, |
|
"loss": 0.8867, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.1391360621258696, |
|
"grad_norm": 0.7822464459121171, |
|
"learning_rate": 9.953746867683807e-06, |
|
"loss": 0.875, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.13994499271962466, |
|
"grad_norm": 0.8269771768105255, |
|
"learning_rate": 9.951810862410426e-06, |
|
"loss": 0.8762, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.14075392331337971, |
|
"grad_norm": 0.766471476016186, |
|
"learning_rate": 9.949835362189215e-06, |
|
"loss": 0.8635, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.14156285390713477, |
|
"grad_norm": 0.8196823476460824, |
|
"learning_rate": 9.947820382776482e-06, |
|
"loss": 0.8933, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.14237178450088983, |
|
"grad_norm": 0.7502232673514019, |
|
"learning_rate": 9.945765940243422e-06, |
|
"loss": 0.9162, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.1431807150946449, |
|
"grad_norm": 0.8243799834314466, |
|
"learning_rate": 9.943672050975979e-06, |
|
"loss": 0.8973, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.14398964568839995, |
|
"grad_norm": 0.7327053926662375, |
|
"learning_rate": 9.94153873167472e-06, |
|
"loss": 0.8754, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.14479857628215498, |
|
"grad_norm": 0.7710933741203212, |
|
"learning_rate": 9.9393659993547e-06, |
|
"loss": 0.8989, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.14560750687591004, |
|
"grad_norm": 0.6969476376928585, |
|
"learning_rate": 9.937153871345326e-06, |
|
"loss": 0.8851, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.1464164374696651, |
|
"grad_norm": 0.7537226761385668, |
|
"learning_rate": 9.934902365290222e-06, |
|
"loss": 0.8678, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.14722536806342015, |
|
"grad_norm": 0.7880392328433623, |
|
"learning_rate": 9.932611499147082e-06, |
|
"loss": 0.8785, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.1480342986571752, |
|
"grad_norm": 0.7979672909632645, |
|
"learning_rate": 9.930281291187534e-06, |
|
"loss": 0.9031, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.14884322925093027, |
|
"grad_norm": 0.769227717875029, |
|
"learning_rate": 9.927911759996989e-06, |
|
"loss": 0.8752, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.14965215984468533, |
|
"grad_norm": 0.7501410474884525, |
|
"learning_rate": 9.925502924474495e-06, |
|
"loss": 0.8818, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.1504610904384404, |
|
"grad_norm": 0.7900709039002747, |
|
"learning_rate": 9.923054803832585e-06, |
|
"loss": 0.8549, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.15127002103219545, |
|
"grad_norm": 0.727528278106019, |
|
"learning_rate": 9.920567417597127e-06, |
|
"loss": 0.869, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.1520789516259505, |
|
"grad_norm": 0.7478261446374327, |
|
"learning_rate": 9.918040785607163e-06, |
|
"loss": 0.8803, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.15288788221970556, |
|
"grad_norm": 0.7798141297679039, |
|
"learning_rate": 9.915474928014754e-06, |
|
"loss": 0.8601, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.1536968128134606, |
|
"grad_norm": 0.743300400063168, |
|
"learning_rate": 9.912869865284821e-06, |
|
"loss": 0.873, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.15450574340721565, |
|
"grad_norm": 0.7855652729779465, |
|
"learning_rate": 9.91022561819498e-06, |
|
"loss": 0.8556, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.1553146740009707, |
|
"grad_norm": 0.7512935005312259, |
|
"learning_rate": 9.90754220783537e-06, |
|
"loss": 0.8761, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.15612360459472577, |
|
"grad_norm": 0.7480633701848225, |
|
"learning_rate": 9.9048196556085e-06, |
|
"loss": 0.884, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.15693253518848083, |
|
"grad_norm": 0.7649617436679113, |
|
"learning_rate": 9.902057983229059e-06, |
|
"loss": 0.8558, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.15774146578223588, |
|
"grad_norm": 0.7423506733665634, |
|
"learning_rate": 9.89925721272376e-06, |
|
"loss": 0.8773, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.15855039637599094, |
|
"grad_norm": 0.7931165102529991, |
|
"learning_rate": 9.89641736643116e-06, |
|
"loss": 0.8846, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.159359326969746, |
|
"grad_norm": 0.7550133131247714, |
|
"learning_rate": 9.893538467001466e-06, |
|
"loss": 0.8727, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.16016825756350106, |
|
"grad_norm": 0.773118150891364, |
|
"learning_rate": 9.89062053739638e-06, |
|
"loss": 0.8808, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.16097718815725612, |
|
"grad_norm": 0.75632598887103, |
|
"learning_rate": 9.887663600888897e-06, |
|
"loss": 0.8713, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.16178611875101118, |
|
"grad_norm": 0.8316783703999336, |
|
"learning_rate": 9.88466768106313e-06, |
|
"loss": 0.8571, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.1625950493447662, |
|
"grad_norm": 0.7454307910719469, |
|
"learning_rate": 9.881632801814112e-06, |
|
"loss": 0.8602, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.16340397993852127, |
|
"grad_norm": 0.7564730753355293, |
|
"learning_rate": 9.878558987347613e-06, |
|
"loss": 0.8694, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.16421291053227632, |
|
"grad_norm": 0.7924034403437432, |
|
"learning_rate": 9.875446262179948e-06, |
|
"loss": 0.8695, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.16502184112603138, |
|
"grad_norm": 0.809312814940399, |
|
"learning_rate": 9.872294651137773e-06, |
|
"loss": 0.8639, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.16583077171978644, |
|
"grad_norm": 0.8496318363889708, |
|
"learning_rate": 9.869104179357898e-06, |
|
"loss": 0.8656, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.1666397023135415, |
|
"grad_norm": 0.8463627183621499, |
|
"learning_rate": 9.865874872287076e-06, |
|
"loss": 0.8721, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.16744863290729656, |
|
"grad_norm": 0.7538421574954219, |
|
"learning_rate": 9.862606755681805e-06, |
|
"loss": 0.887, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.16825756350105162, |
|
"grad_norm": 0.8015977734669262, |
|
"learning_rate": 9.859299855608127e-06, |
|
"loss": 0.8652, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.16906649409480667, |
|
"grad_norm": 0.7608756372365878, |
|
"learning_rate": 9.855954198441411e-06, |
|
"loss": 0.8882, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.16987542468856173, |
|
"grad_norm": 0.8188556394828449, |
|
"learning_rate": 9.852569810866148e-06, |
|
"loss": 0.8694, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.1706843552823168, |
|
"grad_norm": 0.805253431195849, |
|
"learning_rate": 9.849146719875737e-06, |
|
"loss": 0.8754, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.17149328587607182, |
|
"grad_norm": 0.7590085927711259, |
|
"learning_rate": 9.845684952772274e-06, |
|
"loss": 0.8741, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.17230221646982688, |
|
"grad_norm": 0.7493004472784411, |
|
"learning_rate": 9.842184537166326e-06, |
|
"loss": 0.8784, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.17311114706358194, |
|
"grad_norm": 0.8005839214503973, |
|
"learning_rate": 9.838645500976716e-06, |
|
"loss": 0.8668, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.173920077657337, |
|
"grad_norm": 0.7331632983706363, |
|
"learning_rate": 9.835067872430297e-06, |
|
"loss": 0.8577, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.17472900825109206, |
|
"grad_norm": 0.8212489518002695, |
|
"learning_rate": 9.831451680061735e-06, |
|
"loss": 0.8503, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.1755379388448471, |
|
"grad_norm": 0.7423733165945756, |
|
"learning_rate": 9.82779695271327e-06, |
|
"loss": 0.849, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.17634686943860217, |
|
"grad_norm": 0.8436334149956825, |
|
"learning_rate": 9.824103719534497e-06, |
|
"loss": 0.8739, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.17715580003235723, |
|
"grad_norm": 0.8371560523056507, |
|
"learning_rate": 9.820372009982122e-06, |
|
"loss": 0.8885, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.1779647306261123, |
|
"grad_norm": 0.8549576764605578, |
|
"learning_rate": 9.816601853819739e-06, |
|
"loss": 0.8905, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.17877366121986735, |
|
"grad_norm": 0.8302136713634986, |
|
"learning_rate": 9.81279328111758e-06, |
|
"loss": 0.863, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.17958259181362238, |
|
"grad_norm": 0.7692548326236497, |
|
"learning_rate": 9.80894632225229e-06, |
|
"loss": 0.8942, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.18039152240737744, |
|
"grad_norm": 0.8190077972884791, |
|
"learning_rate": 9.805061007906668e-06, |
|
"loss": 0.8721, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.1812004530011325, |
|
"grad_norm": 0.8021866541591391, |
|
"learning_rate": 9.801137369069441e-06, |
|
"loss": 0.8587, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.18200938359488755, |
|
"grad_norm": 0.7295302337882422, |
|
"learning_rate": 9.797175437034997e-06, |
|
"loss": 0.8809, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.1828183141886426, |
|
"grad_norm": 0.7303087771888767, |
|
"learning_rate": 9.79317524340315e-06, |
|
"loss": 0.864, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.18362724478239767, |
|
"grad_norm": 0.7668374049535479, |
|
"learning_rate": 9.789136820078884e-06, |
|
"loss": 0.8768, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.18443617537615273, |
|
"grad_norm": 0.7779749351860815, |
|
"learning_rate": 9.785060199272096e-06, |
|
"loss": 0.8871, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.1852451059699078, |
|
"grad_norm": 0.7605751732056839, |
|
"learning_rate": 9.780945413497339e-06, |
|
"loss": 0.8756, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.18605403656366284, |
|
"grad_norm": 0.7608509607177488, |
|
"learning_rate": 9.776792495573567e-06, |
|
"loss": 0.8651, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.1868629671574179, |
|
"grad_norm": 0.7649020527995475, |
|
"learning_rate": 9.772601478623871e-06, |
|
"loss": 0.8522, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.18767189775117296, |
|
"grad_norm": 0.8284338499023652, |
|
"learning_rate": 9.768372396075213e-06, |
|
"loss": 0.8986, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.188480828344928, |
|
"grad_norm": 0.7588241076036676, |
|
"learning_rate": 9.764105281658161e-06, |
|
"loss": 0.8727, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.18928975893868305, |
|
"grad_norm": 0.7815552643338736, |
|
"learning_rate": 9.759800169406621e-06, |
|
"loss": 0.8639, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.1900986895324381, |
|
"grad_norm": 0.7935036774949533, |
|
"learning_rate": 9.755457093657562e-06, |
|
"loss": 0.8651, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.19090762012619317, |
|
"grad_norm": 0.7411492264667076, |
|
"learning_rate": 9.751076089050747e-06, |
|
"loss": 0.8675, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.19171655071994823, |
|
"grad_norm": 0.7597654203339239, |
|
"learning_rate": 9.746657190528454e-06, |
|
"loss": 0.8421, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.19252548131370328, |
|
"grad_norm": 0.7871189163253436, |
|
"learning_rate": 9.742200433335196e-06, |
|
"loss": 0.877, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.19333441190745834, |
|
"grad_norm": 0.7773214820602937, |
|
"learning_rate": 9.737705853017442e-06, |
|
"loss": 0.8721, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.1941433425012134, |
|
"grad_norm": 0.8230817780724599, |
|
"learning_rate": 9.733173485423333e-06, |
|
"loss": 0.8714, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.19495227309496846, |
|
"grad_norm": 0.7681885249706483, |
|
"learning_rate": 9.7286033667024e-06, |
|
"loss": 0.8594, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.19576120368872352, |
|
"grad_norm": 0.7640895548306019, |
|
"learning_rate": 9.723995533305262e-06, |
|
"loss": 0.8621, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.19657013428247858, |
|
"grad_norm": 0.876302015000407, |
|
"learning_rate": 9.719350021983356e-06, |
|
"loss": 0.873, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.1973790648762336, |
|
"grad_norm": 0.7687883497172476, |
|
"learning_rate": 9.714666869788622e-06, |
|
"loss": 0.8837, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.19818799546998866, |
|
"grad_norm": 0.7918479859070253, |
|
"learning_rate": 9.709946114073231e-06, |
|
"loss": 0.8742, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.19899692606374372, |
|
"grad_norm": 0.7704108440790387, |
|
"learning_rate": 9.705187792489263e-06, |
|
"loss": 0.8831, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.19980585665749878, |
|
"grad_norm": 0.7958664141923848, |
|
"learning_rate": 9.700391942988422e-06, |
|
"loss": 0.8815, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.20061478725125384, |
|
"grad_norm": 0.7299908461383041, |
|
"learning_rate": 9.695558603821735e-06, |
|
"loss": 0.8705, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.2014237178450089, |
|
"grad_norm": 0.7921888129305676, |
|
"learning_rate": 9.69068781353923e-06, |
|
"loss": 0.8651, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.20223264843876396, |
|
"grad_norm": 0.783261666055412, |
|
"learning_rate": 9.68577961098965e-06, |
|
"loss": 0.8657, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.20304157903251902, |
|
"grad_norm": 0.7640970071314145, |
|
"learning_rate": 9.680834035320127e-06, |
|
"loss": 0.8613, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.20385050962627407, |
|
"grad_norm": 0.7480591135260592, |
|
"learning_rate": 9.675851125975879e-06, |
|
"loss": 0.8522, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.20465944022002913, |
|
"grad_norm": 0.764427552884641, |
|
"learning_rate": 9.67083092269989e-06, |
|
"loss": 0.8846, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.2054683708137842, |
|
"grad_norm": 0.7562918652667725, |
|
"learning_rate": 9.665773465532597e-06, |
|
"loss": 0.8538, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.20627730140753922, |
|
"grad_norm": 0.7308192790835369, |
|
"learning_rate": 9.660678794811569e-06, |
|
"loss": 0.8758, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.20708623200129428, |
|
"grad_norm": 0.7327237540533151, |
|
"learning_rate": 9.65554695117118e-06, |
|
"loss": 0.8556, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.20789516259504934, |
|
"grad_norm": 0.8111511896192851, |
|
"learning_rate": 9.650377975542298e-06, |
|
"loss": 0.8845, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.2087040931888044, |
|
"grad_norm": 0.7725083000676402, |
|
"learning_rate": 9.645171909151944e-06, |
|
"loss": 0.8713, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.20951302378255945, |
|
"grad_norm": 0.8087330967911632, |
|
"learning_rate": 9.639928793522976e-06, |
|
"loss": 0.8807, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.2103219543763145, |
|
"grad_norm": 0.7987317654845778, |
|
"learning_rate": 9.634648670473743e-06, |
|
"loss": 0.8887, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.21113088497006957, |
|
"grad_norm": 0.7222991798791983, |
|
"learning_rate": 9.629331582117766e-06, |
|
"loss": 0.8617, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.21193981556382463, |
|
"grad_norm": 0.8205396256241525, |
|
"learning_rate": 9.623977570863398e-06, |
|
"loss": 0.8512, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.2127487461575797, |
|
"grad_norm": 0.7704920928191922, |
|
"learning_rate": 9.618586679413477e-06, |
|
"loss": 0.8545, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.21355767675133475, |
|
"grad_norm": 0.8121022691153029, |
|
"learning_rate": 9.613158950764996e-06, |
|
"loss": 0.8647, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.2143666073450898, |
|
"grad_norm": 0.8068844595908223, |
|
"learning_rate": 9.60769442820876e-06, |
|
"loss": 0.8511, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.21517553793884484, |
|
"grad_norm": 0.779446090355673, |
|
"learning_rate": 9.602193155329029e-06, |
|
"loss": 0.8623, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.2159844685325999, |
|
"grad_norm": 0.7763229071537453, |
|
"learning_rate": 9.596655176003185e-06, |
|
"loss": 0.8662, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.21679339912635495, |
|
"grad_norm": 0.7877601933457381, |
|
"learning_rate": 9.591080534401371e-06, |
|
"loss": 0.87, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.21760232972011, |
|
"grad_norm": 0.7986846021359395, |
|
"learning_rate": 9.585469274986148e-06, |
|
"loss": 0.8716, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.21841126031386507, |
|
"grad_norm": 0.7790084651864484, |
|
"learning_rate": 9.579821442512131e-06, |
|
"loss": 0.8595, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.21922019090762013, |
|
"grad_norm": 0.7565290946235815, |
|
"learning_rate": 9.574137082025639e-06, |
|
"loss": 0.878, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.22002912150137519, |
|
"grad_norm": 0.7534079459249721, |
|
"learning_rate": 9.568416238864335e-06, |
|
"loss": 0.8542, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.22083805209513024, |
|
"grad_norm": 0.7934001624755948, |
|
"learning_rate": 9.562658958656856e-06, |
|
"loss": 0.8662, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.2216469826888853, |
|
"grad_norm": 0.7851224446308791, |
|
"learning_rate": 9.556865287322464e-06, |
|
"loss": 0.8708, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.22245591328264036, |
|
"grad_norm": 0.7292260657124646, |
|
"learning_rate": 9.551035271070665e-06, |
|
"loss": 0.8496, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.22326484387639542, |
|
"grad_norm": 0.8178256842698678, |
|
"learning_rate": 9.54516895640085e-06, |
|
"loss": 0.8615, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.22407377447015045, |
|
"grad_norm": 0.7761082648853441, |
|
"learning_rate": 9.539266390101922e-06, |
|
"loss": 0.8596, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.2248827050639055, |
|
"grad_norm": 0.7510699850597663, |
|
"learning_rate": 9.533327619251921e-06, |
|
"loss": 0.8591, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.22569163565766057, |
|
"grad_norm": 0.7901677982076935, |
|
"learning_rate": 9.527352691217649e-06, |
|
"loss": 0.856, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.22650056625141562, |
|
"grad_norm": 0.8121283027645855, |
|
"learning_rate": 9.52134165365429e-06, |
|
"loss": 0.8686, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.22730949684517068, |
|
"grad_norm": 0.8027876769163085, |
|
"learning_rate": 9.515294554505039e-06, |
|
"loss": 0.8609, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.22811842743892574, |
|
"grad_norm": 0.7677214802906873, |
|
"learning_rate": 9.509211442000705e-06, |
|
"loss": 0.8842, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.2289273580326808, |
|
"grad_norm": 0.7935787903248027, |
|
"learning_rate": 9.503092364659343e-06, |
|
"loss": 0.8571, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.22973628862643586, |
|
"grad_norm": 0.7885818327639111, |
|
"learning_rate": 9.496937371285852e-06, |
|
"loss": 0.8368, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.23054521922019092, |
|
"grad_norm": 0.7664542560047006, |
|
"learning_rate": 9.490746510971595e-06, |
|
"loss": 0.8586, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.23135414981394598, |
|
"grad_norm": 0.7915598638517407, |
|
"learning_rate": 9.484519833094006e-06, |
|
"loss": 0.8594, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.232163080407701, |
|
"grad_norm": 0.8224568006509917, |
|
"learning_rate": 9.478257387316189e-06, |
|
"loss": 0.8604, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.23297201100145606, |
|
"grad_norm": 0.7646329363431662, |
|
"learning_rate": 9.471959223586535e-06, |
|
"loss": 0.8667, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.23378094159521112, |
|
"grad_norm": 0.8054133554377607, |
|
"learning_rate": 9.465625392138314e-06, |
|
"loss": 0.8491, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.23458987218896618, |
|
"grad_norm": 0.7999914607438723, |
|
"learning_rate": 9.459255943489271e-06, |
|
"loss": 0.8787, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.23539880278272124, |
|
"grad_norm": 0.7770791430956507, |
|
"learning_rate": 9.45285092844124e-06, |
|
"loss": 0.8526, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.2362077333764763, |
|
"grad_norm": 0.817105925992962, |
|
"learning_rate": 9.446410398079716e-06, |
|
"loss": 0.8605, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.23701666397023136, |
|
"grad_norm": 0.7238122855218315, |
|
"learning_rate": 9.439934403773468e-06, |
|
"loss": 0.8727, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.23782559456398641, |
|
"grad_norm": 0.7489559356823628, |
|
"learning_rate": 9.433422997174113e-06, |
|
"loss": 0.8672, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.23863452515774147, |
|
"grad_norm": 0.8073337490099621, |
|
"learning_rate": 9.42687623021572e-06, |
|
"loss": 0.8631, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.23944345575149653, |
|
"grad_norm": 0.7504637131946496, |
|
"learning_rate": 9.42029415511438e-06, |
|
"loss": 0.85, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.2402523863452516, |
|
"grad_norm": 0.7888833671784519, |
|
"learning_rate": 9.4136768243678e-06, |
|
"loss": 0.8602, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.24106131693900662, |
|
"grad_norm": 0.8101599644677917, |
|
"learning_rate": 9.40702429075488e-06, |
|
"loss": 0.8535, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.24187024753276168, |
|
"grad_norm": 0.7871350703464342, |
|
"learning_rate": 9.400336607335294e-06, |
|
"loss": 0.8486, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.24267917812651674, |
|
"grad_norm": 0.7878053882806549, |
|
"learning_rate": 9.393613827449064e-06, |
|
"loss": 0.8537, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.2434881087202718, |
|
"grad_norm": 0.7716899022882726, |
|
"learning_rate": 9.38685600471614e-06, |
|
"loss": 0.8751, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.24429703931402685, |
|
"grad_norm": 0.7804644122014117, |
|
"learning_rate": 9.380063193035968e-06, |
|
"loss": 0.8496, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.2451059699077819, |
|
"grad_norm": 0.8146517496395704, |
|
"learning_rate": 9.373235446587055e-06, |
|
"loss": 0.8678, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.24591490050153697, |
|
"grad_norm": 0.7819896556165521, |
|
"learning_rate": 9.366372819826553e-06, |
|
"loss": 0.8572, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.24672383109529203, |
|
"grad_norm": 0.7667264574062115, |
|
"learning_rate": 9.359475367489805e-06, |
|
"loss": 0.8627, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.2475327616890471, |
|
"grad_norm": 0.8236534571118577, |
|
"learning_rate": 9.352543144589923e-06, |
|
"loss": 0.8668, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.24834169228280215, |
|
"grad_norm": 0.7670049125342286, |
|
"learning_rate": 9.345576206417345e-06, |
|
"loss": 0.8418, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.2491506228765572, |
|
"grad_norm": 0.7827102030650855, |
|
"learning_rate": 9.338574608539389e-06, |
|
"loss": 0.8439, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.24995955347031223, |
|
"grad_norm": 0.7839079136756119, |
|
"learning_rate": 9.331538406799815e-06, |
|
"loss": 0.8549, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.2507684840640673, |
|
"grad_norm": 0.8178409940371506, |
|
"learning_rate": 9.324467657318384e-06, |
|
"loss": 0.8534, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.25157741465782235, |
|
"grad_norm": 0.8099902787097547, |
|
"learning_rate": 9.317362416490396e-06, |
|
"loss": 0.8589, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.25238634525157744, |
|
"grad_norm": 0.8397966939088385, |
|
"learning_rate": 9.310222740986258e-06, |
|
"loss": 0.8695, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.25319527584533247, |
|
"grad_norm": 0.8012721851788485, |
|
"learning_rate": 9.303048687751016e-06, |
|
"loss": 0.8647, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.2540042064390875, |
|
"grad_norm": 0.7848654081297167, |
|
"learning_rate": 9.29584031400391e-06, |
|
"loss": 0.848, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.2548131370328426, |
|
"grad_norm": 0.7276468748203139, |
|
"learning_rate": 9.288597677237918e-06, |
|
"loss": 0.8451, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.2556220676265976, |
|
"grad_norm": 0.8006865536918176, |
|
"learning_rate": 9.281320835219294e-06, |
|
"loss": 0.847, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.2564309982203527, |
|
"grad_norm": 0.7628111941438427, |
|
"learning_rate": 9.274009845987106e-06, |
|
"loss": 0.8712, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.25723992881410773, |
|
"grad_norm": 0.7881647592364686, |
|
"learning_rate": 9.26666476785278e-06, |
|
"loss": 0.8678, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.2580488594078628, |
|
"grad_norm": 0.8116538712282235, |
|
"learning_rate": 9.259285659399624e-06, |
|
"loss": 0.8535, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.25885779000161785, |
|
"grad_norm": 0.7813279263856877, |
|
"learning_rate": 9.251872579482373e-06, |
|
"loss": 0.845, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.25966672059537294, |
|
"grad_norm": 0.7940356068886795, |
|
"learning_rate": 9.244425587226708e-06, |
|
"loss": 0.8492, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.26047565118912797, |
|
"grad_norm": 0.8041883146042144, |
|
"learning_rate": 9.236944742028797e-06, |
|
"loss": 0.869, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.26128458178288305, |
|
"grad_norm": 0.8411011003169626, |
|
"learning_rate": 9.229430103554808e-06, |
|
"loss": 0.8702, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.2620935123766381, |
|
"grad_norm": 0.7994752148577935, |
|
"learning_rate": 9.221881731740442e-06, |
|
"loss": 0.846, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.2629024429703931, |
|
"grad_norm": 0.7904404440251053, |
|
"learning_rate": 9.214299686790453e-06, |
|
"loss": 0.8479, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.2637113735641482, |
|
"grad_norm": 0.8285353436413048, |
|
"learning_rate": 9.206684029178166e-06, |
|
"loss": 0.867, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.26452030415790323, |
|
"grad_norm": 0.8167896861662431, |
|
"learning_rate": 9.199034819644997e-06, |
|
"loss": 0.845, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.2653292347516583, |
|
"grad_norm": 0.7625190041745681, |
|
"learning_rate": 9.191352119199965e-06, |
|
"loss": 0.8715, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.26613816534541335, |
|
"grad_norm": 0.766576459090851, |
|
"learning_rate": 9.183635989119211e-06, |
|
"loss": 0.8554, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.26694709593916843, |
|
"grad_norm": 0.7598503790298559, |
|
"learning_rate": 9.175886490945505e-06, |
|
"loss": 0.8657, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.26775602653292346, |
|
"grad_norm": 0.7436415672250105, |
|
"learning_rate": 9.168103686487755e-06, |
|
"loss": 0.8618, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.26856495712667855, |
|
"grad_norm": 0.8147678677174188, |
|
"learning_rate": 9.160287637820514e-06, |
|
"loss": 0.8591, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.2693738877204336, |
|
"grad_norm": 0.7963714833005615, |
|
"learning_rate": 9.152438407283493e-06, |
|
"loss": 0.8651, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.27018281831418867, |
|
"grad_norm": 0.7937419589693009, |
|
"learning_rate": 9.144556057481048e-06, |
|
"loss": 0.8373, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.2709917489079437, |
|
"grad_norm": 0.8239141053085791, |
|
"learning_rate": 9.136640651281694e-06, |
|
"loss": 0.8712, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.2718006795016987, |
|
"grad_norm": 0.7797626292094625, |
|
"learning_rate": 9.128692251817602e-06, |
|
"loss": 0.8629, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.2726096100954538, |
|
"grad_norm": 0.7697095287067577, |
|
"learning_rate": 9.120710922484089e-06, |
|
"loss": 0.8307, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.27341854068920884, |
|
"grad_norm": 0.7710196952612974, |
|
"learning_rate": 9.112696726939112e-06, |
|
"loss": 0.8513, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.27422747128296393, |
|
"grad_norm": 0.7800639021416464, |
|
"learning_rate": 9.104649729102774e-06, |
|
"loss": 0.8582, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.27503640187671896, |
|
"grad_norm": 0.8044706731480418, |
|
"learning_rate": 9.096569993156797e-06, |
|
"loss": 0.8406, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.27584533247047405, |
|
"grad_norm": 0.7699086496292621, |
|
"learning_rate": 9.088457583544022e-06, |
|
"loss": 0.8562, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.2766542630642291, |
|
"grad_norm": 0.8340399404633732, |
|
"learning_rate": 9.080312564967884e-06, |
|
"loss": 0.8694, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.27746319365798416, |
|
"grad_norm": 0.7919025521073914, |
|
"learning_rate": 9.072135002391912e-06, |
|
"loss": 0.8769, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.2782721242517392, |
|
"grad_norm": 0.7754055380467696, |
|
"learning_rate": 9.063924961039195e-06, |
|
"loss": 0.8852, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.2790810548454943, |
|
"grad_norm": 0.76639361552101, |
|
"learning_rate": 9.055682506391866e-06, |
|
"loss": 0.8655, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.2798899854392493, |
|
"grad_norm": 0.7860193334912077, |
|
"learning_rate": 9.04740770419059e-06, |
|
"loss": 0.8572, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.28069891603300434, |
|
"grad_norm": 0.7656637491035695, |
|
"learning_rate": 9.039100620434025e-06, |
|
"loss": 0.8634, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.28150784662675943, |
|
"grad_norm": 0.8287569040947624, |
|
"learning_rate": 9.030761321378303e-06, |
|
"loss": 0.8539, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.28231677722051446, |
|
"grad_norm": 0.778379469664662, |
|
"learning_rate": 9.022389873536505e-06, |
|
"loss": 0.8543, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.28312570781426954, |
|
"grad_norm": 0.8341873388515572, |
|
"learning_rate": 9.01398634367812e-06, |
|
"loss": 0.8378, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.2839346384080246, |
|
"grad_norm": 0.807630286757475, |
|
"learning_rate": 9.005550798828521e-06, |
|
"loss": 0.8661, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.28474356900177966, |
|
"grad_norm": 0.7492013925485693, |
|
"learning_rate": 8.997083306268434e-06, |
|
"loss": 0.8646, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.2855524995955347, |
|
"grad_norm": 0.820329567734956, |
|
"learning_rate": 8.988583933533384e-06, |
|
"loss": 0.866, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.2863614301892898, |
|
"grad_norm": 0.8427359735718537, |
|
"learning_rate": 8.980052748413177e-06, |
|
"loss": 0.8393, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.2871703607830448, |
|
"grad_norm": 0.7315290649122431, |
|
"learning_rate": 8.971489818951347e-06, |
|
"loss": 0.8561, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.2879792913767999, |
|
"grad_norm": 0.7968151029522487, |
|
"learning_rate": 8.962895213444618e-06, |
|
"loss": 0.8449, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.2887882219705549, |
|
"grad_norm": 0.8497262991663913, |
|
"learning_rate": 8.954269000442353e-06, |
|
"loss": 0.8614, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.28959715256430996, |
|
"grad_norm": 0.8884396786756781, |
|
"learning_rate": 8.945611248746015e-06, |
|
"loss": 0.873, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.29040608315806504, |
|
"grad_norm": 0.7822894663944296, |
|
"learning_rate": 8.936922027408618e-06, |
|
"loss": 0.8504, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.2912150137518201, |
|
"grad_norm": 0.7512013193436944, |
|
"learning_rate": 8.928201405734172e-06, |
|
"loss": 0.8638, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.29202394434557516, |
|
"grad_norm": 0.808524098300231, |
|
"learning_rate": 8.919449453277124e-06, |
|
"loss": 0.8555, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.2928328749393302, |
|
"grad_norm": 0.8205569397216151, |
|
"learning_rate": 8.910666239841824e-06, |
|
"loss": 0.8393, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.2936418055330853, |
|
"grad_norm": 0.7685972280574218, |
|
"learning_rate": 8.901851835481947e-06, |
|
"loss": 0.8616, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.2944507361268403, |
|
"grad_norm": 0.7888992044706635, |
|
"learning_rate": 8.893006310499941e-06, |
|
"loss": 0.827, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.2952596667205954, |
|
"grad_norm": 0.8377015460290415, |
|
"learning_rate": 8.884129735446471e-06, |
|
"loss": 0.8699, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.2960685973143504, |
|
"grad_norm": 0.8237212938659825, |
|
"learning_rate": 8.875222181119859e-06, |
|
"loss": 0.8503, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.2968775279081055, |
|
"grad_norm": 0.7665032383816565, |
|
"learning_rate": 8.866283718565498e-06, |
|
"loss": 0.8589, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.29768645850186054, |
|
"grad_norm": 0.8141188029046369, |
|
"learning_rate": 8.857314419075316e-06, |
|
"loss": 0.8571, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.29849538909561557, |
|
"grad_norm": 0.7761079464199027, |
|
"learning_rate": 8.848314354187184e-06, |
|
"loss": 0.8463, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.29930431968937066, |
|
"grad_norm": 0.8097442274081034, |
|
"learning_rate": 8.839283595684355e-06, |
|
"loss": 0.8581, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.3001132502831257, |
|
"grad_norm": 0.8095252549308872, |
|
"learning_rate": 8.83022221559489e-06, |
|
"loss": 0.8557, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.3009221808768808, |
|
"grad_norm": 0.7980409130367505, |
|
"learning_rate": 8.821130286191086e-06, |
|
"loss": 0.8671, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.3017311114706358, |
|
"grad_norm": 0.8173815138889032, |
|
"learning_rate": 8.81200787998889e-06, |
|
"loss": 0.8594, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.3025400420643909, |
|
"grad_norm": 0.8047823405503607, |
|
"learning_rate": 8.802855069747338e-06, |
|
"loss": 0.8596, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.3033489726581459, |
|
"grad_norm": 0.804911943505487, |
|
"learning_rate": 8.793671928467953e-06, |
|
"loss": 0.8397, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.304157903251901, |
|
"grad_norm": 0.7613233673132125, |
|
"learning_rate": 8.784458529394185e-06, |
|
"loss": 0.8407, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.30496683384565604, |
|
"grad_norm": 0.7397028565696792, |
|
"learning_rate": 8.775214946010806e-06, |
|
"loss": 0.8476, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.3057757644394111, |
|
"grad_norm": 0.7728205441604445, |
|
"learning_rate": 8.765941252043341e-06, |
|
"loss": 0.8384, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.30658469503316615, |
|
"grad_norm": 0.8229918467819616, |
|
"learning_rate": 8.756637521457473e-06, |
|
"loss": 0.8488, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.3073936256269212, |
|
"grad_norm": 0.7848703050564307, |
|
"learning_rate": 8.747303828458446e-06, |
|
"loss": 0.8488, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.30820255622067627, |
|
"grad_norm": 0.8185735473058204, |
|
"learning_rate": 8.737940247490487e-06, |
|
"loss": 0.8447, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.3090114868144313, |
|
"grad_norm": 0.8175009758279596, |
|
"learning_rate": 8.728546853236202e-06, |
|
"loss": 0.8468, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.3098204174081864, |
|
"grad_norm": 0.8234001884738343, |
|
"learning_rate": 8.71912372061598e-06, |
|
"loss": 0.8579, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.3106293480019414, |
|
"grad_norm": 0.8135018815005348, |
|
"learning_rate": 8.70967092478741e-06, |
|
"loss": 0.8333, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.3114382785956965, |
|
"grad_norm": 0.8316517603780106, |
|
"learning_rate": 8.700188541144658e-06, |
|
"loss": 0.8152, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.31224720918945154, |
|
"grad_norm": 0.8542272626940559, |
|
"learning_rate": 8.690676645317886e-06, |
|
"loss": 0.8302, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.3130561397832066, |
|
"grad_norm": 0.7850582360070165, |
|
"learning_rate": 8.68113531317264e-06, |
|
"loss": 0.8494, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.31386507037696165, |
|
"grad_norm": 0.8046763932380032, |
|
"learning_rate": 8.671564620809243e-06, |
|
"loss": 0.8512, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.31467400097071674, |
|
"grad_norm": 0.8472122957190853, |
|
"learning_rate": 8.661964644562194e-06, |
|
"loss": 0.8481, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.31548293156447177, |
|
"grad_norm": 0.8098454336630103, |
|
"learning_rate": 8.652335460999554e-06, |
|
"loss": 0.8386, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.3162918621582268, |
|
"grad_norm": 0.81377246582292, |
|
"learning_rate": 8.64267714692234e-06, |
|
"loss": 0.8109, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.3171007927519819, |
|
"grad_norm": 0.7941773089013205, |
|
"learning_rate": 8.632989779363907e-06, |
|
"loss": 0.8223, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.3179097233457369, |
|
"grad_norm": 0.7988879204951932, |
|
"learning_rate": 8.623273435589338e-06, |
|
"loss": 0.8536, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.318718653939492, |
|
"grad_norm": 0.7645549102866793, |
|
"learning_rate": 8.613528193094826e-06, |
|
"loss": 0.841, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.31952758453324703, |
|
"grad_norm": 0.7971671512555748, |
|
"learning_rate": 8.603754129607055e-06, |
|
"loss": 0.8418, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.3203365151270021, |
|
"grad_norm": 0.7570319344787492, |
|
"learning_rate": 8.593951323082586e-06, |
|
"loss": 0.8583, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.32114544572075715, |
|
"grad_norm": 0.8210274942358864, |
|
"learning_rate": 8.584119851707224e-06, |
|
"loss": 0.8493, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.32195437631451224, |
|
"grad_norm": 0.7921398145158576, |
|
"learning_rate": 8.574259793895404e-06, |
|
"loss": 0.8373, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.32276330690826727, |
|
"grad_norm": 0.8093679320396422, |
|
"learning_rate": 8.564371228289563e-06, |
|
"loss": 0.8353, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.32357223750202235, |
|
"grad_norm": 0.7347855423064408, |
|
"learning_rate": 8.554454233759508e-06, |
|
"loss": 0.856, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.3243811680957774, |
|
"grad_norm": 0.7460223907838642, |
|
"learning_rate": 8.544508889401799e-06, |
|
"loss": 0.8437, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 0.3251900986895324, |
|
"grad_norm": 0.7539924040446289, |
|
"learning_rate": 8.534535274539103e-06, |
|
"loss": 0.8448, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.3259990292832875, |
|
"grad_norm": 0.8431518796240063, |
|
"learning_rate": 8.524533468719569e-06, |
|
"loss": 0.8545, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.32680795987704253, |
|
"grad_norm": 0.8116074595962663, |
|
"learning_rate": 8.5145035517162e-06, |
|
"loss": 0.8262, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.3276168904707976, |
|
"grad_norm": 0.7709832113170016, |
|
"learning_rate": 8.504445603526202e-06, |
|
"loss": 0.8718, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.32842582106455265, |
|
"grad_norm": 0.8030729878881923, |
|
"learning_rate": 8.494359704370357e-06, |
|
"loss": 0.8588, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.32923475165830773, |
|
"grad_norm": 0.8146354616260442, |
|
"learning_rate": 8.484245934692379e-06, |
|
"loss": 0.8448, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 0.33004368225206276, |
|
"grad_norm": 0.7977438419631868, |
|
"learning_rate": 8.474104375158277e-06, |
|
"loss": 0.8472, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.33085261284581785, |
|
"grad_norm": 0.7577130929037066, |
|
"learning_rate": 8.463935106655705e-06, |
|
"loss": 0.849, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 0.3316615434395729, |
|
"grad_norm": 0.7702289198050427, |
|
"learning_rate": 8.453738210293316e-06, |
|
"loss": 0.8646, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.33247047403332797, |
|
"grad_norm": 0.8007959864809794, |
|
"learning_rate": 8.443513767400126e-06, |
|
"loss": 0.8488, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 0.333279404627083, |
|
"grad_norm": 0.8288190327751878, |
|
"learning_rate": 8.433261859524856e-06, |
|
"loss": 0.8473, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.33408833522083803, |
|
"grad_norm": 0.7899197293508395, |
|
"learning_rate": 8.422982568435283e-06, |
|
"loss": 0.8503, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 0.3348972658145931, |
|
"grad_norm": 0.7977808601108305, |
|
"learning_rate": 8.412675976117585e-06, |
|
"loss": 0.8711, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.33570619640834815, |
|
"grad_norm": 0.8550202055185513, |
|
"learning_rate": 8.4023421647757e-06, |
|
"loss": 0.8296, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.33651512700210323, |
|
"grad_norm": 0.8370007865096064, |
|
"learning_rate": 8.391981216830651e-06, |
|
"loss": 0.8789, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.33732405759585826, |
|
"grad_norm": 0.7648678747284082, |
|
"learning_rate": 8.381593214919905e-06, |
|
"loss": 0.8615, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 0.33813298818961335, |
|
"grad_norm": 0.8338949429924307, |
|
"learning_rate": 8.371178241896708e-06, |
|
"loss": 0.8546, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.3389419187833684, |
|
"grad_norm": 0.7923771910011903, |
|
"learning_rate": 8.36073638082942e-06, |
|
"loss": 0.8421, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 0.33975084937712347, |
|
"grad_norm": 0.7717061524027085, |
|
"learning_rate": 8.350267715000857e-06, |
|
"loss": 0.8459, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.3405597799708785, |
|
"grad_norm": 0.7706870416392109, |
|
"learning_rate": 8.339772327907627e-06, |
|
"loss": 0.839, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 0.3413687105646336, |
|
"grad_norm": 0.8629826885755151, |
|
"learning_rate": 8.329250303259466e-06, |
|
"loss": 0.847, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.3421776411583886, |
|
"grad_norm": 0.7813813713851825, |
|
"learning_rate": 8.318701724978564e-06, |
|
"loss": 0.8627, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 0.34298657175214364, |
|
"grad_norm": 0.7760292657058412, |
|
"learning_rate": 8.308126677198896e-06, |
|
"loss": 0.843, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.34379550234589873, |
|
"grad_norm": 0.8222906943010696, |
|
"learning_rate": 8.29752524426556e-06, |
|
"loss": 0.8362, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.34460443293965376, |
|
"grad_norm": 0.8292261329180289, |
|
"learning_rate": 8.286897510734098e-06, |
|
"loss": 0.8548, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.34541336353340885, |
|
"grad_norm": 0.7772581002720523, |
|
"learning_rate": 8.276243561369815e-06, |
|
"loss": 0.8475, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 0.3462222941271639, |
|
"grad_norm": 0.8663068319361349, |
|
"learning_rate": 8.265563481147118e-06, |
|
"loss": 0.8656, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.34703122472091896, |
|
"grad_norm": 0.759426780159734, |
|
"learning_rate": 8.254857355248823e-06, |
|
"loss": 0.8272, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 0.347840155314674, |
|
"grad_norm": 0.8045042800571909, |
|
"learning_rate": 8.244125269065492e-06, |
|
"loss": 0.8284, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.3486490859084291, |
|
"grad_norm": 0.7696869214740462, |
|
"learning_rate": 8.233367308194735e-06, |
|
"loss": 0.8441, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 0.3494580165021841, |
|
"grad_norm": 0.8076486642638429, |
|
"learning_rate": 8.222583558440531e-06, |
|
"loss": 0.845, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.35026694709593914, |
|
"grad_norm": 0.7966361567902642, |
|
"learning_rate": 8.21177410581256e-06, |
|
"loss": 0.8512, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 0.3510758776896942, |
|
"grad_norm": 0.8192704344463675, |
|
"learning_rate": 8.200939036525495e-06, |
|
"loss": 0.8475, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.35188480828344926, |
|
"grad_norm": 0.7866644986853055, |
|
"learning_rate": 8.190078436998326e-06, |
|
"loss": 0.8481, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.35269373887720434, |
|
"grad_norm": 0.7784262615486034, |
|
"learning_rate": 8.179192393853667e-06, |
|
"loss": 0.8482, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.3535026694709594, |
|
"grad_norm": 0.7987375076221339, |
|
"learning_rate": 8.168280993917078e-06, |
|
"loss": 0.8315, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 0.35431160006471446, |
|
"grad_norm": 0.7513918277520256, |
|
"learning_rate": 8.15734432421634e-06, |
|
"loss": 0.8547, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.3551205306584695, |
|
"grad_norm": 0.8667321468753193, |
|
"learning_rate": 8.146382471980803e-06, |
|
"loss": 0.8583, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 0.3559294612522246, |
|
"grad_norm": 0.7568485546003565, |
|
"learning_rate": 8.135395524640659e-06, |
|
"loss": 0.8342, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.3567383918459796, |
|
"grad_norm": 0.8356531732827182, |
|
"learning_rate": 8.124383569826253e-06, |
|
"loss": 0.8642, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 0.3575473224397347, |
|
"grad_norm": 0.8080998046421258, |
|
"learning_rate": 8.113346695367393e-06, |
|
"loss": 0.8619, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.3583562530334897, |
|
"grad_norm": 0.7192475502436593, |
|
"learning_rate": 8.102284989292639e-06, |
|
"loss": 0.858, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 0.35916518362724476, |
|
"grad_norm": 0.8273826603550964, |
|
"learning_rate": 8.091198539828601e-06, |
|
"loss": 0.8433, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.35997411422099984, |
|
"grad_norm": 0.8607381682601364, |
|
"learning_rate": 8.080087435399248e-06, |
|
"loss": 0.8382, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.36078304481475487, |
|
"grad_norm": 0.8001621243743261, |
|
"learning_rate": 8.068951764625186e-06, |
|
"loss": 0.8406, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.36159197540850996, |
|
"grad_norm": 0.7615698239071352, |
|
"learning_rate": 8.057791616322958e-06, |
|
"loss": 0.8316, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 0.362400906002265, |
|
"grad_norm": 0.8191316557232389, |
|
"learning_rate": 8.046607079504345e-06, |
|
"loss": 0.8585, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.3632098365960201, |
|
"grad_norm": 0.8329244880945085, |
|
"learning_rate": 8.035398243375636e-06, |
|
"loss": 0.8237, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 0.3640187671897751, |
|
"grad_norm": 0.8151507773643624, |
|
"learning_rate": 8.024165197336934e-06, |
|
"loss": 0.875, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.3648276977835302, |
|
"grad_norm": 0.8450488670993342, |
|
"learning_rate": 8.012908030981442e-06, |
|
"loss": 0.8336, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 0.3656366283772852, |
|
"grad_norm": 0.7623122756894671, |
|
"learning_rate": 8.00162683409473e-06, |
|
"loss": 0.8535, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.3664455589710403, |
|
"grad_norm": 0.7904848999057988, |
|
"learning_rate": 7.990321696654044e-06, |
|
"loss": 0.8566, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 0.36725448956479534, |
|
"grad_norm": 0.7803510833606502, |
|
"learning_rate": 7.978992708827571e-06, |
|
"loss": 0.8544, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.36806342015855037, |
|
"grad_norm": 0.8005678354543746, |
|
"learning_rate": 7.967639960973727e-06, |
|
"loss": 0.8461, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.36887235075230546, |
|
"grad_norm": 0.8007519176589866, |
|
"learning_rate": 7.956263543640432e-06, |
|
"loss": 0.8409, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.3696812813460605, |
|
"grad_norm": 0.8474393359723089, |
|
"learning_rate": 7.944863547564396e-06, |
|
"loss": 0.8585, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 0.3704902119398156, |
|
"grad_norm": 0.8486471733965024, |
|
"learning_rate": 7.933440063670383e-06, |
|
"loss": 0.8399, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.3712991425335706, |
|
"grad_norm": 0.7861689750278411, |
|
"learning_rate": 7.921993183070497e-06, |
|
"loss": 0.8496, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 0.3721080731273257, |
|
"grad_norm": 0.8805931639748436, |
|
"learning_rate": 7.910522997063451e-06, |
|
"loss": 0.8351, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.3729170037210807, |
|
"grad_norm": 0.7946263432456658, |
|
"learning_rate": 7.899029597133836e-06, |
|
"loss": 0.8444, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 0.3737259343148358, |
|
"grad_norm": 0.8337819610743522, |
|
"learning_rate": 7.887513074951397e-06, |
|
"loss": 0.8491, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.37453486490859084, |
|
"grad_norm": 0.8157784141608686, |
|
"learning_rate": 7.875973522370294e-06, |
|
"loss": 0.835, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 0.3753437955023459, |
|
"grad_norm": 0.8127539829554938, |
|
"learning_rate": 7.864411031428379e-06, |
|
"loss": 0.8425, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.37615272609610095, |
|
"grad_norm": 0.7999538507762206, |
|
"learning_rate": 7.852825694346455e-06, |
|
"loss": 0.8364, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.376961656689856, |
|
"grad_norm": 0.8590002826350661, |
|
"learning_rate": 7.841217603527545e-06, |
|
"loss": 0.8415, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.37777058728361107, |
|
"grad_norm": 0.8424318909298105, |
|
"learning_rate": 7.82958685155615e-06, |
|
"loss": 0.8559, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 0.3785795178773661, |
|
"grad_norm": 0.7727695207961051, |
|
"learning_rate": 7.817933531197513e-06, |
|
"loss": 0.8479, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.3793884484711212, |
|
"grad_norm": 0.8080185911172371, |
|
"learning_rate": 7.806257735396879e-06, |
|
"loss": 0.832, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 0.3801973790648762, |
|
"grad_norm": 0.8144308118806884, |
|
"learning_rate": 7.794559557278757e-06, |
|
"loss": 0.8449, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.3810063096586313, |
|
"grad_norm": 0.8257947013812411, |
|
"learning_rate": 7.782839090146172e-06, |
|
"loss": 0.8579, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 0.38181524025238633, |
|
"grad_norm": 0.7945614016073755, |
|
"learning_rate": 7.771096427479925e-06, |
|
"loss": 0.8298, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.3826241708461414, |
|
"grad_norm": 0.8410802459537005, |
|
"learning_rate": 7.759331662937841e-06, |
|
"loss": 0.8491, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 0.38343310143989645, |
|
"grad_norm": 0.8308755908758182, |
|
"learning_rate": 7.747544890354031e-06, |
|
"loss": 0.8573, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.38424203203365154, |
|
"grad_norm": 0.8040651764511801, |
|
"learning_rate": 7.735736203738137e-06, |
|
"loss": 0.8344, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.38505096262740657, |
|
"grad_norm": 0.8125819462471874, |
|
"learning_rate": 7.723905697274586e-06, |
|
"loss": 0.8424, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.3858598932211616, |
|
"grad_norm": 0.7539076215491992, |
|
"learning_rate": 7.71205346532183e-06, |
|
"loss": 0.8495, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 0.3866688238149167, |
|
"grad_norm": 0.8155202307231507, |
|
"learning_rate": 7.700179602411615e-06, |
|
"loss": 0.8342, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.3874777544086717, |
|
"grad_norm": 0.8220478382780729, |
|
"learning_rate": 7.688284203248197e-06, |
|
"loss": 0.846, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3882866850024268, |
|
"grad_norm": 0.8673932844304086, |
|
"learning_rate": 7.67636736270761e-06, |
|
"loss": 0.875, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.38909561559618183, |
|
"grad_norm": 0.803133188698084, |
|
"learning_rate": 7.664429175836903e-06, |
|
"loss": 0.8672, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 0.3899045461899369, |
|
"grad_norm": 0.7796141487305905, |
|
"learning_rate": 7.652469737853372e-06, |
|
"loss": 0.8542, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.39071347678369195, |
|
"grad_norm": 0.7721547412514925, |
|
"learning_rate": 7.64048914414382e-06, |
|
"loss": 0.8274, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 0.39152240737744703, |
|
"grad_norm": 0.8076354344928516, |
|
"learning_rate": 7.628487490263779e-06, |
|
"loss": 0.8315, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.39233133797120207, |
|
"grad_norm": 0.8310201097209512, |
|
"learning_rate": 7.616464871936748e-06, |
|
"loss": 0.8478, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.39314026856495715, |
|
"grad_norm": 0.8088889309070267, |
|
"learning_rate": 7.60442138505345e-06, |
|
"loss": 0.8434, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.3939491991587122, |
|
"grad_norm": 0.8845883583499743, |
|
"learning_rate": 7.5923571256710396e-06, |
|
"loss": 0.827, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 0.3947581297524672, |
|
"grad_norm": 0.8075412429174051, |
|
"learning_rate": 7.580272190012357e-06, |
|
"loss": 0.8395, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.3955670603462223, |
|
"grad_norm": 0.781477252019071, |
|
"learning_rate": 7.5681666744651505e-06, |
|
"loss": 0.8157, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 0.39637599093997733, |
|
"grad_norm": 0.7999629302259256, |
|
"learning_rate": 7.556040675581311e-06, |
|
"loss": 0.8288, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.3971849215337324, |
|
"grad_norm": 0.8292218777187262, |
|
"learning_rate": 7.5438942900761035e-06, |
|
"loss": 0.8417, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 0.39799385212748745, |
|
"grad_norm": 0.7796119585780503, |
|
"learning_rate": 7.531727614827392e-06, |
|
"loss": 0.8512, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.39880278272124253, |
|
"grad_norm": 1.0387110538612172, |
|
"learning_rate": 7.519540746874869e-06, |
|
"loss": 0.8261, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 0.39961171331499756, |
|
"grad_norm": 0.8332505684625972, |
|
"learning_rate": 7.507333783419282e-06, |
|
"loss": 0.8643, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.40042064390875265, |
|
"grad_norm": 0.80731246004178, |
|
"learning_rate": 7.495106821821656e-06, |
|
"loss": 0.8542, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.4012295745025077, |
|
"grad_norm": 0.8894304174472868, |
|
"learning_rate": 7.482859959602518e-06, |
|
"loss": 0.8403, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.40203850509626277, |
|
"grad_norm": 0.8754692672700708, |
|
"learning_rate": 7.470593294441124e-06, |
|
"loss": 0.8516, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 0.4028474356900178, |
|
"grad_norm": 0.8789987496906461, |
|
"learning_rate": 7.4583069241746696e-06, |
|
"loss": 0.834, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.4036563662837728, |
|
"grad_norm": 0.841200352411981, |
|
"learning_rate": 7.446000946797519e-06, |
|
"loss": 0.8241, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 0.4044652968775279, |
|
"grad_norm": 0.8282889658691387, |
|
"learning_rate": 7.433675460460419e-06, |
|
"loss": 0.845, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.40527422747128294, |
|
"grad_norm": 0.8171176352536702, |
|
"learning_rate": 7.421330563469717e-06, |
|
"loss": 0.8361, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 0.40608315806503803, |
|
"grad_norm": 0.7424149733511415, |
|
"learning_rate": 7.408966354286575e-06, |
|
"loss": 0.8394, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.40689208865879306, |
|
"grad_norm": 0.772435215520305, |
|
"learning_rate": 7.396582931526194e-06, |
|
"loss": 0.8364, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 0.40770101925254815, |
|
"grad_norm": 0.8577362984373131, |
|
"learning_rate": 7.38418039395701e-06, |
|
"loss": 0.8522, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.4085099498463032, |
|
"grad_norm": 0.8324922368152244, |
|
"learning_rate": 7.371758840499921e-06, |
|
"loss": 0.837, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.40931888044005826, |
|
"grad_norm": 0.8167673718582185, |
|
"learning_rate": 7.359318370227494e-06, |
|
"loss": 0.8551, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.4101278110338133, |
|
"grad_norm": 0.8183481719062728, |
|
"learning_rate": 7.346859082363172e-06, |
|
"loss": 0.8653, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 0.4109367416275684, |
|
"grad_norm": 0.8526948175700703, |
|
"learning_rate": 7.334381076280483e-06, |
|
"loss": 0.8321, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.4117456722213234, |
|
"grad_norm": 0.8683127108850153, |
|
"learning_rate": 7.321884451502252e-06, |
|
"loss": 0.8391, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 0.41255460281507844, |
|
"grad_norm": 0.8156307062993907, |
|
"learning_rate": 7.309369307699802e-06, |
|
"loss": 0.8281, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.41336353340883353, |
|
"grad_norm": 0.831536181417552, |
|
"learning_rate": 7.296835744692163e-06, |
|
"loss": 0.8417, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 0.41417246400258856, |
|
"grad_norm": 0.840572967544839, |
|
"learning_rate": 7.2842838624452725e-06, |
|
"loss": 0.8172, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.41498139459634364, |
|
"grad_norm": 0.8063608871323279, |
|
"learning_rate": 7.271713761071181e-06, |
|
"loss": 0.8533, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 0.4157903251900987, |
|
"grad_norm": 0.9598879655118182, |
|
"learning_rate": 7.259125540827248e-06, |
|
"loss": 0.8323, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.41659925578385376, |
|
"grad_norm": 0.77131131850428, |
|
"learning_rate": 7.246519302115355e-06, |
|
"loss": 0.8202, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.4174081863776088, |
|
"grad_norm": 0.7740177253879231, |
|
"learning_rate": 7.233895145481086e-06, |
|
"loss": 0.8169, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.4182171169713639, |
|
"grad_norm": 0.7993458055375153, |
|
"learning_rate": 7.221253171612944e-06, |
|
"loss": 0.8281, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 0.4190260475651189, |
|
"grad_norm": 0.8017047616887952, |
|
"learning_rate": 7.208593481341536e-06, |
|
"loss": 0.8551, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.419834978158874, |
|
"grad_norm": 0.7947124171074315, |
|
"learning_rate": 7.195916175638773e-06, |
|
"loss": 0.8347, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 0.420643908752629, |
|
"grad_norm": 0.8878564841823726, |
|
"learning_rate": 7.183221355617065e-06, |
|
"loss": 0.8377, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.42145283934638406, |
|
"grad_norm": 0.8125554403726852, |
|
"learning_rate": 7.170509122528511e-06, |
|
"loss": 0.8322, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 0.42226176994013914, |
|
"grad_norm": 0.8652065540252103, |
|
"learning_rate": 7.157779577764099e-06, |
|
"loss": 0.8367, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.4230707005338942, |
|
"grad_norm": 0.8610622568512487, |
|
"learning_rate": 7.145032822852889e-06, |
|
"loss": 0.8378, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 0.42387963112764926, |
|
"grad_norm": 0.7876508067233663, |
|
"learning_rate": 7.132268959461209e-06, |
|
"loss": 0.8375, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.4246885617214043, |
|
"grad_norm": 0.8282333937739974, |
|
"learning_rate": 7.119488089391836e-06, |
|
"loss": 0.8406, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.4254974923151594, |
|
"grad_norm": 0.8305925610445289, |
|
"learning_rate": 7.106690314583199e-06, |
|
"loss": 0.8414, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.4263064229089144, |
|
"grad_norm": 0.789770771549953, |
|
"learning_rate": 7.0938757371085485e-06, |
|
"loss": 0.853, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 0.4271153535026695, |
|
"grad_norm": 0.8456657346625689, |
|
"learning_rate": 7.081044459175159e-06, |
|
"loss": 0.8308, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.4279242840964245, |
|
"grad_norm": 0.8352288349586666, |
|
"learning_rate": 7.068196583123495e-06, |
|
"loss": 0.8269, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 0.4287332146901796, |
|
"grad_norm": 0.8448828847445977, |
|
"learning_rate": 7.055332211426417e-06, |
|
"loss": 0.8388, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.42954214528393464, |
|
"grad_norm": 0.858853034822238, |
|
"learning_rate": 7.042451446688342e-06, |
|
"loss": 0.8298, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 0.43035107587768967, |
|
"grad_norm": 0.8391759096265385, |
|
"learning_rate": 7.029554391644441e-06, |
|
"loss": 0.8371, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.43116000647144476, |
|
"grad_norm": 0.7973420420214934, |
|
"learning_rate": 7.016641149159816e-06, |
|
"loss": 0.8377, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 0.4319689370651998, |
|
"grad_norm": 0.8010191599865539, |
|
"learning_rate": 7.00371182222867e-06, |
|
"loss": 0.8483, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.4327778676589549, |
|
"grad_norm": 0.8013088566432139, |
|
"learning_rate": 6.9907665139735035e-06, |
|
"loss": 0.8366, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.4335867982527099, |
|
"grad_norm": 0.8481738940133525, |
|
"learning_rate": 6.977805327644275e-06, |
|
"loss": 0.8311, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.434395728846465, |
|
"grad_norm": 0.7925446967652235, |
|
"learning_rate": 6.964828366617583e-06, |
|
"loss": 0.8466, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 0.43520465944022, |
|
"grad_norm": 0.810554276504509, |
|
"learning_rate": 6.95183573439585e-06, |
|
"loss": 0.8468, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.4360135900339751, |
|
"grad_norm": 0.8018391904385581, |
|
"learning_rate": 6.938827534606484e-06, |
|
"loss": 0.8356, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 0.43682252062773014, |
|
"grad_norm": 0.8273266309236444, |
|
"learning_rate": 6.925803871001058e-06, |
|
"loss": 0.8358, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.4376314512214852, |
|
"grad_norm": 0.867800658645478, |
|
"learning_rate": 6.912764847454485e-06, |
|
"loss": 0.8332, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 0.43844038181524025, |
|
"grad_norm": 0.8366321913701985, |
|
"learning_rate": 6.899710567964184e-06, |
|
"loss": 0.8289, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.4392493124089953, |
|
"grad_norm": 0.851318956301739, |
|
"learning_rate": 6.8866411366492546e-06, |
|
"loss": 0.8252, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 0.44005824300275037, |
|
"grad_norm": 0.8661629567364111, |
|
"learning_rate": 6.873556657749646e-06, |
|
"loss": 0.8433, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.4408671735965054, |
|
"grad_norm": 0.8890212561668989, |
|
"learning_rate": 6.860457235625322e-06, |
|
"loss": 0.8327, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.4416761041902605, |
|
"grad_norm": 0.8479920834756562, |
|
"learning_rate": 6.847342974755435e-06, |
|
"loss": 0.8408, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.4424850347840155, |
|
"grad_norm": 0.8040960156855093, |
|
"learning_rate": 6.834213979737488e-06, |
|
"loss": 0.8667, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 0.4432939653777706, |
|
"grad_norm": 0.8449566845357666, |
|
"learning_rate": 6.821070355286498e-06, |
|
"loss": 0.8318, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.44410289597152564, |
|
"grad_norm": 0.8008355458582818, |
|
"learning_rate": 6.807912206234168e-06, |
|
"loss": 0.8233, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 0.4449118265652807, |
|
"grad_norm": 0.8051699329730306, |
|
"learning_rate": 6.794739637528051e-06, |
|
"loss": 0.8386, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.44572075715903575, |
|
"grad_norm": 0.8602857948468763, |
|
"learning_rate": 6.7815527542307e-06, |
|
"loss": 0.8399, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 0.44652968775279084, |
|
"grad_norm": 0.8187853844017233, |
|
"learning_rate": 6.768351661518845e-06, |
|
"loss": 0.8176, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.44733861834654587, |
|
"grad_norm": 0.8053208482891812, |
|
"learning_rate": 6.755136464682546e-06, |
|
"loss": 0.8475, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 0.4481475489403009, |
|
"grad_norm": 0.8750589437015153, |
|
"learning_rate": 6.741907269124358e-06, |
|
"loss": 0.8279, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.448956479534056, |
|
"grad_norm": 0.8815876030660124, |
|
"learning_rate": 6.728664180358487e-06, |
|
"loss": 0.8438, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.449765410127811, |
|
"grad_norm": 0.8318847722342249, |
|
"learning_rate": 6.715407304009948e-06, |
|
"loss": 0.852, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.4505743407215661, |
|
"grad_norm": 0.8913600813483957, |
|
"learning_rate": 6.702136745813721e-06, |
|
"loss": 0.8435, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45138327131532113, |
|
"grad_norm": 0.8207757002268261, |
|
"learning_rate": 6.688852611613921e-06, |
|
"loss": 0.8432, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.4521922019090762, |
|
"grad_norm": 0.8964348508265096, |
|
"learning_rate": 6.675555007362931e-06, |
|
"loss": 0.8492, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 0.45300113250283125, |
|
"grad_norm": 0.88701161837332, |
|
"learning_rate": 6.662244039120575e-06, |
|
"loss": 0.8488, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.45381006309658634, |
|
"grad_norm": 0.7897608660219696, |
|
"learning_rate": 6.648919813053266e-06, |
|
"loss": 0.8436, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 0.45461899369034137, |
|
"grad_norm": 0.8277360620764583, |
|
"learning_rate": 6.635582435433161e-06, |
|
"loss": 0.8473, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.4554279242840964, |
|
"grad_norm": 0.8476898435569198, |
|
"learning_rate": 6.6222320126373105e-06, |
|
"loss": 0.8248, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 0.4562368548778515, |
|
"grad_norm": 0.8213309435302791, |
|
"learning_rate": 6.6088686511468106e-06, |
|
"loss": 0.8208, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.4570457854716065, |
|
"grad_norm": 0.8314089865069589, |
|
"learning_rate": 6.595492457545953e-06, |
|
"loss": 0.8654, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.4578547160653616, |
|
"grad_norm": 0.8098173131527258, |
|
"learning_rate": 6.582103538521383e-06, |
|
"loss": 0.817, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.45866364665911663, |
|
"grad_norm": 0.8530608603749926, |
|
"learning_rate": 6.568702000861234e-06, |
|
"loss": 0.858, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 0.4594725772528717, |
|
"grad_norm": 0.884209021072577, |
|
"learning_rate": 6.5552879514542915e-06, |
|
"loss": 0.8314, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.46028150784662675, |
|
"grad_norm": 0.8215391384460337, |
|
"learning_rate": 6.541861497289126e-06, |
|
"loss": 0.8284, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 0.46109043844038183, |
|
"grad_norm": 0.7972506723621342, |
|
"learning_rate": 6.528422745453251e-06, |
|
"loss": 0.8439, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.46189936903413686, |
|
"grad_norm": 0.8602949746737556, |
|
"learning_rate": 6.514971803132264e-06, |
|
"loss": 0.8366, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 0.46270829962789195, |
|
"grad_norm": 0.8506988739569268, |
|
"learning_rate": 6.50150877760899e-06, |
|
"loss": 0.86, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.463517230221647, |
|
"grad_norm": 0.7981906008336666, |
|
"learning_rate": 6.488033776262632e-06, |
|
"loss": 0.8531, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 0.464326160815402, |
|
"grad_norm": 0.7798400777450988, |
|
"learning_rate": 6.474546906567905e-06, |
|
"loss": 0.8457, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.4651350914091571, |
|
"grad_norm": 0.8391735146467554, |
|
"learning_rate": 6.46104827609419e-06, |
|
"loss": 0.8408, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.46594402200291213, |
|
"grad_norm": 0.8052808401629455, |
|
"learning_rate": 6.447537992504663e-06, |
|
"loss": 0.8486, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.4667529525966672, |
|
"grad_norm": 0.7988941689518605, |
|
"learning_rate": 6.434016163555452e-06, |
|
"loss": 0.846, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 0.46756188319042224, |
|
"grad_norm": 0.8165894688556871, |
|
"learning_rate": 6.420482897094764e-06, |
|
"loss": 0.8319, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.46837081378417733, |
|
"grad_norm": 0.8121894000178076, |
|
"learning_rate": 6.406938301062031e-06, |
|
"loss": 0.8388, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 0.46917974437793236, |
|
"grad_norm": 0.835869232540653, |
|
"learning_rate": 6.393382483487049e-06, |
|
"loss": 0.8228, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.46998867497168745, |
|
"grad_norm": 0.912311843916525, |
|
"learning_rate": 6.379815552489112e-06, |
|
"loss": 0.819, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 0.4707976055654425, |
|
"grad_norm": 0.8202079981924546, |
|
"learning_rate": 6.366237616276161e-06, |
|
"loss": 0.825, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.47160653615919756, |
|
"grad_norm": 0.82471875114036, |
|
"learning_rate": 6.3526487831439045e-06, |
|
"loss": 0.8238, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 0.4724154667529526, |
|
"grad_norm": 0.9037615912491918, |
|
"learning_rate": 6.339049161474965e-06, |
|
"loss": 0.8202, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.4732243973467076, |
|
"grad_norm": 0.8152831104583218, |
|
"learning_rate": 6.3254388597380165e-06, |
|
"loss": 0.8448, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.4740333279404627, |
|
"grad_norm": 0.8118468395706068, |
|
"learning_rate": 6.311817986486917e-06, |
|
"loss": 0.8342, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.47484225853421774, |
|
"grad_norm": 0.8075212743777721, |
|
"learning_rate": 6.298186650359832e-06, |
|
"loss": 0.8274, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 0.47565118912797283, |
|
"grad_norm": 0.859392711702304, |
|
"learning_rate": 6.284544960078387e-06, |
|
"loss": 0.8401, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.47646011972172786, |
|
"grad_norm": 0.8858782454028935, |
|
"learning_rate": 6.270893024446788e-06, |
|
"loss": 0.8624, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 0.47726905031548295, |
|
"grad_norm": 0.8427992758295506, |
|
"learning_rate": 6.257230952350954e-06, |
|
"loss": 0.839, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.478077980909238, |
|
"grad_norm": 0.8680923960328171, |
|
"learning_rate": 6.243558852757654e-06, |
|
"loss": 0.8577, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 0.47888691150299306, |
|
"grad_norm": 0.8682921557831497, |
|
"learning_rate": 6.229876834713633e-06, |
|
"loss": 0.839, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.4796958420967481, |
|
"grad_norm": 0.8040852911882872, |
|
"learning_rate": 6.216185007344745e-06, |
|
"loss": 0.8501, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 0.4805047726905032, |
|
"grad_norm": 0.7631057139150006, |
|
"learning_rate": 6.202483479855083e-06, |
|
"loss": 0.8268, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.4813137032842582, |
|
"grad_norm": 0.813256135509401, |
|
"learning_rate": 6.188772361526104e-06, |
|
"loss": 0.8196, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.48212263387801324, |
|
"grad_norm": 0.8297884897498811, |
|
"learning_rate": 6.175051761715762e-06, |
|
"loss": 0.848, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.4829315644717683, |
|
"grad_norm": 0.8334678385544815, |
|
"learning_rate": 6.161321789857635e-06, |
|
"loss": 0.8367, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 0.48374049506552336, |
|
"grad_norm": 0.8985811450220941, |
|
"learning_rate": 6.147582555460048e-06, |
|
"loss": 0.843, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.48454942565927844, |
|
"grad_norm": 0.8362819535768706, |
|
"learning_rate": 6.133834168105206e-06, |
|
"loss": 0.8385, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 0.4853583562530335, |
|
"grad_norm": 0.8317479151299757, |
|
"learning_rate": 6.120076737448314e-06, |
|
"loss": 0.8331, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.48616728684678856, |
|
"grad_norm": 0.816034350431961, |
|
"learning_rate": 6.106310373216706e-06, |
|
"loss": 0.8389, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 0.4869762174405436, |
|
"grad_norm": 0.8435438416835686, |
|
"learning_rate": 6.092535185208973e-06, |
|
"loss": 0.8144, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.4877851480342987, |
|
"grad_norm": 0.8650473536522931, |
|
"learning_rate": 6.078751283294075e-06, |
|
"loss": 0.8415, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 0.4885940786280537, |
|
"grad_norm": 0.8631976632792485, |
|
"learning_rate": 6.0649587774104775e-06, |
|
"loss": 0.8384, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.4894030092218088, |
|
"grad_norm": 0.8051930942211001, |
|
"learning_rate": 6.0511577775652744e-06, |
|
"loss": 0.8307, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.4902119398155638, |
|
"grad_norm": 0.9000033790636423, |
|
"learning_rate": 6.037348393833298e-06, |
|
"loss": 0.8288, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.49102087040931885, |
|
"grad_norm": 0.8370590038208652, |
|
"learning_rate": 6.0235307363562524e-06, |
|
"loss": 0.8331, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 0.49182980100307394, |
|
"grad_norm": 0.8387976240360758, |
|
"learning_rate": 6.009704915341835e-06, |
|
"loss": 0.8429, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.49263873159682897, |
|
"grad_norm": 0.849367045833896, |
|
"learning_rate": 5.9958710410628515e-06, |
|
"loss": 0.8245, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 0.49344766219058406, |
|
"grad_norm": 0.817264343403589, |
|
"learning_rate": 5.9820292238563404e-06, |
|
"loss": 0.8279, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.4942565927843391, |
|
"grad_norm": 0.8397242273647842, |
|
"learning_rate": 5.96817957412269e-06, |
|
"loss": 0.8211, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 0.4950655233780942, |
|
"grad_norm": 0.8292950717170571, |
|
"learning_rate": 5.954322202324759e-06, |
|
"loss": 0.8447, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.4958744539718492, |
|
"grad_norm": 0.8153225989185483, |
|
"learning_rate": 5.940457218987003e-06, |
|
"loss": 0.849, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 0.4966833845656043, |
|
"grad_norm": 0.8817065793496368, |
|
"learning_rate": 5.926584734694579e-06, |
|
"loss": 0.831, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.4974923151593593, |
|
"grad_norm": 0.8223135029609022, |
|
"learning_rate": 5.912704860092473e-06, |
|
"loss": 0.8442, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.4983012457531144, |
|
"grad_norm": 0.8254057255135826, |
|
"learning_rate": 5.898817705884615e-06, |
|
"loss": 0.8231, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.49911017634686944, |
|
"grad_norm": 0.8516771176045018, |
|
"learning_rate": 5.8849233828329964e-06, |
|
"loss": 0.8325, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 0.49991910694062447, |
|
"grad_norm": 0.8770116756575128, |
|
"learning_rate": 5.871022001756786e-06, |
|
"loss": 0.8353, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.5007280375343796, |
|
"grad_norm": 0.8416392905846662, |
|
"learning_rate": 5.857113673531446e-06, |
|
"loss": 0.8411, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 0.5015369681281346, |
|
"grad_norm": 0.8727681516631783, |
|
"learning_rate": 5.843198509087848e-06, |
|
"loss": 0.8327, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.5023458987218896, |
|
"grad_norm": 0.8122836804739928, |
|
"learning_rate": 5.829276619411392e-06, |
|
"loss": 0.8156, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 0.5031548293156447, |
|
"grad_norm": 0.8419116014056326, |
|
"learning_rate": 5.815348115541112e-06, |
|
"loss": 0.8406, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.5039637599093998, |
|
"grad_norm": 0.8496791968992736, |
|
"learning_rate": 5.801413108568798e-06, |
|
"loss": 0.8428, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 0.5047726905031549, |
|
"grad_norm": 0.8370866951694025, |
|
"learning_rate": 5.787471709638109e-06, |
|
"loss": 0.8397, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.5055816210969099, |
|
"grad_norm": 0.8807740954559549, |
|
"learning_rate": 5.7735240299436825e-06, |
|
"loss": 0.8281, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.5063905516906649, |
|
"grad_norm": 0.880842454890424, |
|
"learning_rate": 5.759570180730255e-06, |
|
"loss": 0.8261, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.50719948228442, |
|
"grad_norm": 0.8501925674960217, |
|
"learning_rate": 5.745610273291766e-06, |
|
"loss": 0.8182, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 0.508008412878175, |
|
"grad_norm": 0.7885937426308233, |
|
"learning_rate": 5.731644418970478e-06, |
|
"loss": 0.8006, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.5088173434719301, |
|
"grad_norm": 0.8453255098699243, |
|
"learning_rate": 5.717672729156082e-06, |
|
"loss": 0.8408, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 0.5096262740656852, |
|
"grad_norm": 0.9027259053341657, |
|
"learning_rate": 5.703695315284814e-06, |
|
"loss": 0.8415, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.5104352046594403, |
|
"grad_norm": 0.8760720225211055, |
|
"learning_rate": 5.689712288838561e-06, |
|
"loss": 0.8459, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 0.5112441352531952, |
|
"grad_norm": 0.8371540278282537, |
|
"learning_rate": 5.675723761343981e-06, |
|
"loss": 0.8217, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.5120530658469503, |
|
"grad_norm": 0.8832530229547365, |
|
"learning_rate": 5.661729844371602e-06, |
|
"loss": 0.8321, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 0.5128619964407054, |
|
"grad_norm": 0.8466920163343052, |
|
"learning_rate": 5.64773064953494e-06, |
|
"loss": 0.8443, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.5136709270344605, |
|
"grad_norm": 0.8386270391117758, |
|
"learning_rate": 5.633726288489609e-06, |
|
"loss": 0.8215, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.5144798576282155, |
|
"grad_norm": 0.8968915867410968, |
|
"learning_rate": 5.619716872932422e-06, |
|
"loss": 0.8369, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.5152887882219706, |
|
"grad_norm": 0.8572970506555265, |
|
"learning_rate": 5.6057025146005125e-06, |
|
"loss": 0.8312, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 0.5160977188157256, |
|
"grad_norm": 0.856451059153177, |
|
"learning_rate": 5.591683325270434e-06, |
|
"loss": 0.828, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.5169066494094806, |
|
"grad_norm": 0.8777484476404416, |
|
"learning_rate": 5.577659416757267e-06, |
|
"loss": 0.8408, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 0.5177155800032357, |
|
"grad_norm": 0.8571851342757004, |
|
"learning_rate": 5.5636309009137404e-06, |
|
"loss": 0.8337, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.5185245105969908, |
|
"grad_norm": 0.829598576934011, |
|
"learning_rate": 5.549597889629325e-06, |
|
"loss": 0.8507, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 0.5193334411907459, |
|
"grad_norm": 0.8763914543424554, |
|
"learning_rate": 5.535560494829345e-06, |
|
"loss": 0.8363, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.5201423717845008, |
|
"grad_norm": 0.8614187326018883, |
|
"learning_rate": 5.521518828474092e-06, |
|
"loss": 0.8254, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 0.5209513023782559, |
|
"grad_norm": 0.8803171537853229, |
|
"learning_rate": 5.507473002557922e-06, |
|
"loss": 0.822, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.521760232972011, |
|
"grad_norm": 0.8826978372002413, |
|
"learning_rate": 5.493423129108373e-06, |
|
"loss": 0.8249, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.5225691635657661, |
|
"grad_norm": 0.8712751190362638, |
|
"learning_rate": 5.47936932018526e-06, |
|
"loss": 0.8423, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.5233780941595211, |
|
"grad_norm": 0.8238326567914839, |
|
"learning_rate": 5.465311687879785e-06, |
|
"loss": 0.824, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 0.5241870247532762, |
|
"grad_norm": 0.8899415075708137, |
|
"learning_rate": 5.4512503443136555e-06, |
|
"loss": 0.8663, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.5249959553470313, |
|
"grad_norm": 0.9621216340368667, |
|
"learning_rate": 5.437185401638168e-06, |
|
"loss": 0.8414, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 0.5258048859407862, |
|
"grad_norm": 0.8156708167135952, |
|
"learning_rate": 5.423116972033332e-06, |
|
"loss": 0.8249, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.5266138165345413, |
|
"grad_norm": 0.8179857114204797, |
|
"learning_rate": 5.409045167706962e-06, |
|
"loss": 0.8421, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.5274227471282964, |
|
"grad_norm": 0.9127466183923784, |
|
"learning_rate": 5.394970100893797e-06, |
|
"loss": 0.8359, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.5282316777220515, |
|
"grad_norm": 0.8430058361090182, |
|
"learning_rate": 5.380891883854591e-06, |
|
"loss": 0.8152, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 0.5290406083158065, |
|
"grad_norm": 0.9184787098267886, |
|
"learning_rate": 5.366810628875226e-06, |
|
"loss": 0.8253, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.5298495389095615, |
|
"grad_norm": 0.8924101947820497, |
|
"learning_rate": 5.352726448265808e-06, |
|
"loss": 0.8386, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.5306584695033166, |
|
"grad_norm": 0.7686705583327001, |
|
"learning_rate": 5.338639454359792e-06, |
|
"loss": 0.8305, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.5314674000970717, |
|
"grad_norm": 0.8264073173393284, |
|
"learning_rate": 5.324549759513058e-06, |
|
"loss": 0.8279, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 0.5322763306908267, |
|
"grad_norm": 0.7898714553275273, |
|
"learning_rate": 5.310457476103033e-06, |
|
"loss": 0.832, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.5330852612845818, |
|
"grad_norm": 0.915049917158178, |
|
"learning_rate": 5.2963627165277884e-06, |
|
"loss": 0.8423, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 0.5338941918783369, |
|
"grad_norm": 0.8390857749633213, |
|
"learning_rate": 5.28226559320515e-06, |
|
"loss": 0.7937, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.5347031224720918, |
|
"grad_norm": 0.8766888011605986, |
|
"learning_rate": 5.268166218571792e-06, |
|
"loss": 0.8409, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 0.5355120530658469, |
|
"grad_norm": 0.9055264757342382, |
|
"learning_rate": 5.254064705082345e-06, |
|
"loss": 0.8374, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.536320983659602, |
|
"grad_norm": 0.9082065836332707, |
|
"learning_rate": 5.239961165208499e-06, |
|
"loss": 0.8424, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 0.5371299142533571, |
|
"grad_norm": 0.8630555734437255, |
|
"learning_rate": 5.2258557114381085e-06, |
|
"loss": 0.8492, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.5379388448471121, |
|
"grad_norm": 0.8387927972292225, |
|
"learning_rate": 5.211748456274291e-06, |
|
"loss": 0.8162, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.5387477754408672, |
|
"grad_norm": 0.8838319893113061, |
|
"learning_rate": 5.197639512234532e-06, |
|
"loss": 0.8302, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.5395567060346222, |
|
"grad_norm": 0.8810506540467933, |
|
"learning_rate": 5.183528991849784e-06, |
|
"loss": 0.8246, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 0.5403656366283773, |
|
"grad_norm": 0.9032568956242902, |
|
"learning_rate": 5.16941700766358e-06, |
|
"loss": 0.8302, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.5411745672221323, |
|
"grad_norm": 0.9878346022756306, |
|
"learning_rate": 5.155303672231123e-06, |
|
"loss": 0.8429, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 0.5419834978158874, |
|
"grad_norm": 0.7990363456989189, |
|
"learning_rate": 5.141189098118392e-06, |
|
"loss": 0.8295, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.5427924284096425, |
|
"grad_norm": 0.849668666846281, |
|
"learning_rate": 5.127073397901248e-06, |
|
"loss": 0.8447, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 0.5436013590033975, |
|
"grad_norm": 0.8870097311869496, |
|
"learning_rate": 5.112956684164532e-06, |
|
"loss": 0.8291, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.5444102895971525, |
|
"grad_norm": 0.8786319797950302, |
|
"learning_rate": 5.09883906950117e-06, |
|
"loss": 0.8325, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 0.5452192201909076, |
|
"grad_norm": 0.8559879223353866, |
|
"learning_rate": 5.084720666511276e-06, |
|
"loss": 0.8343, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.5460281507846627, |
|
"grad_norm": 0.8393771435379772, |
|
"learning_rate": 5.070601587801246e-06, |
|
"loss": 0.8264, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.5468370813784177, |
|
"grad_norm": 0.8756680886902389, |
|
"learning_rate": 5.056481945982871e-06, |
|
"loss": 0.8363, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.5476460119721728, |
|
"grad_norm": 0.8643664631148287, |
|
"learning_rate": 5.042361853672429e-06, |
|
"loss": 0.8508, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 0.5484549425659279, |
|
"grad_norm": 0.8396247416018382, |
|
"learning_rate": 5.0282414234897905e-06, |
|
"loss": 0.8313, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.549263873159683, |
|
"grad_norm": 0.8799962735756717, |
|
"learning_rate": 5.014120768057526e-06, |
|
"loss": 0.8394, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 0.5500728037534379, |
|
"grad_norm": 0.8585807475699035, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8259, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.550881734347193, |
|
"grad_norm": 0.8967809635243733, |
|
"learning_rate": 4.985879231942475e-06, |
|
"loss": 0.8232, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 0.5516906649409481, |
|
"grad_norm": 0.8853595174049823, |
|
"learning_rate": 4.97175857651021e-06, |
|
"loss": 0.8196, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.5524995955347031, |
|
"grad_norm": 0.839604935690722, |
|
"learning_rate": 4.957638146327575e-06, |
|
"loss": 0.8361, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 0.5533085261284582, |
|
"grad_norm": 0.8391467792592279, |
|
"learning_rate": 4.943518054017131e-06, |
|
"loss": 0.8272, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.5541174567222132, |
|
"grad_norm": 0.8844616497057926, |
|
"learning_rate": 4.929398412198756e-06, |
|
"loss": 0.8233, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.5549263873159683, |
|
"grad_norm": 0.8052750334117075, |
|
"learning_rate": 4.915279333488726e-06, |
|
"loss": 0.8479, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.5557353179097233, |
|
"grad_norm": 0.8647851317448605, |
|
"learning_rate": 4.90116093049883e-06, |
|
"loss": 0.819, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 0.5565442485034784, |
|
"grad_norm": 0.8453777869125878, |
|
"learning_rate": 4.887043315835469e-06, |
|
"loss": 0.8536, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.5573531790972335, |
|
"grad_norm": 0.8705052285115737, |
|
"learning_rate": 4.872926602098756e-06, |
|
"loss": 0.8277, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 0.5581621096909886, |
|
"grad_norm": 0.861071586334625, |
|
"learning_rate": 4.85881090188161e-06, |
|
"loss": 0.8081, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.5589710402847435, |
|
"grad_norm": 0.8558895543087007, |
|
"learning_rate": 4.844696327768878e-06, |
|
"loss": 0.8086, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 0.5597799708784986, |
|
"grad_norm": 0.9566303499060257, |
|
"learning_rate": 4.83058299233642e-06, |
|
"loss": 0.8388, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.5605889014722537, |
|
"grad_norm": 0.8480580281776552, |
|
"learning_rate": 4.8164710081502165e-06, |
|
"loss": 0.8464, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 0.5613978320660087, |
|
"grad_norm": 0.8577968798548672, |
|
"learning_rate": 4.802360487765471e-06, |
|
"loss": 0.8331, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.5622067626597638, |
|
"grad_norm": 0.8683642459249713, |
|
"learning_rate": 4.788251543725711e-06, |
|
"loss": 0.8379, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.5630156932535189, |
|
"grad_norm": 0.8893404896273497, |
|
"learning_rate": 4.774144288561893e-06, |
|
"loss": 0.8138, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.5638246238472739, |
|
"grad_norm": 0.8886059419047985, |
|
"learning_rate": 4.7600388347915025e-06, |
|
"loss": 0.8306, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 0.5646335544410289, |
|
"grad_norm": 0.8630514971290992, |
|
"learning_rate": 4.745935294917658e-06, |
|
"loss": 0.8348, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.565442485034784, |
|
"grad_norm": 0.8662086181368763, |
|
"learning_rate": 4.731833781428208e-06, |
|
"loss": 0.8272, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 0.5662514156285391, |
|
"grad_norm": 0.8982952976447754, |
|
"learning_rate": 4.7177344067948526e-06, |
|
"loss": 0.8525, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.5670603462222942, |
|
"grad_norm": 0.8422016831549223, |
|
"learning_rate": 4.703637283472213e-06, |
|
"loss": 0.8501, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 0.5678692768160492, |
|
"grad_norm": 0.9548881231159024, |
|
"learning_rate": 4.689542523896969e-06, |
|
"loss": 0.8328, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.5686782074098042, |
|
"grad_norm": 0.9166786357525355, |
|
"learning_rate": 4.6754502404869434e-06, |
|
"loss": 0.8205, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 0.5694871380035593, |
|
"grad_norm": 0.9061529794904933, |
|
"learning_rate": 4.661360545640209e-06, |
|
"loss": 0.8352, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.5702960685973143, |
|
"grad_norm": 0.8412734635963403, |
|
"learning_rate": 4.647273551734193e-06, |
|
"loss": 0.8348, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.5711049991910694, |
|
"grad_norm": 0.8664917924780835, |
|
"learning_rate": 4.633189371124778e-06, |
|
"loss": 0.8274, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.5719139297848245, |
|
"grad_norm": 0.8421186404143896, |
|
"learning_rate": 4.619108116145411e-06, |
|
"loss": 0.8354, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 0.5727228603785796, |
|
"grad_norm": 0.9394380176490387, |
|
"learning_rate": 4.6050298991062045e-06, |
|
"loss": 0.8269, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.5735317909723345, |
|
"grad_norm": 0.8660091216092733, |
|
"learning_rate": 4.590954832293039e-06, |
|
"loss": 0.8431, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 0.5743407215660896, |
|
"grad_norm": 0.8453775529626361, |
|
"learning_rate": 4.57688302796667e-06, |
|
"loss": 0.8189, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.5751496521598447, |
|
"grad_norm": 0.9293223006762544, |
|
"learning_rate": 4.562814598361834e-06, |
|
"loss": 0.8272, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 0.5759585827535998, |
|
"grad_norm": 0.8682075221715769, |
|
"learning_rate": 4.548749655686346e-06, |
|
"loss": 0.8345, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.5767675133473548, |
|
"grad_norm": 0.8962654121926869, |
|
"learning_rate": 4.534688312120216e-06, |
|
"loss": 0.8328, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 0.5775764439411099, |
|
"grad_norm": 0.8465460016413964, |
|
"learning_rate": 4.520630679814742e-06, |
|
"loss": 0.8203, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.5783853745348649, |
|
"grad_norm": 0.8542030722056939, |
|
"learning_rate": 4.506576870891628e-06, |
|
"loss": 0.8356, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.5791943051286199, |
|
"grad_norm": 0.8664072736373561, |
|
"learning_rate": 4.49252699744208e-06, |
|
"loss": 0.8282, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.580003235722375, |
|
"grad_norm": 0.8908304692018135, |
|
"learning_rate": 4.47848117152591e-06, |
|
"loss": 0.8486, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 0.5808121663161301, |
|
"grad_norm": 0.9604353292765517, |
|
"learning_rate": 4.464439505170656e-06, |
|
"loss": 0.828, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.5816210969098852, |
|
"grad_norm": 0.8847638351175713, |
|
"learning_rate": 4.450402110370677e-06, |
|
"loss": 0.8282, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 0.5824300275036401, |
|
"grad_norm": 0.8448125761174995, |
|
"learning_rate": 4.43636909908626e-06, |
|
"loss": 0.8381, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.5832389580973952, |
|
"grad_norm": 0.899925581945171, |
|
"learning_rate": 4.422340583242733e-06, |
|
"loss": 0.8316, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 0.5840478886911503, |
|
"grad_norm": 0.8270612926284294, |
|
"learning_rate": 4.408316674729569e-06, |
|
"loss": 0.8369, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.5848568192849054, |
|
"grad_norm": 0.8514310722916417, |
|
"learning_rate": 4.394297485399488e-06, |
|
"loss": 0.8123, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 0.5856657498786604, |
|
"grad_norm": 0.847607926161957, |
|
"learning_rate": 4.3802831270675785e-06, |
|
"loss": 0.8125, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.5864746804724155, |
|
"grad_norm": 0.8743645770925027, |
|
"learning_rate": 4.3662737115103925e-06, |
|
"loss": 0.8175, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.5872836110661706, |
|
"grad_norm": 0.8872414789266562, |
|
"learning_rate": 4.35226935046506e-06, |
|
"loss": 0.8128, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.5880925416599255, |
|
"grad_norm": 0.8748952049996549, |
|
"learning_rate": 4.338270155628401e-06, |
|
"loss": 0.8323, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 0.5889014722536806, |
|
"grad_norm": 0.9131605811984947, |
|
"learning_rate": 4.324276238656023e-06, |
|
"loss": 0.8266, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.5897104028474357, |
|
"grad_norm": 0.9378108739015335, |
|
"learning_rate": 4.31028771116144e-06, |
|
"loss": 0.8394, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 0.5905193334411908, |
|
"grad_norm": 0.9270250400148877, |
|
"learning_rate": 4.2963046847151875e-06, |
|
"loss": 0.8415, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.5913282640349458, |
|
"grad_norm": 0.8427644597969671, |
|
"learning_rate": 4.282327270843919e-06, |
|
"loss": 0.8206, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 0.5921371946287008, |
|
"grad_norm": 0.9019710217551041, |
|
"learning_rate": 4.268355581029523e-06, |
|
"loss": 0.8414, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.5929461252224559, |
|
"grad_norm": 0.9361559273358337, |
|
"learning_rate": 4.254389726708234e-06, |
|
"loss": 0.8388, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 0.593755055816211, |
|
"grad_norm": 0.8479002824514841, |
|
"learning_rate": 4.240429819269746e-06, |
|
"loss": 0.8151, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.594563986409966, |
|
"grad_norm": 0.8837530926010035, |
|
"learning_rate": 4.226475970056318e-06, |
|
"loss": 0.8212, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.5953729170037211, |
|
"grad_norm": 0.8660636262138016, |
|
"learning_rate": 4.212528290361893e-06, |
|
"loss": 0.8571, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.5961818475974762, |
|
"grad_norm": 0.9246667975571174, |
|
"learning_rate": 4.198586891431203e-06, |
|
"loss": 0.8397, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 0.5969907781912311, |
|
"grad_norm": 0.90386970265181, |
|
"learning_rate": 4.1846518844588906e-06, |
|
"loss": 0.8365, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.5977997087849862, |
|
"grad_norm": 0.9258635632218947, |
|
"learning_rate": 4.170723380588609e-06, |
|
"loss": 0.8403, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 0.5986086393787413, |
|
"grad_norm": 0.8741091788263792, |
|
"learning_rate": 4.156801490912153e-06, |
|
"loss": 0.822, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.5994175699724964, |
|
"grad_norm": 0.8552618808724114, |
|
"learning_rate": 4.142886326468556e-06, |
|
"loss": 0.8414, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 0.6002265005662514, |
|
"grad_norm": 0.8577590928960753, |
|
"learning_rate": 4.128977998243215e-06, |
|
"loss": 0.8234, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.6010354311600065, |
|
"grad_norm": 0.9640946415628675, |
|
"learning_rate": 4.115076617167004e-06, |
|
"loss": 0.8344, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 0.6018443617537615, |
|
"grad_norm": 0.8840143210557655, |
|
"learning_rate": 4.101182294115388e-06, |
|
"loss": 0.8211, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.6026532923475166, |
|
"grad_norm": 0.8574994142653399, |
|
"learning_rate": 4.087295139907528e-06, |
|
"loss": 0.8166, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.6034622229412716, |
|
"grad_norm": 0.8964092793127811, |
|
"learning_rate": 4.073415265305422e-06, |
|
"loss": 0.8171, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.6042711535350267, |
|
"grad_norm": 0.963355011643523, |
|
"learning_rate": 4.059542781012998e-06, |
|
"loss": 0.8223, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 0.6050800841287818, |
|
"grad_norm": 0.8678396011067857, |
|
"learning_rate": 4.045677797675242e-06, |
|
"loss": 0.8203, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.6058890147225368, |
|
"grad_norm": 0.9121282453763787, |
|
"learning_rate": 4.031820425877313e-06, |
|
"loss": 0.8219, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 0.6066979453162918, |
|
"grad_norm": 0.9216469820283191, |
|
"learning_rate": 4.017970776143662e-06, |
|
"loss": 0.8352, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.6075068759100469, |
|
"grad_norm": 0.908885616786373, |
|
"learning_rate": 4.004128958937149e-06, |
|
"loss": 0.8241, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 0.608315806503802, |
|
"grad_norm": 0.8942143952561401, |
|
"learning_rate": 3.990295084658166e-06, |
|
"loss": 0.8361, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.609124737097557, |
|
"grad_norm": 0.8730532396217757, |
|
"learning_rate": 3.976469263643748e-06, |
|
"loss": 0.8384, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 0.6099336676913121, |
|
"grad_norm": 0.9177096179325701, |
|
"learning_rate": 3.962651606166703e-06, |
|
"loss": 0.8184, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.6107425982850672, |
|
"grad_norm": 0.9852181301325248, |
|
"learning_rate": 3.948842222434728e-06, |
|
"loss": 0.817, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.6115515288788222, |
|
"grad_norm": 0.8956044101647403, |
|
"learning_rate": 3.935041222589524e-06, |
|
"loss": 0.818, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.6123604594725772, |
|
"grad_norm": 0.9733581516182005, |
|
"learning_rate": 3.9212487167059265e-06, |
|
"loss": 0.8223, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 0.6131693900663323, |
|
"grad_norm": 0.8650513303090718, |
|
"learning_rate": 3.907464814791029e-06, |
|
"loss": 0.8033, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.6139783206600874, |
|
"grad_norm": 0.99672664240277, |
|
"learning_rate": 3.893689626783294e-06, |
|
"loss": 0.836, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 0.6147872512538424, |
|
"grad_norm": 0.908345822113972, |
|
"learning_rate": 3.8799232625516884e-06, |
|
"loss": 0.8478, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.6155961818475975, |
|
"grad_norm": 0.8836902180965445, |
|
"learning_rate": 3.866165831894796e-06, |
|
"loss": 0.8235, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 0.6164051124413525, |
|
"grad_norm": 0.9264471672860939, |
|
"learning_rate": 3.852417444539953e-06, |
|
"loss": 0.8247, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.6172140430351076, |
|
"grad_norm": 0.894173445093799, |
|
"learning_rate": 3.8386782101423665e-06, |
|
"loss": 0.8235, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 0.6180229736288626, |
|
"grad_norm": 0.9438338724077253, |
|
"learning_rate": 3.824948238284238e-06, |
|
"loss": 0.8339, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.6188319042226177, |
|
"grad_norm": 0.9012049433189601, |
|
"learning_rate": 3.8112276384738973e-06, |
|
"loss": 0.8217, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.6196408348163728, |
|
"grad_norm": 0.9631072787593842, |
|
"learning_rate": 3.797516520144919e-06, |
|
"loss": 0.821, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.6204497654101279, |
|
"grad_norm": 0.8562741768165036, |
|
"learning_rate": 3.7838149926552565e-06, |
|
"loss": 0.8254, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 0.6212586960038828, |
|
"grad_norm": 0.9007576913836781, |
|
"learning_rate": 3.770123165286369e-06, |
|
"loss": 0.8187, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.6220676265976379, |
|
"grad_norm": 1.0271531293622427, |
|
"learning_rate": 3.7564411472423467e-06, |
|
"loss": 0.8358, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 0.622876557191393, |
|
"grad_norm": 0.8914948323397264, |
|
"learning_rate": 3.7427690476490462e-06, |
|
"loss": 0.8275, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.623685487785148, |
|
"grad_norm": 0.9328213480538451, |
|
"learning_rate": 3.7291069755532146e-06, |
|
"loss": 0.8167, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 0.6244944183789031, |
|
"grad_norm": 0.8892937385561739, |
|
"learning_rate": 3.7154550399216137e-06, |
|
"loss": 0.8347, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.6253033489726582, |
|
"grad_norm": 0.8798858697832553, |
|
"learning_rate": 3.7018133496401688e-06, |
|
"loss": 0.8383, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 0.6261122795664132, |
|
"grad_norm": 0.98439076252917, |
|
"learning_rate": 3.688182013513085e-06, |
|
"loss": 0.8274, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.6269212101601682, |
|
"grad_norm": 0.8370789210688374, |
|
"learning_rate": 3.674561140261983e-06, |
|
"loss": 0.8178, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.6277301407539233, |
|
"grad_norm": 0.9153894187914345, |
|
"learning_rate": 3.660950838525036e-06, |
|
"loss": 0.8258, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.6285390713476784, |
|
"grad_norm": 0.8922726078616626, |
|
"learning_rate": 3.647351216856099e-06, |
|
"loss": 0.8581, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 0.6293480019414335, |
|
"grad_norm": 0.9058364873912355, |
|
"learning_rate": 3.633762383723841e-06, |
|
"loss": 0.827, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.6301569325351885, |
|
"grad_norm": 0.8703352651889255, |
|
"learning_rate": 3.6201844475108884e-06, |
|
"loss": 0.8491, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 0.6309658631289435, |
|
"grad_norm": 0.892101387058938, |
|
"learning_rate": 3.606617516512953e-06, |
|
"loss": 0.8175, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.6317747937226986, |
|
"grad_norm": 0.9365570155133393, |
|
"learning_rate": 3.5930616989379697e-06, |
|
"loss": 0.8388, |
|
"step": 3905 |
|
}, |
|
{ |
|
"epoch": 0.6325837243164536, |
|
"grad_norm": 0.8773819774765554, |
|
"learning_rate": 3.5795171029052383e-06, |
|
"loss": 0.8294, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.6333926549102087, |
|
"grad_norm": 0.8714533378375239, |
|
"learning_rate": 3.5659838364445505e-06, |
|
"loss": 0.8346, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 0.6342015855039638, |
|
"grad_norm": 0.9825642646199596, |
|
"learning_rate": 3.552462007495338e-06, |
|
"loss": 0.8376, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.6350105160977189, |
|
"grad_norm": 0.899943511678241, |
|
"learning_rate": 3.5389517239058126e-06, |
|
"loss": 0.8251, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.6358194466914738, |
|
"grad_norm": 0.9817959591645147, |
|
"learning_rate": 3.5254530934320956e-06, |
|
"loss": 0.8332, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.6366283772852289, |
|
"grad_norm": 0.939309664119045, |
|
"learning_rate": 3.5119662237373684e-06, |
|
"loss": 0.8172, |
|
"step": 3935 |
|
}, |
|
{ |
|
"epoch": 0.637437307878984, |
|
"grad_norm": 0.9545145278463266, |
|
"learning_rate": 3.4984912223910105e-06, |
|
"loss": 0.8137, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.6382462384727391, |
|
"grad_norm": 0.9284835788915782, |
|
"learning_rate": 3.485028196867738e-06, |
|
"loss": 0.825, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 0.6390551690664941, |
|
"grad_norm": 1.0025398605998348, |
|
"learning_rate": 3.4715772545467507e-06, |
|
"loss": 0.8359, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.6398640996602492, |
|
"grad_norm": 0.9250686960694907, |
|
"learning_rate": 3.458138502710876e-06, |
|
"loss": 0.8445, |
|
"step": 3955 |
|
}, |
|
{ |
|
"epoch": 0.6406730302540042, |
|
"grad_norm": 0.885581537464598, |
|
"learning_rate": 3.444712048545711e-06, |
|
"loss": 0.8241, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.6414819608477592, |
|
"grad_norm": 0.9162743922784746, |
|
"learning_rate": 3.431297999138768e-06, |
|
"loss": 0.8178, |
|
"step": 3965 |
|
}, |
|
{ |
|
"epoch": 0.6422908914415143, |
|
"grad_norm": 0.9997360512776706, |
|
"learning_rate": 3.417896461478619e-06, |
|
"loss": 0.849, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.6430998220352694, |
|
"grad_norm": 0.9230165688594162, |
|
"learning_rate": 3.4045075424540484e-06, |
|
"loss": 0.8048, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.6439087526290245, |
|
"grad_norm": 0.8899064142475545, |
|
"learning_rate": 3.3911313488531907e-06, |
|
"loss": 0.8286, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.6447176832227794, |
|
"grad_norm": 0.8669101247237674, |
|
"learning_rate": 3.37776798736269e-06, |
|
"loss": 0.8119, |
|
"step": 3985 |
|
}, |
|
{ |
|
"epoch": 0.6455266138165345, |
|
"grad_norm": 0.9167759750434173, |
|
"learning_rate": 3.364417564566839e-06, |
|
"loss": 0.8472, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.6463355444102896, |
|
"grad_norm": 0.9521570771011725, |
|
"learning_rate": 3.351080186946736e-06, |
|
"loss": 0.8251, |
|
"step": 3995 |
|
}, |
|
{ |
|
"epoch": 0.6471444750040447, |
|
"grad_norm": 0.8382337519442962, |
|
"learning_rate": 3.3377559608794273e-06, |
|
"loss": 0.8453, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.6479534055977997, |
|
"grad_norm": 0.9439756459989987, |
|
"learning_rate": 3.3244449926370713e-06, |
|
"loss": 0.8359, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 0.6487623361915548, |
|
"grad_norm": 0.9306298379750102, |
|
"learning_rate": 3.3111473883860813e-06, |
|
"loss": 0.8196, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.6495712667853099, |
|
"grad_norm": 0.8706086854228088, |
|
"learning_rate": 3.2978632541862788e-06, |
|
"loss": 0.8252, |
|
"step": 4015 |
|
}, |
|
{ |
|
"epoch": 0.6503801973790648, |
|
"grad_norm": 0.918543174205597, |
|
"learning_rate": 3.2845926959900555e-06, |
|
"loss": 0.8194, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.6511891279728199, |
|
"grad_norm": 0.9355133643042608, |
|
"learning_rate": 3.2713358196415147e-06, |
|
"loss": 0.8279, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.651998058566575, |
|
"grad_norm": 0.8875680347636398, |
|
"learning_rate": 3.2580927308756426e-06, |
|
"loss": 0.8336, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.6528069891603301, |
|
"grad_norm": 0.974049905246242, |
|
"learning_rate": 3.2448635353174553e-06, |
|
"loss": 0.8334, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 0.6536159197540851, |
|
"grad_norm": 0.8679703990844894, |
|
"learning_rate": 3.231648338481157e-06, |
|
"loss": 0.8355, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.6544248503478401, |
|
"grad_norm": 0.9299757764874559, |
|
"learning_rate": 3.2184472457693005e-06, |
|
"loss": 0.8401, |
|
"step": 4045 |
|
}, |
|
{ |
|
"epoch": 0.6552337809415952, |
|
"grad_norm": 0.8615524033230236, |
|
"learning_rate": 3.2052603624719516e-06, |
|
"loss": 0.8217, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.6560427115353503, |
|
"grad_norm": 0.9060239627418345, |
|
"learning_rate": 3.1920877937658325e-06, |
|
"loss": 0.8173, |
|
"step": 4055 |
|
}, |
|
{ |
|
"epoch": 0.6568516421291053, |
|
"grad_norm": 0.9398327715146758, |
|
"learning_rate": 3.178929644713504e-06, |
|
"loss": 0.8412, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.6576605727228604, |
|
"grad_norm": 0.8393737863171705, |
|
"learning_rate": 3.1657860202625145e-06, |
|
"loss": 0.8368, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 0.6584695033166155, |
|
"grad_norm": 0.8802074487722802, |
|
"learning_rate": 3.1526570252445665e-06, |
|
"loss": 0.8203, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.6592784339103704, |
|
"grad_norm": 0.9133104861467479, |
|
"learning_rate": 3.1395427643746802e-06, |
|
"loss": 0.8168, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.6600873645041255, |
|
"grad_norm": 0.9090731300453533, |
|
"learning_rate": 3.1264433422503564e-06, |
|
"loss": 0.8466, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.6608962950978806, |
|
"grad_norm": 0.9229902190782812, |
|
"learning_rate": 3.113358863350747e-06, |
|
"loss": 0.8334, |
|
"step": 4085 |
|
}, |
|
{ |
|
"epoch": 0.6617052256916357, |
|
"grad_norm": 0.8999096831432444, |
|
"learning_rate": 3.100289432035818e-06, |
|
"loss": 0.8145, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.6625141562853907, |
|
"grad_norm": 0.9226844151049607, |
|
"learning_rate": 3.087235152545517e-06, |
|
"loss": 0.8245, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 0.6633230868791458, |
|
"grad_norm": 0.9923339907314755, |
|
"learning_rate": 3.074196128998943e-06, |
|
"loss": 0.8067, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.6641320174729008, |
|
"grad_norm": 0.8793470002054651, |
|
"learning_rate": 3.0611724653935184e-06, |
|
"loss": 0.8154, |
|
"step": 4105 |
|
}, |
|
{ |
|
"epoch": 0.6649409480666559, |
|
"grad_norm": 0.9952922559372246, |
|
"learning_rate": 3.048164265604152e-06, |
|
"loss": 0.8339, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.6657498786604109, |
|
"grad_norm": 0.8934191696631508, |
|
"learning_rate": 3.035171633382419e-06, |
|
"loss": 0.8126, |
|
"step": 4115 |
|
}, |
|
{ |
|
"epoch": 0.666558809254166, |
|
"grad_norm": 0.9074584444838903, |
|
"learning_rate": 3.0221946723557274e-06, |
|
"loss": 0.814, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.6673677398479211, |
|
"grad_norm": 0.8823615491806162, |
|
"learning_rate": 3.009233486026497e-06, |
|
"loss": 0.8231, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.6681766704416761, |
|
"grad_norm": 0.8985312723090128, |
|
"learning_rate": 2.9962881777713326e-06, |
|
"loss": 0.8279, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.6689856010354311, |
|
"grad_norm": 0.8759610512792171, |
|
"learning_rate": 2.983358850840187e-06, |
|
"loss": 0.8503, |
|
"step": 4135 |
|
}, |
|
{ |
|
"epoch": 0.6697945316291862, |
|
"grad_norm": 0.875080183012667, |
|
"learning_rate": 2.97044560835556e-06, |
|
"loss": 0.8361, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.6706034622229413, |
|
"grad_norm": 0.9343434705098017, |
|
"learning_rate": 2.9575485533116597e-06, |
|
"loss": 0.8422, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 0.6714123928166963, |
|
"grad_norm": 0.9509754443485635, |
|
"learning_rate": 2.944667788573584e-06, |
|
"loss": 0.824, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.6722213234104514, |
|
"grad_norm": 0.8941436073858621, |
|
"learning_rate": 2.9318034168765048e-06, |
|
"loss": 0.8329, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 0.6730302540042065, |
|
"grad_norm": 0.9216572906600096, |
|
"learning_rate": 2.9189555408248436e-06, |
|
"loss": 0.838, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.6738391845979615, |
|
"grad_norm": 0.9049144748685041, |
|
"learning_rate": 2.906124262891451e-06, |
|
"loss": 0.8392, |
|
"step": 4165 |
|
}, |
|
{ |
|
"epoch": 0.6746481151917165, |
|
"grad_norm": 0.9088265471558502, |
|
"learning_rate": 2.893309685416802e-06, |
|
"loss": 0.8219, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.6754570457854716, |
|
"grad_norm": 0.958093818584568, |
|
"learning_rate": 2.880511910608164e-06, |
|
"loss": 0.8361, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.6762659763792267, |
|
"grad_norm": 0.9461864165794798, |
|
"learning_rate": 2.8677310405387926e-06, |
|
"loss": 0.8145, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.6770749069729817, |
|
"grad_norm": 0.8862588950970776, |
|
"learning_rate": 2.854967177147113e-06, |
|
"loss": 0.8101, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 0.6778838375667368, |
|
"grad_norm": 0.95284924784784, |
|
"learning_rate": 2.8422204222359027e-06, |
|
"loss": 0.8355, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.6786927681604918, |
|
"grad_norm": 0.9473709868364139, |
|
"learning_rate": 2.829490877471491e-06, |
|
"loss": 0.8213, |
|
"step": 4195 |
|
}, |
|
{ |
|
"epoch": 0.6795016987542469, |
|
"grad_norm": 0.8870378301848979, |
|
"learning_rate": 2.816778644382938e-06, |
|
"loss": 0.8143, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.6803106293480019, |
|
"grad_norm": 0.9018325718849391, |
|
"learning_rate": 2.804083824361229e-06, |
|
"loss": 0.8265, |
|
"step": 4205 |
|
}, |
|
{ |
|
"epoch": 0.681119559941757, |
|
"grad_norm": 0.8727955391410293, |
|
"learning_rate": 2.7914065186584637e-06, |
|
"loss": 0.8254, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.6819284905355121, |
|
"grad_norm": 1.0739665165244483, |
|
"learning_rate": 2.778746828387058e-06, |
|
"loss": 0.82, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 0.6827374211292672, |
|
"grad_norm": 0.9350732250295917, |
|
"learning_rate": 2.766104854518915e-06, |
|
"loss": 0.8391, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.6835463517230221, |
|
"grad_norm": 1.0087459108347565, |
|
"learning_rate": 2.753480697884647e-06, |
|
"loss": 0.8191, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.6843552823167772, |
|
"grad_norm": 0.9139563110633456, |
|
"learning_rate": 2.740874459172752e-06, |
|
"loss": 0.8148, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.6851642129105323, |
|
"grad_norm": 0.8530060707424758, |
|
"learning_rate": 2.728286238928821e-06, |
|
"loss": 0.8053, |
|
"step": 4235 |
|
}, |
|
{ |
|
"epoch": 0.6859731435042873, |
|
"grad_norm": 0.8777686426009732, |
|
"learning_rate": 2.7157161375547304e-06, |
|
"loss": 0.8114, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.6867820740980424, |
|
"grad_norm": 0.897912961167507, |
|
"learning_rate": 2.7031642553078374e-06, |
|
"loss": 0.8418, |
|
"step": 4245 |
|
}, |
|
{ |
|
"epoch": 0.6875910046917975, |
|
"grad_norm": 0.9279693202641751, |
|
"learning_rate": 2.690630692300199e-06, |
|
"loss": 0.8404, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.6883999352855525, |
|
"grad_norm": 0.9248453334693622, |
|
"learning_rate": 2.6781155484977495e-06, |
|
"loss": 0.8494, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 0.6892088658793075, |
|
"grad_norm": 0.9396606484969379, |
|
"learning_rate": 2.6656189237195186e-06, |
|
"loss": 0.8226, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.6900177964730626, |
|
"grad_norm": 0.8634815816211927, |
|
"learning_rate": 2.6531409176368296e-06, |
|
"loss": 0.8265, |
|
"step": 4265 |
|
}, |
|
{ |
|
"epoch": 0.6908267270668177, |
|
"grad_norm": 0.9135686768769312, |
|
"learning_rate": 2.6406816297725086e-06, |
|
"loss": 0.8248, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.6916356576605728, |
|
"grad_norm": 0.92902484349033, |
|
"learning_rate": 2.628241159500081e-06, |
|
"loss": 0.8173, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.6924445882543278, |
|
"grad_norm": 0.8687123389790895, |
|
"learning_rate": 2.6158196060429926e-06, |
|
"loss": 0.8398, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.6932535188480828, |
|
"grad_norm": 0.9217730078223995, |
|
"learning_rate": 2.6034170684738065e-06, |
|
"loss": 0.8309, |
|
"step": 4285 |
|
}, |
|
{ |
|
"epoch": 0.6940624494418379, |
|
"grad_norm": 0.9769331954136037, |
|
"learning_rate": 2.591033645713424e-06, |
|
"loss": 0.8223, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.6948713800355929, |
|
"grad_norm": 0.908044715178041, |
|
"learning_rate": 2.5786694365302855e-06, |
|
"loss": 0.8281, |
|
"step": 4295 |
|
}, |
|
{ |
|
"epoch": 0.695680310629348, |
|
"grad_norm": 0.9111114311128146, |
|
"learning_rate": 2.566324539539583e-06, |
|
"loss": 0.8266, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.6964892412231031, |
|
"grad_norm": 0.9303322655983225, |
|
"learning_rate": 2.5539990532024827e-06, |
|
"loss": 0.8111, |
|
"step": 4305 |
|
}, |
|
{ |
|
"epoch": 0.6972981718168582, |
|
"grad_norm": 0.8418692144813529, |
|
"learning_rate": 2.5416930758253317e-06, |
|
"loss": 0.8256, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.6981071024106131, |
|
"grad_norm": 0.9992022955905648, |
|
"learning_rate": 2.5294067055588765e-06, |
|
"loss": 0.8032, |
|
"step": 4315 |
|
}, |
|
{ |
|
"epoch": 0.6989160330043682, |
|
"grad_norm": 0.8813775901696349, |
|
"learning_rate": 2.517140040397482e-06, |
|
"loss": 0.8168, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.6997249635981233, |
|
"grad_norm": 0.9034982645101618, |
|
"learning_rate": 2.5048931781783457e-06, |
|
"loss": 0.8363, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.7005338941918783, |
|
"grad_norm": 0.9291076912208419, |
|
"learning_rate": 2.492666216580719e-06, |
|
"loss": 0.8309, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.7013428247856334, |
|
"grad_norm": 0.9136441839215993, |
|
"learning_rate": 2.480459253125132e-06, |
|
"loss": 0.8446, |
|
"step": 4335 |
|
}, |
|
{ |
|
"epoch": 0.7021517553793885, |
|
"grad_norm": 0.8945013641006528, |
|
"learning_rate": 2.468272385172609e-06, |
|
"loss": 0.8061, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.7029606859731435, |
|
"grad_norm": 0.9843334357090001, |
|
"learning_rate": 2.4561057099238973e-06, |
|
"loss": 0.8245, |
|
"step": 4345 |
|
}, |
|
{ |
|
"epoch": 0.7037696165668985, |
|
"grad_norm": 0.9061201723841746, |
|
"learning_rate": 2.4439593244186914e-06, |
|
"loss": 0.8156, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.7045785471606536, |
|
"grad_norm": 0.9595180789878346, |
|
"learning_rate": 2.4318333255348524e-06, |
|
"loss": 0.8362, |
|
"step": 4355 |
|
}, |
|
{ |
|
"epoch": 0.7053874777544087, |
|
"grad_norm": 0.8905904032697742, |
|
"learning_rate": 2.4197278099876458e-06, |
|
"loss": 0.8013, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.7061964083481638, |
|
"grad_norm": 0.9061698973868614, |
|
"learning_rate": 2.407642874328961e-06, |
|
"loss": 0.805, |
|
"step": 4365 |
|
}, |
|
{ |
|
"epoch": 0.7070053389419187, |
|
"grad_norm": 0.9192998017239877, |
|
"learning_rate": 2.3955786149465505e-06, |
|
"loss": 0.8445, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.7078142695356738, |
|
"grad_norm": 1.0236876517448235, |
|
"learning_rate": 2.3835351280632514e-06, |
|
"loss": 0.8484, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.7086232001294289, |
|
"grad_norm": 0.9092749965225744, |
|
"learning_rate": 2.3715125097362246e-06, |
|
"loss": 0.8134, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.7094321307231839, |
|
"grad_norm": 0.8913146454343571, |
|
"learning_rate": 2.3595108558561814e-06, |
|
"loss": 0.8231, |
|
"step": 4385 |
|
}, |
|
{ |
|
"epoch": 0.710241061316939, |
|
"grad_norm": 0.9163485544067049, |
|
"learning_rate": 2.347530262146629e-06, |
|
"loss": 0.8227, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.7110499919106941, |
|
"grad_norm": 0.9692923072785117, |
|
"learning_rate": 2.3355708241630998e-06, |
|
"loss": 0.8114, |
|
"step": 4395 |
|
}, |
|
{ |
|
"epoch": 0.7118589225044492, |
|
"grad_norm": 0.9952307777966306, |
|
"learning_rate": 2.3236326372923913e-06, |
|
"loss": 0.8265, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.7126678530982041, |
|
"grad_norm": 0.8908142322793587, |
|
"learning_rate": 2.3117157967518052e-06, |
|
"loss": 0.8306, |
|
"step": 4405 |
|
}, |
|
{ |
|
"epoch": 0.7134767836919592, |
|
"grad_norm": 0.9192836496709301, |
|
"learning_rate": 2.299820397588387e-06, |
|
"loss": 0.8104, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 0.90827967789888, |
|
"learning_rate": 2.2879465346781703e-06, |
|
"loss": 0.8106, |
|
"step": 4415 |
|
}, |
|
{ |
|
"epoch": 0.7150946448794694, |
|
"grad_norm": 0.9019495653500154, |
|
"learning_rate": 2.2760943027254168e-06, |
|
"loss": 0.8162, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.7159035754732244, |
|
"grad_norm": 0.9225860245555908, |
|
"learning_rate": 2.264263796261864e-06, |
|
"loss": 0.8385, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.7167125060669794, |
|
"grad_norm": 0.9905684037464217, |
|
"learning_rate": 2.2524551096459703e-06, |
|
"loss": 0.836, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.7175214366607345, |
|
"grad_norm": 0.9642215008382582, |
|
"learning_rate": 2.240668337062162e-06, |
|
"loss": 0.8382, |
|
"step": 4435 |
|
}, |
|
{ |
|
"epoch": 0.7183303672544895, |
|
"grad_norm": 1.02786094199622, |
|
"learning_rate": 2.228903572520076e-06, |
|
"loss": 0.8106, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.7191392978482446, |
|
"grad_norm": 0.885661361219246, |
|
"learning_rate": 2.2171609098538275e-06, |
|
"loss": 0.8363, |
|
"step": 4445 |
|
}, |
|
{ |
|
"epoch": 0.7199482284419997, |
|
"grad_norm": 0.9432094404967053, |
|
"learning_rate": 2.2054404427212427e-06, |
|
"loss": 0.8295, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.7207571590357548, |
|
"grad_norm": 0.8867439139171059, |
|
"learning_rate": 2.1937422646031216e-06, |
|
"loss": 0.826, |
|
"step": 4455 |
|
}, |
|
{ |
|
"epoch": 0.7215660896295097, |
|
"grad_norm": 0.8799592291488351, |
|
"learning_rate": 2.1820664688024885e-06, |
|
"loss": 0.808, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.7223750202232648, |
|
"grad_norm": 0.9187111254961942, |
|
"learning_rate": 2.1704131484438523e-06, |
|
"loss": 0.8284, |
|
"step": 4465 |
|
}, |
|
{ |
|
"epoch": 0.7231839508170199, |
|
"grad_norm": 0.8585062008980405, |
|
"learning_rate": 2.1587823964724564e-06, |
|
"loss": 0.8333, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.723992881410775, |
|
"grad_norm": 0.8979409403030056, |
|
"learning_rate": 2.1471743056535455e-06, |
|
"loss": 0.8401, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.72480181200453, |
|
"grad_norm": 0.9181507434036461, |
|
"learning_rate": 2.1355889685716225e-06, |
|
"loss": 0.8209, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.7256107425982851, |
|
"grad_norm": 0.942590732475379, |
|
"learning_rate": 2.124026477629706e-06, |
|
"loss": 0.8066, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 0.7264196731920401, |
|
"grad_norm": 0.9429750289429594, |
|
"learning_rate": 2.1124869250486053e-06, |
|
"loss": 0.8224, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.7272286037857951, |
|
"grad_norm": 0.9542611725247429, |
|
"learning_rate": 2.1009704028661643e-06, |
|
"loss": 0.8384, |
|
"step": 4495 |
|
}, |
|
{ |
|
"epoch": 0.7280375343795502, |
|
"grad_norm": 0.9040683126789256, |
|
"learning_rate": 2.08947700293655e-06, |
|
"loss": 0.8418, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.7288464649733053, |
|
"grad_norm": 0.9481483200578114, |
|
"learning_rate": 2.078006816929503e-06, |
|
"loss": 0.8144, |
|
"step": 4505 |
|
}, |
|
{ |
|
"epoch": 0.7296553955670604, |
|
"grad_norm": 0.8940640126195414, |
|
"learning_rate": 2.066559936329618e-06, |
|
"loss": 0.8059, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.7304643261608154, |
|
"grad_norm": 0.9261592631490639, |
|
"learning_rate": 2.0551364524356054e-06, |
|
"loss": 0.8261, |
|
"step": 4515 |
|
}, |
|
{ |
|
"epoch": 0.7312732567545704, |
|
"grad_norm": 0.896600436452953, |
|
"learning_rate": 2.0437364563595686e-06, |
|
"loss": 0.8437, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.7320821873483255, |
|
"grad_norm": 0.8919258636639074, |
|
"learning_rate": 2.0323600390262743e-06, |
|
"loss": 0.8226, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.7328911179420806, |
|
"grad_norm": 0.9826507277779656, |
|
"learning_rate": 2.0210072911724293e-06, |
|
"loss": 0.8215, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.7337000485358356, |
|
"grad_norm": 0.9182603432302395, |
|
"learning_rate": 2.0096783033459564e-06, |
|
"loss": 0.804, |
|
"step": 4535 |
|
}, |
|
{ |
|
"epoch": 0.7345089791295907, |
|
"grad_norm": 0.904398523841377, |
|
"learning_rate": 1.99837316590527e-06, |
|
"loss": 0.8385, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.7353179097233458, |
|
"grad_norm": 0.8989331803375635, |
|
"learning_rate": 1.987091969018561e-06, |
|
"loss": 0.8225, |
|
"step": 4545 |
|
}, |
|
{ |
|
"epoch": 0.7361268403171007, |
|
"grad_norm": 0.8975780407900149, |
|
"learning_rate": 1.9758348026630663e-06, |
|
"loss": 0.8388, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.7369357709108558, |
|
"grad_norm": 0.9121042226321563, |
|
"learning_rate": 1.964601756624366e-06, |
|
"loss": 0.8215, |
|
"step": 4555 |
|
}, |
|
{ |
|
"epoch": 0.7377447015046109, |
|
"grad_norm": 0.9214704840321436, |
|
"learning_rate": 1.9533929204956553e-06, |
|
"loss": 0.8142, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.738553632098366, |
|
"grad_norm": 0.9189755165680765, |
|
"learning_rate": 1.9422083836770405e-06, |
|
"loss": 0.8307, |
|
"step": 4565 |
|
}, |
|
{ |
|
"epoch": 0.739362562692121, |
|
"grad_norm": 0.8679073250223568, |
|
"learning_rate": 1.9310482353748146e-06, |
|
"loss": 0.8118, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.7401714932858761, |
|
"grad_norm": 0.9016222202190433, |
|
"learning_rate": 1.9199125646007533e-06, |
|
"loss": 0.8043, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.7409804238796311, |
|
"grad_norm": 0.9105339155753437, |
|
"learning_rate": 1.9088014601713993e-06, |
|
"loss": 0.8376, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.7417893544733862, |
|
"grad_norm": 1.005707657930975, |
|
"learning_rate": 1.8977150107073632e-06, |
|
"loss": 0.8219, |
|
"step": 4585 |
|
}, |
|
{ |
|
"epoch": 0.7425982850671412, |
|
"grad_norm": 0.8230187430989078, |
|
"learning_rate": 1.8866533046326086e-06, |
|
"loss": 0.839, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.7434072156608963, |
|
"grad_norm": 0.9054157125777134, |
|
"learning_rate": 1.8756164301737478e-06, |
|
"loss": 0.8215, |
|
"step": 4595 |
|
}, |
|
{ |
|
"epoch": 0.7442161462546514, |
|
"grad_norm": 0.9386133499303263, |
|
"learning_rate": 1.8646044753593429e-06, |
|
"loss": 0.8139, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.7450250768484064, |
|
"grad_norm": 0.8974275889331189, |
|
"learning_rate": 1.8536175280191971e-06, |
|
"loss": 0.8269, |
|
"step": 4605 |
|
}, |
|
{ |
|
"epoch": 0.7458340074421614, |
|
"grad_norm": 0.9291872406905795, |
|
"learning_rate": 1.8426556757836594e-06, |
|
"loss": 0.8104, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.7466429380359165, |
|
"grad_norm": 0.921360125713896, |
|
"learning_rate": 1.8317190060829242e-06, |
|
"loss": 0.8078, |
|
"step": 4615 |
|
}, |
|
{ |
|
"epoch": 0.7474518686296716, |
|
"grad_norm": 0.9154624659332359, |
|
"learning_rate": 1.820807606146332e-06, |
|
"loss": 0.8228, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.7482607992234266, |
|
"grad_norm": 0.9910039258360209, |
|
"learning_rate": 1.8099215630016759e-06, |
|
"loss": 0.8162, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.7490697298171817, |
|
"grad_norm": 0.9527574985037581, |
|
"learning_rate": 1.7990609634745087e-06, |
|
"loss": 0.8206, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.7498786604109368, |
|
"grad_norm": 0.8617520277776372, |
|
"learning_rate": 1.788225894187443e-06, |
|
"loss": 0.82, |
|
"step": 4635 |
|
}, |
|
{ |
|
"epoch": 0.7506875910046918, |
|
"grad_norm": 1.0275065944999713, |
|
"learning_rate": 1.7774164415594692e-06, |
|
"loss": 0.8474, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.7514965215984468, |
|
"grad_norm": 1.0046608971251627, |
|
"learning_rate": 1.7666326918052667e-06, |
|
"loss": 0.8144, |
|
"step": 4645 |
|
}, |
|
{ |
|
"epoch": 0.7523054521922019, |
|
"grad_norm": 0.9234744279179873, |
|
"learning_rate": 1.7558747309345075e-06, |
|
"loss": 0.8258, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.753114382785957, |
|
"grad_norm": 0.987941691115384, |
|
"learning_rate": 1.7451426447511771e-06, |
|
"loss": 0.8161, |
|
"step": 4655 |
|
}, |
|
{ |
|
"epoch": 0.753923313379712, |
|
"grad_norm": 1.0005046357018823, |
|
"learning_rate": 1.7344365188528844e-06, |
|
"loss": 0.817, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.754732243973467, |
|
"grad_norm": 0.9561730998580897, |
|
"learning_rate": 1.7237564386301869e-06, |
|
"loss": 0.8088, |
|
"step": 4665 |
|
}, |
|
{ |
|
"epoch": 0.7555411745672221, |
|
"grad_norm": 0.9447102620547789, |
|
"learning_rate": 1.7131024892659048e-06, |
|
"loss": 0.8118, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.7563501051609772, |
|
"grad_norm": 0.9539633697057311, |
|
"learning_rate": 1.7024747557344411e-06, |
|
"loss": 0.844, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.7571590357547322, |
|
"grad_norm": 1.0424885395489862, |
|
"learning_rate": 1.6918733228011058e-06, |
|
"loss": 0.8315, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.7579679663484873, |
|
"grad_norm": 0.9458202947476442, |
|
"learning_rate": 1.6812982750214385e-06, |
|
"loss": 0.8304, |
|
"step": 4685 |
|
}, |
|
{ |
|
"epoch": 0.7587768969422424, |
|
"grad_norm": 0.918182742402279, |
|
"learning_rate": 1.6707496967405347e-06, |
|
"loss": 0.8115, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.7595858275359975, |
|
"grad_norm": 0.9291030646393732, |
|
"learning_rate": 1.6602276720923732e-06, |
|
"loss": 0.8163, |
|
"step": 4695 |
|
}, |
|
{ |
|
"epoch": 0.7603947581297524, |
|
"grad_norm": 0.9588497352744614, |
|
"learning_rate": 1.6497322849991443e-06, |
|
"loss": 0.8266, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.7612036887235075, |
|
"grad_norm": 0.9290169660288218, |
|
"learning_rate": 1.6392636191705818e-06, |
|
"loss": 0.8231, |
|
"step": 4705 |
|
}, |
|
{ |
|
"epoch": 0.7620126193172626, |
|
"grad_norm": 0.9469265731424403, |
|
"learning_rate": 1.6288217581032945e-06, |
|
"loss": 0.8283, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.7628215499110176, |
|
"grad_norm": 0.9739835381552143, |
|
"learning_rate": 1.618406785080095e-06, |
|
"loss": 0.8358, |
|
"step": 4715 |
|
}, |
|
{ |
|
"epoch": 0.7636304805047727, |
|
"grad_norm": 0.9140268328463288, |
|
"learning_rate": 1.60801878316935e-06, |
|
"loss": 0.8173, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.7644394110985278, |
|
"grad_norm": 0.9519079424417303, |
|
"learning_rate": 1.5976578352243015e-06, |
|
"loss": 0.8013, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.7652483416922828, |
|
"grad_norm": 1.0216658413405784, |
|
"learning_rate": 1.587324023882415e-06, |
|
"loss": 0.827, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.7660572722860378, |
|
"grad_norm": 0.981342150260489, |
|
"learning_rate": 1.5770174315647185e-06, |
|
"loss": 0.8172, |
|
"step": 4735 |
|
}, |
|
{ |
|
"epoch": 0.7668662028797929, |
|
"grad_norm": 0.9841540010061296, |
|
"learning_rate": 1.566738140475146e-06, |
|
"loss": 0.8298, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.767675133473548, |
|
"grad_norm": 0.9647924689522289, |
|
"learning_rate": 1.5564862325998754e-06, |
|
"loss": 0.8349, |
|
"step": 4745 |
|
}, |
|
{ |
|
"epoch": 0.7684840640673031, |
|
"grad_norm": 0.9071693118539406, |
|
"learning_rate": 1.5462617897066863e-06, |
|
"loss": 0.8266, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.769292994661058, |
|
"grad_norm": 0.9672024055908947, |
|
"learning_rate": 1.536064893344298e-06, |
|
"loss": 0.8062, |
|
"step": 4755 |
|
}, |
|
{ |
|
"epoch": 0.7701019252548131, |
|
"grad_norm": 0.9479852003456001, |
|
"learning_rate": 1.5258956248417228e-06, |
|
"loss": 0.8171, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.7709108558485682, |
|
"grad_norm": 1.0847747231670404, |
|
"learning_rate": 1.515754065307622e-06, |
|
"loss": 0.8343, |
|
"step": 4765 |
|
}, |
|
{ |
|
"epoch": 0.7717197864423232, |
|
"grad_norm": 0.8930428867327512, |
|
"learning_rate": 1.505640295629645e-06, |
|
"loss": 0.8166, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.7725287170360783, |
|
"grad_norm": 0.9689746816489254, |
|
"learning_rate": 1.4955543964738e-06, |
|
"loss": 0.8357, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.7733376476298334, |
|
"grad_norm": 0.9559994524411303, |
|
"learning_rate": 1.4854964482838014e-06, |
|
"loss": 0.8497, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.7741465782235885, |
|
"grad_norm": 0.9337905415707888, |
|
"learning_rate": 1.4754665312804311e-06, |
|
"loss": 0.8312, |
|
"step": 4785 |
|
}, |
|
{ |
|
"epoch": 0.7749555088173434, |
|
"grad_norm": 0.9273467544337363, |
|
"learning_rate": 1.4654647254608988e-06, |
|
"loss": 0.8234, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.7757644394110985, |
|
"grad_norm": 0.9792032991670162, |
|
"learning_rate": 1.4554911105982022e-06, |
|
"loss": 0.8468, |
|
"step": 4795 |
|
}, |
|
{ |
|
"epoch": 0.7765733700048536, |
|
"grad_norm": 0.9045240623424784, |
|
"learning_rate": 1.445545766240492e-06, |
|
"loss": 0.8315, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.7773823005986087, |
|
"grad_norm": 0.9547782673298132, |
|
"learning_rate": 1.4356287717104384e-06, |
|
"loss": 0.8259, |
|
"step": 4805 |
|
}, |
|
{ |
|
"epoch": 0.7781912311923637, |
|
"grad_norm": 1.0002987188424177, |
|
"learning_rate": 1.4257402061045966e-06, |
|
"loss": 0.8242, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.7790001617861187, |
|
"grad_norm": 0.9750841170635594, |
|
"learning_rate": 1.4158801482927764e-06, |
|
"loss": 0.8253, |
|
"step": 4815 |
|
}, |
|
{ |
|
"epoch": 0.7798090923798738, |
|
"grad_norm": 1.0204763580485683, |
|
"learning_rate": 1.4060486769174158e-06, |
|
"loss": 0.8124, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.7806180229736288, |
|
"grad_norm": 0.9140777892397043, |
|
"learning_rate": 1.396245870392946e-06, |
|
"loss": 0.8209, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.7814269535673839, |
|
"grad_norm": 0.9326381455666519, |
|
"learning_rate": 1.3864718069051765e-06, |
|
"loss": 0.8085, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.782235884161139, |
|
"grad_norm": 0.8927418792043738, |
|
"learning_rate": 1.376726564410663e-06, |
|
"loss": 0.8281, |
|
"step": 4835 |
|
}, |
|
{ |
|
"epoch": 0.7830448147548941, |
|
"grad_norm": 0.9878798053606094, |
|
"learning_rate": 1.3670102206360936e-06, |
|
"loss": 0.8393, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.783853745348649, |
|
"grad_norm": 0.9345336173306545, |
|
"learning_rate": 1.3573228530776605e-06, |
|
"loss": 0.8044, |
|
"step": 4845 |
|
}, |
|
{ |
|
"epoch": 0.7846626759424041, |
|
"grad_norm": 0.9721628031873526, |
|
"learning_rate": 1.3476645390004473e-06, |
|
"loss": 0.8386, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.7854716065361592, |
|
"grad_norm": 0.915209628149697, |
|
"learning_rate": 1.3380353554378074e-06, |
|
"loss": 0.8357, |
|
"step": 4855 |
|
}, |
|
{ |
|
"epoch": 0.7862805371299143, |
|
"grad_norm": 0.9820722069486887, |
|
"learning_rate": 1.3284353791907584e-06, |
|
"loss": 0.8333, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.7870894677236693, |
|
"grad_norm": 0.9732881075592806, |
|
"learning_rate": 1.3188646868273615e-06, |
|
"loss": 0.8239, |
|
"step": 4865 |
|
}, |
|
{ |
|
"epoch": 0.7878983983174244, |
|
"grad_norm": 0.8961967498766817, |
|
"learning_rate": 1.3093233546821143e-06, |
|
"loss": 0.824, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.7887073289111795, |
|
"grad_norm": 0.9052548621214841, |
|
"learning_rate": 1.2998114588553429e-06, |
|
"loss": 0.8267, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.7895162595049344, |
|
"grad_norm": 1.0580362890932773, |
|
"learning_rate": 1.2903290752125914e-06, |
|
"loss": 0.8218, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.7903251900986895, |
|
"grad_norm": 0.853800216222736, |
|
"learning_rate": 1.28087627938402e-06, |
|
"loss": 0.8166, |
|
"step": 4885 |
|
}, |
|
{ |
|
"epoch": 0.7911341206924446, |
|
"grad_norm": 0.9652580037959974, |
|
"learning_rate": 1.2714531467637998e-06, |
|
"loss": 0.8238, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.7919430512861997, |
|
"grad_norm": 0.8842437766967849, |
|
"learning_rate": 1.2620597525095135e-06, |
|
"loss": 0.817, |
|
"step": 4895 |
|
}, |
|
{ |
|
"epoch": 0.7927519818799547, |
|
"grad_norm": 0.9258752585803977, |
|
"learning_rate": 1.2526961715415542e-06, |
|
"loss": 0.8279, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.7935609124737097, |
|
"grad_norm": 0.8928563496158387, |
|
"learning_rate": 1.2433624785425291e-06, |
|
"loss": 0.8162, |
|
"step": 4905 |
|
}, |
|
{ |
|
"epoch": 0.7943698430674648, |
|
"grad_norm": 0.982000276972604, |
|
"learning_rate": 1.2340587479566597e-06, |
|
"loss": 0.8356, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.7951787736612199, |
|
"grad_norm": 0.9282259377376256, |
|
"learning_rate": 1.2247850539891947e-06, |
|
"loss": 0.8253, |
|
"step": 4915 |
|
}, |
|
{ |
|
"epoch": 0.7959877042549749, |
|
"grad_norm": 0.9525465075992243, |
|
"learning_rate": 1.2155414706058172e-06, |
|
"loss": 0.8169, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.79679663484873, |
|
"grad_norm": 0.9231085182420589, |
|
"learning_rate": 1.206328071532048e-06, |
|
"loss": 0.8138, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.7976055654424851, |
|
"grad_norm": 0.9253319119591827, |
|
"learning_rate": 1.197144930252665e-06, |
|
"loss": 0.8225, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.79841449603624, |
|
"grad_norm": 0.9002303313112969, |
|
"learning_rate": 1.187992120011111e-06, |
|
"loss": 0.8291, |
|
"step": 4935 |
|
}, |
|
{ |
|
"epoch": 0.7992234266299951, |
|
"grad_norm": 0.9376615616843144, |
|
"learning_rate": 1.178869713808916e-06, |
|
"loss": 0.8341, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.8000323572237502, |
|
"grad_norm": 0.9698576047423542, |
|
"learning_rate": 1.1697777844051105e-06, |
|
"loss": 0.8147, |
|
"step": 4945 |
|
}, |
|
{ |
|
"epoch": 0.8008412878175053, |
|
"grad_norm": 0.9273363034177441, |
|
"learning_rate": 1.1607164043156455e-06, |
|
"loss": 0.8152, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.8016502184112603, |
|
"grad_norm": 1.0253181360844876, |
|
"learning_rate": 1.1516856458128167e-06, |
|
"loss": 0.8084, |
|
"step": 4955 |
|
}, |
|
{ |
|
"epoch": 0.8024591490050154, |
|
"grad_norm": 0.9400581889744581, |
|
"learning_rate": 1.1426855809246846e-06, |
|
"loss": 0.8232, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.8032680795987704, |
|
"grad_norm": 0.9017439278181727, |
|
"learning_rate": 1.133716281434502e-06, |
|
"loss": 0.8212, |
|
"step": 4965 |
|
}, |
|
{ |
|
"epoch": 0.8040770101925255, |
|
"grad_norm": 0.9129823712372367, |
|
"learning_rate": 1.1247778188801428e-06, |
|
"loss": 0.8214, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.8048859407862805, |
|
"grad_norm": 0.9288433560645074, |
|
"learning_rate": 1.1158702645535285e-06, |
|
"loss": 0.8474, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.8056948713800356, |
|
"grad_norm": 0.9211800178241987, |
|
"learning_rate": 1.1069936895000605e-06, |
|
"loss": 0.8087, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.8065038019737907, |
|
"grad_norm": 1.0542799666763112, |
|
"learning_rate": 1.0981481645180563e-06, |
|
"loss": 0.842, |
|
"step": 4985 |
|
}, |
|
{ |
|
"epoch": 0.8073127325675457, |
|
"grad_norm": 0.9078222972795535, |
|
"learning_rate": 1.0893337601581766e-06, |
|
"loss": 0.8171, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.8081216631613007, |
|
"grad_norm": 0.8965670623934641, |
|
"learning_rate": 1.0805505467228762e-06, |
|
"loss": 0.8106, |
|
"step": 4995 |
|
}, |
|
{ |
|
"epoch": 0.8089305937550558, |
|
"grad_norm": 0.9618251738798234, |
|
"learning_rate": 1.0717985942658299e-06, |
|
"loss": 0.8146, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.8097395243488109, |
|
"grad_norm": 0.950502760029193, |
|
"learning_rate": 1.063077972591382e-06, |
|
"loss": 0.8215, |
|
"step": 5005 |
|
}, |
|
{ |
|
"epoch": 0.8105484549425659, |
|
"grad_norm": 0.9493168241113547, |
|
"learning_rate": 1.0543887512539851e-06, |
|
"loss": 0.8194, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.811357385536321, |
|
"grad_norm": 0.8823063979936068, |
|
"learning_rate": 1.0457309995576498e-06, |
|
"loss": 0.8081, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 0.8121663161300761, |
|
"grad_norm": 0.9596843704171958, |
|
"learning_rate": 1.0371047865553847e-06, |
|
"loss": 0.7973, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.8129752467238311, |
|
"grad_norm": 0.9284122972047167, |
|
"learning_rate": 1.0285101810486535e-06, |
|
"loss": 0.8468, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.8137841773175861, |
|
"grad_norm": 0.922908519391859, |
|
"learning_rate": 1.0199472515868235e-06, |
|
"loss": 0.8602, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.8145931079113412, |
|
"grad_norm": 0.9150347591949571, |
|
"learning_rate": 1.0114160664666156e-06, |
|
"loss": 0.8285, |
|
"step": 5035 |
|
}, |
|
{ |
|
"epoch": 0.8154020385050963, |
|
"grad_norm": 0.961458868733065, |
|
"learning_rate": 1.0029166937315682e-06, |
|
"loss": 0.8261, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.8162109690988513, |
|
"grad_norm": 0.9795182791817205, |
|
"learning_rate": 9.94449201171479e-07, |
|
"loss": 0.8077, |
|
"step": 5045 |
|
}, |
|
{ |
|
"epoch": 0.8170198996926064, |
|
"grad_norm": 0.9712437822816266, |
|
"learning_rate": 9.86013656321882e-07, |
|
"loss": 0.8136, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.8178288302863614, |
|
"grad_norm": 0.9243414819653508, |
|
"learning_rate": 9.77610126463497e-07, |
|
"loss": 0.8116, |
|
"step": 5055 |
|
}, |
|
{ |
|
"epoch": 0.8186377608801165, |
|
"grad_norm": 0.8953653000547861, |
|
"learning_rate": 9.69238678621698e-07, |
|
"loss": 0.8052, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.8194466914738715, |
|
"grad_norm": 0.9571083425955552, |
|
"learning_rate": 9.608993795659766e-07, |
|
"loss": 0.8361, |
|
"step": 5065 |
|
}, |
|
{ |
|
"epoch": 0.8202556220676266, |
|
"grad_norm": 0.8812425656788687, |
|
"learning_rate": 9.525922958094113e-07, |
|
"loss": 0.8091, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.8210645526613817, |
|
"grad_norm": 0.9769325986983122, |
|
"learning_rate": 9.443174936081345e-07, |
|
"loss": 0.8384, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.8218734832551368, |
|
"grad_norm": 0.9429765500848418, |
|
"learning_rate": 9.360750389608069e-07, |
|
"loss": 0.823, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.8226824138488917, |
|
"grad_norm": 0.921228115104811, |
|
"learning_rate": 9.278649976080889e-07, |
|
"loss": 0.8331, |
|
"step": 5085 |
|
}, |
|
{ |
|
"epoch": 0.8234913444426468, |
|
"grad_norm": 0.9819735566785222, |
|
"learning_rate": 9.196874350321161e-07, |
|
"loss": 0.829, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.8243002750364019, |
|
"grad_norm": 0.8804316419490247, |
|
"learning_rate": 9.11542416455981e-07, |
|
"loss": 0.8253, |
|
"step": 5095 |
|
}, |
|
{ |
|
"epoch": 0.8251092056301569, |
|
"grad_norm": 0.9309408261176141, |
|
"learning_rate": 9.034300068432045e-07, |
|
"loss": 0.8294, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.825918136223912, |
|
"grad_norm": 0.8366500179121683, |
|
"learning_rate": 8.953502708972279e-07, |
|
"loss": 0.8309, |
|
"step": 5105 |
|
}, |
|
{ |
|
"epoch": 0.8267270668176671, |
|
"grad_norm": 0.9984582201604612, |
|
"learning_rate": 8.873032730608883e-07, |
|
"loss": 0.8288, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.8275359974114221, |
|
"grad_norm": 1.0257992816612576, |
|
"learning_rate": 8.792890775159124e-07, |
|
"loss": 0.832, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 0.8283449280051771, |
|
"grad_norm": 0.9357127400993555, |
|
"learning_rate": 8.713077481823978e-07, |
|
"loss": 0.8377, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.8291538585989322, |
|
"grad_norm": 1.042951179385166, |
|
"learning_rate": 8.633593487183067e-07, |
|
"loss": 0.8171, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.8299627891926873, |
|
"grad_norm": 1.0012604552671969, |
|
"learning_rate": 8.554439425189537e-07, |
|
"loss": 0.8017, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.8307717197864424, |
|
"grad_norm": 0.9393560978003819, |
|
"learning_rate": 8.475615927165093e-07, |
|
"loss": 0.8208, |
|
"step": 5135 |
|
}, |
|
{ |
|
"epoch": 0.8315806503801974, |
|
"grad_norm": 0.9829342370330283, |
|
"learning_rate": 8.397123621794867e-07, |
|
"loss": 0.8213, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.8323895809739524, |
|
"grad_norm": 0.993397225397711, |
|
"learning_rate": 8.318963135122471e-07, |
|
"loss": 0.8386, |
|
"step": 5145 |
|
}, |
|
{ |
|
"epoch": 0.8331985115677075, |
|
"grad_norm": 0.9203262711299934, |
|
"learning_rate": 8.241135090544966e-07, |
|
"loss": 0.8135, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.8340074421614625, |
|
"grad_norm": 0.9284177490644239, |
|
"learning_rate": 8.163640108807897e-07, |
|
"loss": 0.834, |
|
"step": 5155 |
|
}, |
|
{ |
|
"epoch": 0.8348163727552176, |
|
"grad_norm": 0.9517537835456206, |
|
"learning_rate": 8.086478808000359e-07, |
|
"loss": 0.7949, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.8356253033489727, |
|
"grad_norm": 0.9372135921171859, |
|
"learning_rate": 8.009651803550045e-07, |
|
"loss": 0.8289, |
|
"step": 5165 |
|
}, |
|
{ |
|
"epoch": 0.8364342339427278, |
|
"grad_norm": 1.000548748366353, |
|
"learning_rate": 7.933159708218347e-07, |
|
"loss": 0.8268, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.8372431645364827, |
|
"grad_norm": 0.9612535821852152, |
|
"learning_rate": 7.857003132095481e-07, |
|
"loss": 0.8182, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.8380520951302378, |
|
"grad_norm": 0.9551451813093564, |
|
"learning_rate": 7.7811826825956e-07, |
|
"loss": 0.8399, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.8388610257239929, |
|
"grad_norm": 0.9803361111190585, |
|
"learning_rate": 7.705698964451941e-07, |
|
"loss": 0.8345, |
|
"step": 5185 |
|
}, |
|
{ |
|
"epoch": 0.839669956317748, |
|
"grad_norm": 0.9615430440473038, |
|
"learning_rate": 7.630552579712041e-07, |
|
"loss": 0.8482, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.840478886911503, |
|
"grad_norm": 0.8717103769599843, |
|
"learning_rate": 7.555744127732922e-07, |
|
"loss": 0.8133, |
|
"step": 5195 |
|
}, |
|
{ |
|
"epoch": 0.841287817505258, |
|
"grad_norm": 0.9559551946507219, |
|
"learning_rate": 7.481274205176286e-07, |
|
"loss": 0.8396, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.8420967480990131, |
|
"grad_norm": 0.9386431006557889, |
|
"learning_rate": 7.407143406003781e-07, |
|
"loss": 0.8166, |
|
"step": 5205 |
|
}, |
|
{ |
|
"epoch": 0.8429056786927681, |
|
"grad_norm": 0.9453424188819014, |
|
"learning_rate": 7.33335232147222e-07, |
|
"loss": 0.8318, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.8437146092865232, |
|
"grad_norm": 0.9335912363260552, |
|
"learning_rate": 7.25990154012895e-07, |
|
"loss": 0.8043, |
|
"step": 5215 |
|
}, |
|
{ |
|
"epoch": 0.8445235398802783, |
|
"grad_norm": 0.9242358250100629, |
|
"learning_rate": 7.186791647807078e-07, |
|
"loss": 0.8216, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.8453324704740334, |
|
"grad_norm": 0.8917056191120556, |
|
"learning_rate": 7.114023227620831e-07, |
|
"loss": 0.8263, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.8461414010677883, |
|
"grad_norm": 0.8756042557078504, |
|
"learning_rate": 7.04159685996092e-07, |
|
"loss": 0.8234, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.8469503316615434, |
|
"grad_norm": 0.9669924510507611, |
|
"learning_rate": 6.969513122489862e-07, |
|
"loss": 0.8317, |
|
"step": 5235 |
|
}, |
|
{ |
|
"epoch": 0.8477592622552985, |
|
"grad_norm": 0.9690987783521561, |
|
"learning_rate": 6.897772590137436e-07, |
|
"loss": 0.8196, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.8485681928490536, |
|
"grad_norm": 1.032895557175751, |
|
"learning_rate": 6.826375835096038e-07, |
|
"loss": 0.8017, |
|
"step": 5245 |
|
}, |
|
{ |
|
"epoch": 0.8493771234428086, |
|
"grad_norm": 1.0187566680457716, |
|
"learning_rate": 6.755323426816169e-07, |
|
"loss": 0.8354, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.8501860540365637, |
|
"grad_norm": 0.8816437739999422, |
|
"learning_rate": 6.684615932001847e-07, |
|
"loss": 0.8581, |
|
"step": 5255 |
|
}, |
|
{ |
|
"epoch": 0.8509949846303188, |
|
"grad_norm": 0.9088385041537731, |
|
"learning_rate": 6.614253914606134e-07, |
|
"loss": 0.8036, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.8518039152240737, |
|
"grad_norm": 0.9703072936941138, |
|
"learning_rate": 6.544237935826558e-07, |
|
"loss": 0.8277, |
|
"step": 5265 |
|
}, |
|
{ |
|
"epoch": 0.8526128458178288, |
|
"grad_norm": 0.9508642384505709, |
|
"learning_rate": 6.474568554100768e-07, |
|
"loss": 0.8243, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.8534217764115839, |
|
"grad_norm": 0.9433685567993545, |
|
"learning_rate": 6.405246325101955e-07, |
|
"loss": 0.8056, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.854230707005339, |
|
"grad_norm": 0.8978485349230313, |
|
"learning_rate": 6.336271801734479e-07, |
|
"loss": 0.8016, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.855039637599094, |
|
"grad_norm": 0.9591854547149082, |
|
"learning_rate": 6.267645534129446e-07, |
|
"loss": 0.8184, |
|
"step": 5285 |
|
}, |
|
{ |
|
"epoch": 0.855848568192849, |
|
"grad_norm": 0.9368695693533301, |
|
"learning_rate": 6.199368069640343e-07, |
|
"loss": 0.8345, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.8566574987866041, |
|
"grad_norm": 0.9363387262284574, |
|
"learning_rate": 6.131439952838608e-07, |
|
"loss": 0.8143, |
|
"step": 5295 |
|
}, |
|
{ |
|
"epoch": 0.8574664293803592, |
|
"grad_norm": 0.9449827101149645, |
|
"learning_rate": 6.063861725509374e-07, |
|
"loss": 0.801, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.8582753599741142, |
|
"grad_norm": 0.9911552846071767, |
|
"learning_rate": 5.996633926647083e-07, |
|
"loss": 0.8248, |
|
"step": 5305 |
|
}, |
|
{ |
|
"epoch": 0.8590842905678693, |
|
"grad_norm": 0.9088842745097252, |
|
"learning_rate": 5.92975709245121e-07, |
|
"loss": 0.8047, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.8598932211616244, |
|
"grad_norm": 0.8908028970825979, |
|
"learning_rate": 5.863231756322019e-07, |
|
"loss": 0.8242, |
|
"step": 5315 |
|
}, |
|
{ |
|
"epoch": 0.8607021517553793, |
|
"grad_norm": 0.9019135577959921, |
|
"learning_rate": 5.797058448856213e-07, |
|
"loss": 0.8182, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.8615110823491344, |
|
"grad_norm": 0.9938100484918152, |
|
"learning_rate": 5.731237697842811e-07, |
|
"loss": 0.8381, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.8623200129428895, |
|
"grad_norm": 0.9522671534271417, |
|
"learning_rate": 5.665770028258876e-07, |
|
"loss": 0.8103, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.8631289435366446, |
|
"grad_norm": 1.0241489828374508, |
|
"learning_rate": 5.600655962265345e-07, |
|
"loss": 0.8253, |
|
"step": 5335 |
|
}, |
|
{ |
|
"epoch": 0.8639378741303996, |
|
"grad_norm": 0.8339409337033903, |
|
"learning_rate": 5.535896019202853e-07, |
|
"loss": 0.8091, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.8647468047241547, |
|
"grad_norm": 0.953851217822305, |
|
"learning_rate": 5.471490715587618e-07, |
|
"loss": 0.8277, |
|
"step": 5345 |
|
}, |
|
{ |
|
"epoch": 0.8655557353179097, |
|
"grad_norm": 0.9253602828275306, |
|
"learning_rate": 5.407440565107291e-07, |
|
"loss": 0.8352, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.8663646659116648, |
|
"grad_norm": 0.8979717645863395, |
|
"learning_rate": 5.343746078616879e-07, |
|
"loss": 0.8346, |
|
"step": 5355 |
|
}, |
|
{ |
|
"epoch": 0.8671735965054198, |
|
"grad_norm": 0.9052656857430195, |
|
"learning_rate": 5.280407764134648e-07, |
|
"loss": 0.8098, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.8679825270991749, |
|
"grad_norm": 0.9265736364407307, |
|
"learning_rate": 5.21742612683811e-07, |
|
"loss": 0.8134, |
|
"step": 5365 |
|
}, |
|
{ |
|
"epoch": 0.86879145769293, |
|
"grad_norm": 0.9905238729956866, |
|
"learning_rate": 5.154801669059961e-07, |
|
"loss": 0.8057, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.869600388286685, |
|
"grad_norm": 0.9346637770934627, |
|
"learning_rate": 5.092534890284057e-07, |
|
"loss": 0.8319, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.87040931888044, |
|
"grad_norm": 0.9036166328056819, |
|
"learning_rate": 5.030626287141488e-07, |
|
"loss": 0.8242, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.8712182494741951, |
|
"grad_norm": 0.957351416925992, |
|
"learning_rate": 4.969076353406571e-07, |
|
"loss": 0.81, |
|
"step": 5385 |
|
}, |
|
{ |
|
"epoch": 0.8720271800679502, |
|
"grad_norm": 0.9998030765172857, |
|
"learning_rate": 4.907885579992943e-07, |
|
"loss": 0.7931, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.8728361106617052, |
|
"grad_norm": 0.9015423928862624, |
|
"learning_rate": 4.847054454949617e-07, |
|
"loss": 0.8321, |
|
"step": 5395 |
|
}, |
|
{ |
|
"epoch": 0.8736450412554603, |
|
"grad_norm": 0.8951302299296955, |
|
"learning_rate": 4.786583463457106e-07, |
|
"loss": 0.8175, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.8744539718492154, |
|
"grad_norm": 0.987288936023476, |
|
"learning_rate": 4.726473087823524e-07, |
|
"loss": 0.8149, |
|
"step": 5405 |
|
}, |
|
{ |
|
"epoch": 0.8752629024429704, |
|
"grad_norm": 0.9607690476849372, |
|
"learning_rate": 4.666723807480794e-07, |
|
"loss": 0.8395, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.8760718330367254, |
|
"grad_norm": 0.9412173085678744, |
|
"learning_rate": 4.6073360989807805e-07, |
|
"loss": 0.8131, |
|
"step": 5415 |
|
}, |
|
{ |
|
"epoch": 0.8768807636304805, |
|
"grad_norm": 0.9953808945790064, |
|
"learning_rate": 4.5483104359915095e-07, |
|
"loss": 0.8337, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.8776896942242356, |
|
"grad_norm": 0.9224438419258796, |
|
"learning_rate": 4.4896472892933693e-07, |
|
"loss": 0.827, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.8784986248179906, |
|
"grad_norm": 0.9067448515242419, |
|
"learning_rate": 4.431347126775382e-07, |
|
"loss": 0.8225, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.8793075554117457, |
|
"grad_norm": 0.9160793517967267, |
|
"learning_rate": 4.3734104134314505e-07, |
|
"loss": 0.8093, |
|
"step": 5435 |
|
}, |
|
{ |
|
"epoch": 0.8801164860055007, |
|
"grad_norm": 0.8957867534440033, |
|
"learning_rate": 4.3158376113566656e-07, |
|
"loss": 0.8255, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.8809254165992558, |
|
"grad_norm": 0.9414164683136745, |
|
"learning_rate": 4.258629179743612e-07, |
|
"loss": 0.8307, |
|
"step": 5445 |
|
}, |
|
{ |
|
"epoch": 0.8817343471930108, |
|
"grad_norm": 1.0412360371956955, |
|
"learning_rate": 4.2017855748786975e-07, |
|
"loss": 0.8167, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.8825432777867659, |
|
"grad_norm": 0.9168749499249985, |
|
"learning_rate": 4.1453072501385415e-07, |
|
"loss": 0.807, |
|
"step": 5455 |
|
}, |
|
{ |
|
"epoch": 0.883352208380521, |
|
"grad_norm": 0.885047475917418, |
|
"learning_rate": 4.089194655986306e-07, |
|
"loss": 0.831, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.8841611389742761, |
|
"grad_norm": 0.9248545364082547, |
|
"learning_rate": 4.0334482399681684e-07, |
|
"loss": 0.836, |
|
"step": 5465 |
|
}, |
|
{ |
|
"epoch": 0.884970069568031, |
|
"grad_norm": 0.9565631081942835, |
|
"learning_rate": 3.978068446709721e-07, |
|
"loss": 0.8206, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.8857790001617861, |
|
"grad_norm": 0.9313667515936709, |
|
"learning_rate": 3.923055717912411e-07, |
|
"loss": 0.8315, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.8865879307555412, |
|
"grad_norm": 0.9153735916152161, |
|
"learning_rate": 3.868410492350044e-07, |
|
"loss": 0.8315, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.8873968613492962, |
|
"grad_norm": 0.8996467266796176, |
|
"learning_rate": 3.8141332058652447e-07, |
|
"loss": 0.8039, |
|
"step": 5485 |
|
}, |
|
{ |
|
"epoch": 0.8882057919430513, |
|
"grad_norm": 0.8739192973106119, |
|
"learning_rate": 3.7602242913660325e-07, |
|
"loss": 0.8196, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.8890147225368064, |
|
"grad_norm": 1.0836849051236812, |
|
"learning_rate": 3.7066841788223394e-07, |
|
"loss": 0.8414, |
|
"step": 5495 |
|
}, |
|
{ |
|
"epoch": 0.8898236531305614, |
|
"grad_norm": 0.8886207081976889, |
|
"learning_rate": 3.6535132952625784e-07, |
|
"loss": 0.8041, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.8906325837243164, |
|
"grad_norm": 0.8954848811354755, |
|
"learning_rate": 3.6007120647702566e-07, |
|
"loss": 0.836, |
|
"step": 5505 |
|
}, |
|
{ |
|
"epoch": 0.8914415143180715, |
|
"grad_norm": 0.9763361834402516, |
|
"learning_rate": 3.548280908480556e-07, |
|
"loss": 0.8193, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.8922504449118266, |
|
"grad_norm": 0.9676090733901674, |
|
"learning_rate": 3.496220244577025e-07, |
|
"loss": 0.8247, |
|
"step": 5515 |
|
}, |
|
{ |
|
"epoch": 0.8930593755055817, |
|
"grad_norm": 0.9583454675377773, |
|
"learning_rate": 3.4445304882882056e-07, |
|
"loss": 0.8232, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.8938683060993367, |
|
"grad_norm": 0.9149362551378993, |
|
"learning_rate": 3.3932120518843315e-07, |
|
"loss": 0.8378, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.8946772366930917, |
|
"grad_norm": 0.9981131323135035, |
|
"learning_rate": 3.342265344674034e-07, |
|
"loss": 0.8364, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.8954861672868468, |
|
"grad_norm": 0.9602463248675637, |
|
"learning_rate": 3.291690773001116e-07, |
|
"loss": 0.8148, |
|
"step": 5535 |
|
}, |
|
{ |
|
"epoch": 0.8962950978806018, |
|
"grad_norm": 0.9061694557646229, |
|
"learning_rate": 3.241488740241222e-07, |
|
"loss": 0.8135, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.8971040284743569, |
|
"grad_norm": 0.9504624920131578, |
|
"learning_rate": 3.1916596467987395e-07, |
|
"loss": 0.8, |
|
"step": 5545 |
|
}, |
|
{ |
|
"epoch": 0.897912959068112, |
|
"grad_norm": 1.0083410313503633, |
|
"learning_rate": 3.142203890103512e-07, |
|
"loss": 0.8194, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.8987218896618671, |
|
"grad_norm": 0.981599869718165, |
|
"learning_rate": 3.093121864607707e-07, |
|
"loss": 0.816, |
|
"step": 5555 |
|
}, |
|
{ |
|
"epoch": 0.899530820255622, |
|
"grad_norm": 0.927185199167883, |
|
"learning_rate": 3.0444139617826605e-07, |
|
"loss": 0.8133, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.9003397508493771, |
|
"grad_norm": 1.0064121664416639, |
|
"learning_rate": 2.996080570115778e-07, |
|
"loss": 0.8262, |
|
"step": 5565 |
|
}, |
|
{ |
|
"epoch": 0.9011486814431322, |
|
"grad_norm": 0.9554574033151684, |
|
"learning_rate": 2.9481220751073793e-07, |
|
"loss": 0.8247, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.9019576120368872, |
|
"grad_norm": 0.9588661920794765, |
|
"learning_rate": 2.9005388592676987e-07, |
|
"loss": 0.8273, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.9027665426306423, |
|
"grad_norm": 0.885495966716363, |
|
"learning_rate": 2.8533313021137765e-07, |
|
"loss": 0.8326, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.9035754732243974, |
|
"grad_norm": 0.9273868198730849, |
|
"learning_rate": 2.806499780166455e-07, |
|
"loss": 0.8073, |
|
"step": 5585 |
|
}, |
|
{ |
|
"epoch": 0.9043844038181524, |
|
"grad_norm": 1.0414377339942757, |
|
"learning_rate": 2.760044666947387e-07, |
|
"loss": 0.8324, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.9051933344119074, |
|
"grad_norm": 0.9699497983834742, |
|
"learning_rate": 2.7139663329760203e-07, |
|
"loss": 0.8222, |
|
"step": 5595 |
|
}, |
|
{ |
|
"epoch": 0.9060022650056625, |
|
"grad_norm": 0.9269981190661996, |
|
"learning_rate": 2.668265145766669e-07, |
|
"loss": 0.8298, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.9068111955994176, |
|
"grad_norm": 1.0012454133607382, |
|
"learning_rate": 2.6229414698255907e-07, |
|
"loss": 0.8293, |
|
"step": 5605 |
|
}, |
|
{ |
|
"epoch": 0.9076201261931727, |
|
"grad_norm": 0.9294225278070521, |
|
"learning_rate": 2.577995666648053e-07, |
|
"loss": 0.8228, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.9084290567869276, |
|
"grad_norm": 0.940534897284778, |
|
"learning_rate": 2.533428094715473e-07, |
|
"loss": 0.8232, |
|
"step": 5615 |
|
}, |
|
{ |
|
"epoch": 0.9092379873806827, |
|
"grad_norm": 0.9310916767297108, |
|
"learning_rate": 2.489239109492536e-07, |
|
"loss": 0.8196, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.9100469179744378, |
|
"grad_norm": 1.012544196651258, |
|
"learning_rate": 2.4454290634243927e-07, |
|
"loss": 0.8092, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.9108558485681928, |
|
"grad_norm": 0.9941950459630348, |
|
"learning_rate": 2.4019983059338005e-07, |
|
"loss": 0.8219, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.9116647791619479, |
|
"grad_norm": 1.0349621117968182, |
|
"learning_rate": 2.3589471834183975e-07, |
|
"loss": 0.7977, |
|
"step": 5635 |
|
}, |
|
{ |
|
"epoch": 0.912473709755703, |
|
"grad_norm": 0.9021187223816023, |
|
"learning_rate": 2.3162760392478777e-07, |
|
"loss": 0.8256, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.913282640349458, |
|
"grad_norm": 0.9107025655941831, |
|
"learning_rate": 2.2739852137612984e-07, |
|
"loss": 0.8279, |
|
"step": 5645 |
|
}, |
|
{ |
|
"epoch": 0.914091570943213, |
|
"grad_norm": 0.9459797428393495, |
|
"learning_rate": 2.2320750442643423e-07, |
|
"loss": 0.8077, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.9149005015369681, |
|
"grad_norm": 1.1288128324454283, |
|
"learning_rate": 2.1905458650266276e-07, |
|
"loss": 0.8407, |
|
"step": 5655 |
|
}, |
|
{ |
|
"epoch": 0.9157094321307232, |
|
"grad_norm": 0.9221316094387947, |
|
"learning_rate": 2.149398007279052e-07, |
|
"loss": 0.8209, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.9165183627244783, |
|
"grad_norm": 0.9047592079692445, |
|
"learning_rate": 2.108631799211158e-07, |
|
"loss": 0.8037, |
|
"step": 5665 |
|
}, |
|
{ |
|
"epoch": 0.9173272933182333, |
|
"grad_norm": 0.9186500058678413, |
|
"learning_rate": 2.0682475659684953e-07, |
|
"loss": 0.836, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.9181362239119883, |
|
"grad_norm": 0.8965926605186532, |
|
"learning_rate": 2.0282456296500385e-07, |
|
"loss": 0.8303, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.9189451545057434, |
|
"grad_norm": 0.9119449898107598, |
|
"learning_rate": 1.988626309305597e-07, |
|
"loss": 0.838, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.9197540850994984, |
|
"grad_norm": 0.9512693378379136, |
|
"learning_rate": 1.9493899209333146e-07, |
|
"loss": 0.8027, |
|
"step": 5685 |
|
}, |
|
{ |
|
"epoch": 0.9205630156932535, |
|
"grad_norm": 0.9690819188637002, |
|
"learning_rate": 1.9105367774771122e-07, |
|
"loss": 0.8125, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.9213719462870086, |
|
"grad_norm": 0.9571795973045409, |
|
"learning_rate": 1.8720671888242058e-07, |
|
"loss": 0.8271, |
|
"step": 5695 |
|
}, |
|
{ |
|
"epoch": 0.9221808768807637, |
|
"grad_norm": 0.909548062163855, |
|
"learning_rate": 1.8339814618026252e-07, |
|
"loss": 0.833, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.9229898074745186, |
|
"grad_norm": 0.9572976405315803, |
|
"learning_rate": 1.7962799001787823e-07, |
|
"loss": 0.8103, |
|
"step": 5705 |
|
}, |
|
{ |
|
"epoch": 0.9237987380682737, |
|
"grad_norm": 1.0064204020940664, |
|
"learning_rate": 1.7589628046550343e-07, |
|
"loss": 0.8369, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.9246076686620288, |
|
"grad_norm": 0.9734981320914123, |
|
"learning_rate": 1.7220304728672977e-07, |
|
"loss": 0.8187, |
|
"step": 5715 |
|
}, |
|
{ |
|
"epoch": 0.9254165992557839, |
|
"grad_norm": 0.9779299096983186, |
|
"learning_rate": 1.6854831993826591e-07, |
|
"loss": 0.8212, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.9262255298495389, |
|
"grad_norm": 0.8990160141351417, |
|
"learning_rate": 1.6493212756970356e-07, |
|
"loss": 0.8194, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.927034460443294, |
|
"grad_norm": 0.9133770658973546, |
|
"learning_rate": 1.6135449902328627e-07, |
|
"loss": 0.8377, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.927843391037049, |
|
"grad_norm": 0.8935752099105897, |
|
"learning_rate": 1.578154628336753e-07, |
|
"loss": 0.8247, |
|
"step": 5735 |
|
}, |
|
{ |
|
"epoch": 0.928652321630804, |
|
"grad_norm": 0.9599763188366742, |
|
"learning_rate": 1.5431504722772605e-07, |
|
"loss": 0.8259, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.9294612522245591, |
|
"grad_norm": 0.9543468652422167, |
|
"learning_rate": 1.5085328012426293e-07, |
|
"loss": 0.8239, |
|
"step": 5745 |
|
}, |
|
{ |
|
"epoch": 0.9302701828183142, |
|
"grad_norm": 0.9093126800081541, |
|
"learning_rate": 1.4743018913385308e-07, |
|
"loss": 0.8313, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.9310791134120693, |
|
"grad_norm": 1.002356581515829, |
|
"learning_rate": 1.4404580155859106e-07, |
|
"loss": 0.8288, |
|
"step": 5755 |
|
}, |
|
{ |
|
"epoch": 0.9318880440058243, |
|
"grad_norm": 0.900910009096251, |
|
"learning_rate": 1.407001443918743e-07, |
|
"loss": 0.824, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.9326969745995793, |
|
"grad_norm": 0.9316638128288214, |
|
"learning_rate": 1.373932443181958e-07, |
|
"loss": 0.8376, |
|
"step": 5765 |
|
}, |
|
{ |
|
"epoch": 0.9335059051933344, |
|
"grad_norm": 0.8859091785962655, |
|
"learning_rate": 1.3412512771292574e-07, |
|
"loss": 0.8262, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.9343148357870895, |
|
"grad_norm": 0.9517796063815803, |
|
"learning_rate": 1.3089582064210293e-07, |
|
"loss": 0.8332, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.9351237663808445, |
|
"grad_norm": 0.956425135920899, |
|
"learning_rate": 1.2770534886222709e-07, |
|
"loss": 0.8441, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.9359326969745996, |
|
"grad_norm": 0.9436260486622394, |
|
"learning_rate": 1.2455373782005343e-07, |
|
"loss": 0.8246, |
|
"step": 5785 |
|
}, |
|
{ |
|
"epoch": 0.9367416275683547, |
|
"grad_norm": 0.8907054197780977, |
|
"learning_rate": 1.2144101265238795e-07, |
|
"loss": 0.8226, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.9375505581621096, |
|
"grad_norm": 0.9599764091397304, |
|
"learning_rate": 1.1836719818588971e-07, |
|
"loss": 0.8197, |
|
"step": 5795 |
|
}, |
|
{ |
|
"epoch": 0.9383594887558647, |
|
"grad_norm": 0.89616146112279, |
|
"learning_rate": 1.1533231893687158e-07, |
|
"loss": 0.8292, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.9391684193496198, |
|
"grad_norm": 0.9254286869197753, |
|
"learning_rate": 1.1233639911110317e-07, |
|
"loss": 0.8106, |
|
"step": 5805 |
|
}, |
|
{ |
|
"epoch": 0.9399773499433749, |
|
"grad_norm": 0.9336133063869916, |
|
"learning_rate": 1.0937946260362154e-07, |
|
"loss": 0.8174, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.9407862805371299, |
|
"grad_norm": 0.9193663783506368, |
|
"learning_rate": 1.0646153299853523e-07, |
|
"loss": 0.8219, |
|
"step": 5815 |
|
}, |
|
{ |
|
"epoch": 0.941595211130885, |
|
"grad_norm": 0.9922561636241934, |
|
"learning_rate": 1.0358263356884223e-07, |
|
"loss": 0.8192, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.94240414172464, |
|
"grad_norm": 0.9269516184874731, |
|
"learning_rate": 1.0074278727623955e-07, |
|
"loss": 0.8342, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.9432130723183951, |
|
"grad_norm": 0.9567075615786765, |
|
"learning_rate": 9.794201677094162e-08, |
|
"loss": 0.8456, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.9440220029121501, |
|
"grad_norm": 0.900361140775962, |
|
"learning_rate": 9.51803443915017e-08, |
|
"loss": 0.8071, |
|
"step": 5835 |
|
}, |
|
{ |
|
"epoch": 0.9448309335059052, |
|
"grad_norm": 0.8928244797252599, |
|
"learning_rate": 9.245779216463024e-08, |
|
"loss": 0.8184, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.9456398640996603, |
|
"grad_norm": 1.0034007579574744, |
|
"learning_rate": 8.977438180502118e-08, |
|
"loss": 0.8356, |
|
"step": 5845 |
|
}, |
|
{ |
|
"epoch": 0.9464487946934153, |
|
"grad_norm": 0.9678539855117049, |
|
"learning_rate": 8.713013471517873e-08, |
|
"loss": 0.8346, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.9472577252871703, |
|
"grad_norm": 0.9529843175929286, |
|
"learning_rate": 8.452507198524584e-08, |
|
"loss": 0.8101, |
|
"step": 5855 |
|
}, |
|
{ |
|
"epoch": 0.9480666558809254, |
|
"grad_norm": 0.9262321483199513, |
|
"learning_rate": 8.195921439283716e-08, |
|
"loss": 0.8242, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.9488755864746805, |
|
"grad_norm": 0.885777314821936, |
|
"learning_rate": 7.943258240287355e-08, |
|
"loss": 0.8201, |
|
"step": 5865 |
|
}, |
|
{ |
|
"epoch": 0.9496845170684355, |
|
"grad_norm": 0.9553802642628383, |
|
"learning_rate": 7.694519616741503e-08, |
|
"loss": 0.8331, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.9504934476621906, |
|
"grad_norm": 0.9768522628826408, |
|
"learning_rate": 7.449707552550533e-08, |
|
"loss": 0.8147, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.9513023782559457, |
|
"grad_norm": 0.9068927130823696, |
|
"learning_rate": 7.208824000301151e-08, |
|
"loss": 0.8302, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.9521113088497007, |
|
"grad_norm": 0.9575224681944164, |
|
"learning_rate": 6.971870881246678e-08, |
|
"loss": 0.8145, |
|
"step": 5885 |
|
}, |
|
{ |
|
"epoch": 0.9529202394434557, |
|
"grad_norm": 0.8832000510088256, |
|
"learning_rate": 6.738850085291904e-08, |
|
"loss": 0.8177, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.9537291700372108, |
|
"grad_norm": 0.9566260863289568, |
|
"learning_rate": 6.509763470977925e-08, |
|
"loss": 0.8321, |
|
"step": 5895 |
|
}, |
|
{ |
|
"epoch": 0.9545381006309659, |
|
"grad_norm": 0.9104128242370191, |
|
"learning_rate": 6.284612865467499e-08, |
|
"loss": 0.8271, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.9553470312247209, |
|
"grad_norm": 0.9365563187928431, |
|
"learning_rate": 6.063400064530155e-08, |
|
"loss": 0.8251, |
|
"step": 5905 |
|
}, |
|
{ |
|
"epoch": 0.956155961818476, |
|
"grad_norm": 0.8777597042529512, |
|
"learning_rate": 5.8461268325281096e-08, |
|
"loss": 0.8021, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.956964892412231, |
|
"grad_norm": 0.9899981204211554, |
|
"learning_rate": 5.632794902402205e-08, |
|
"loss": 0.8099, |
|
"step": 5915 |
|
}, |
|
{ |
|
"epoch": 0.9577738230059861, |
|
"grad_norm": 0.900474026342178, |
|
"learning_rate": 5.423405975657936e-08, |
|
"loss": 0.8367, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.9585827535997411, |
|
"grad_norm": 1.0047500482028124, |
|
"learning_rate": 5.217961722351894e-08, |
|
"loss": 0.8343, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.9593916841934962, |
|
"grad_norm": 0.9477198924309056, |
|
"learning_rate": 5.016463781078618e-08, |
|
"loss": 0.8317, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.9602006147872513, |
|
"grad_norm": 0.8807307401976437, |
|
"learning_rate": 4.818913758957378e-08, |
|
"loss": 0.8239, |
|
"step": 5935 |
|
}, |
|
{ |
|
"epoch": 0.9610095453810064, |
|
"grad_norm": 0.9428322183465226, |
|
"learning_rate": 4.6253132316194103e-08, |
|
"loss": 0.8296, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.9618184759747613, |
|
"grad_norm": 0.9762495130698056, |
|
"learning_rate": 4.4356637431953734e-08, |
|
"loss": 0.8035, |
|
"step": 5945 |
|
}, |
|
{ |
|
"epoch": 0.9626274065685164, |
|
"grad_norm": 0.9782271222260458, |
|
"learning_rate": 4.2499668063029075e-08, |
|
"loss": 0.8079, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.9634363371622715, |
|
"grad_norm": 1.0490239258099892, |
|
"learning_rate": 4.068223902034651e-08, |
|
"loss": 0.827, |
|
"step": 5955 |
|
}, |
|
{ |
|
"epoch": 0.9642452677560265, |
|
"grad_norm": 0.9547918261422726, |
|
"learning_rate": 3.89043647994658e-08, |
|
"loss": 0.8309, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.9650541983497816, |
|
"grad_norm": 0.9195603493515522, |
|
"learning_rate": 3.716605958046071e-08, |
|
"loss": 0.8324, |
|
"step": 5965 |
|
}, |
|
{ |
|
"epoch": 0.9658631289435367, |
|
"grad_norm": 0.9215008996007901, |
|
"learning_rate": 3.546733722781026e-08, |
|
"loss": 0.8202, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.9666720595372917, |
|
"grad_norm": 0.9117492446000369, |
|
"learning_rate": 3.3808211290284886e-08, |
|
"loss": 0.8283, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.9674809901310467, |
|
"grad_norm": 0.9238030724612377, |
|
"learning_rate": 3.218869500084099e-08, |
|
"loss": 0.8203, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.9682899207248018, |
|
"grad_norm": 0.9683332126948087, |
|
"learning_rate": 3.0608801276511556e-08, |
|
"loss": 0.8063, |
|
"step": 5985 |
|
}, |
|
{ |
|
"epoch": 0.9690988513185569, |
|
"grad_norm": 0.9578244707163264, |
|
"learning_rate": 2.9068542718307947e-08, |
|
"loss": 0.8178, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.969907781912312, |
|
"grad_norm": 1.0002466766621863, |
|
"learning_rate": 2.7567931611116037e-08, |
|
"loss": 0.8298, |
|
"step": 5995 |
|
}, |
|
{ |
|
"epoch": 0.970716712506067, |
|
"grad_norm": 0.9033323404105812, |
|
"learning_rate": 2.6106979923599117e-08, |
|
"loss": 0.8314, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.971525643099822, |
|
"grad_norm": 0.9981727314544666, |
|
"learning_rate": 2.4685699308102385e-08, |
|
"loss": 0.8245, |
|
"step": 6005 |
|
}, |
|
{ |
|
"epoch": 0.9723345736935771, |
|
"grad_norm": 0.90324273915133, |
|
"learning_rate": 2.330410110056025e-08, |
|
"loss": 0.8128, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.9731435042873321, |
|
"grad_norm": 0.9184000593665834, |
|
"learning_rate": 2.1962196320406416e-08, |
|
"loss": 0.8234, |
|
"step": 6015 |
|
}, |
|
{ |
|
"epoch": 0.9739524348810872, |
|
"grad_norm": 1.0208286644340638, |
|
"learning_rate": 2.065999567048449e-08, |
|
"loss": 0.8313, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.9747613654748423, |
|
"grad_norm": 0.8848130385791, |
|
"learning_rate": 1.9397509536964177e-08, |
|
"loss": 0.8045, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.9755702960685974, |
|
"grad_norm": 0.8868943198262542, |
|
"learning_rate": 1.8174747989258002e-08, |
|
"loss": 0.8181, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.9763792266623523, |
|
"grad_norm": 0.9860697437983315, |
|
"learning_rate": 1.6991720779939157e-08, |
|
"loss": 0.8269, |
|
"step": 6035 |
|
}, |
|
{ |
|
"epoch": 0.9771881572561074, |
|
"grad_norm": 0.9169269167928082, |
|
"learning_rate": 1.5848437344667124e-08, |
|
"loss": 0.8155, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.9779970878498625, |
|
"grad_norm": 0.8899129024501534, |
|
"learning_rate": 1.4744906802110493e-08, |
|
"loss": 0.814, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 0.9788060184436176, |
|
"grad_norm": 0.9676626630215127, |
|
"learning_rate": 1.3681137953872604e-08, |
|
"loss": 0.8248, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.9796149490373726, |
|
"grad_norm": 1.008805163098565, |
|
"learning_rate": 1.2657139284425468e-08, |
|
"loss": 0.8442, |
|
"step": 6055 |
|
}, |
|
{ |
|
"epoch": 0.9804238796311276, |
|
"grad_norm": 0.9275436235635999, |
|
"learning_rate": 1.167291896103817e-08, |
|
"loss": 0.8255, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.9812328102248827, |
|
"grad_norm": 0.9859360057982008, |
|
"learning_rate": 1.0728484833713581e-08, |
|
"loss": 0.8231, |
|
"step": 6065 |
|
}, |
|
{ |
|
"epoch": 0.9820417408186377, |
|
"grad_norm": 0.9537378986780001, |
|
"learning_rate": 9.823844435126184e-09, |
|
"loss": 0.8204, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.9828506714123928, |
|
"grad_norm": 0.9029585570220431, |
|
"learning_rate": 8.959004980559905e-09, |
|
"loss": 0.8351, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.9836596020061479, |
|
"grad_norm": 0.9441776141647337, |
|
"learning_rate": 8.133973367853708e-09, |
|
"loss": 0.8401, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.984468532599903, |
|
"grad_norm": 0.9139424255052578, |
|
"learning_rate": 7.348756177343319e-09, |
|
"loss": 0.8228, |
|
"step": 6085 |
|
}, |
|
{ |
|
"epoch": 0.9852774631936579, |
|
"grad_norm": 0.9582142736178327, |
|
"learning_rate": 6.603359671810694e-09, |
|
"loss": 0.8239, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.986086393787413, |
|
"grad_norm": 0.9470852983866349, |
|
"learning_rate": 5.8977897964335174e-09, |
|
"loss": 0.8164, |
|
"step": 6095 |
|
}, |
|
{ |
|
"epoch": 0.9868953243811681, |
|
"grad_norm": 0.9016603805684872, |
|
"learning_rate": 5.232052178738567e-09, |
|
"loss": 0.8232, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.9877042549749232, |
|
"grad_norm": 0.9549402446433066, |
|
"learning_rate": 4.606152128555086e-09, |
|
"loss": 0.8088, |
|
"step": 6105 |
|
}, |
|
{ |
|
"epoch": 0.9885131855686782, |
|
"grad_norm": 0.9529061460613366, |
|
"learning_rate": 4.020094637973704e-09, |
|
"loss": 0.8162, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.9893221161624333, |
|
"grad_norm": 0.9500963750803435, |
|
"learning_rate": 3.4738843813075795e-09, |
|
"loss": 0.813, |
|
"step": 6115 |
|
}, |
|
{ |
|
"epoch": 0.9901310467561883, |
|
"grad_norm": 0.9329204750410419, |
|
"learning_rate": 2.967525715052433e-09, |
|
"loss": 0.839, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.9909399773499433, |
|
"grad_norm": 0.9456290946732857, |
|
"learning_rate": 2.5010226778537927e-09, |
|
"loss": 0.8231, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.9917489079436984, |
|
"grad_norm": 0.8612020082608306, |
|
"learning_rate": 2.074378990474246e-09, |
|
"loss": 0.8139, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.9925578385374535, |
|
"grad_norm": 0.9267109066168538, |
|
"learning_rate": 1.687598055764017e-09, |
|
"loss": 0.8351, |
|
"step": 6135 |
|
}, |
|
{ |
|
"epoch": 0.9933667691312086, |
|
"grad_norm": 0.9928733972042116, |
|
"learning_rate": 1.3406829586337656e-09, |
|
"loss": 0.8033, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.9941756997249636, |
|
"grad_norm": 0.882709413718118, |
|
"learning_rate": 1.0336364660290532e-09, |
|
"loss": 0.8098, |
|
"step": 6145 |
|
}, |
|
{ |
|
"epoch": 0.9949846303187186, |
|
"grad_norm": 0.9193633255003315, |
|
"learning_rate": 7.664610269103589e-10, |
|
"loss": 0.8285, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.9957935609124737, |
|
"grad_norm": 0.9524125245282298, |
|
"learning_rate": 5.391587722303194e-10, |
|
"loss": 0.8087, |
|
"step": 6155 |
|
}, |
|
{ |
|
"epoch": 0.9966024915062288, |
|
"grad_norm": 0.9577558462345785, |
|
"learning_rate": 3.5173151492096104e-10, |
|
"loss": 0.8309, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.9974114220999838, |
|
"grad_norm": 0.892739545774551, |
|
"learning_rate": 2.0418074987538229e-10, |
|
"loss": 0.8144, |
|
"step": 6165 |
|
}, |
|
{ |
|
"epoch": 0.9982203526937389, |
|
"grad_norm": 0.973028280918144, |
|
"learning_rate": 9.650765393720563e-11, |
|
"loss": 0.8147, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.999029283287494, |
|
"grad_norm": 0.9043858725591138, |
|
"learning_rate": 2.8713085892806415e-11, |
|
"loss": 0.8312, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.9998382138812489, |
|
"grad_norm": 0.9568219074455289, |
|
"learning_rate": 7.975864613207762e-13, |
|
"loss": 0.8409, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.8599082231521606, |
|
"eval_runtime": 3.0154, |
|
"eval_samples_per_second": 3.316, |
|
"eval_steps_per_second": 0.995, |
|
"step": 6181 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 6181, |
|
"total_flos": 2571089024581632.0, |
|
"train_loss": 0.8638364601301117, |
|
"train_runtime": 28571.7374, |
|
"train_samples_per_second": 3.461, |
|
"train_steps_per_second": 0.216 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 6181, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2571089024581632.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|