|
{ |
|
"best_metric": 0.90963293, |
|
"best_model_checkpoint": "/volume/output/run2/checkpoint-252560", |
|
"epoch": 20.0, |
|
"eval_steps": 500, |
|
"global_step": 252560, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05004751346214761, |
|
"grad_norm": 1.831141710281372, |
|
"learning_rate": 9.974976243268926e-05, |
|
"loss": 1.5344, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.10009502692429521, |
|
"grad_norm": 3.449512004852295, |
|
"learning_rate": 9.949952486537853e-05, |
|
"loss": 0.3225, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 0.1501425403864428, |
|
"grad_norm": 1.8513766527175903, |
|
"learning_rate": 9.92492872980678e-05, |
|
"loss": 0.2466, |
|
"step": 1896 |
|
}, |
|
{ |
|
"epoch": 0.20019005384859043, |
|
"grad_norm": 1.0473196506500244, |
|
"learning_rate": 9.899904973075705e-05, |
|
"loss": 0.2176, |
|
"step": 2528 |
|
}, |
|
{ |
|
"epoch": 0.25023756731073804, |
|
"grad_norm": 1.3224372863769531, |
|
"learning_rate": 9.874881216344632e-05, |
|
"loss": 0.1852, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.3002850807728856, |
|
"grad_norm": 1.1163212060928345, |
|
"learning_rate": 9.849857459613557e-05, |
|
"loss": 0.1725, |
|
"step": 3792 |
|
}, |
|
{ |
|
"epoch": 0.35033259423503327, |
|
"grad_norm": 0.2766059339046478, |
|
"learning_rate": 9.824833702882485e-05, |
|
"loss": 0.1613, |
|
"step": 4424 |
|
}, |
|
{ |
|
"epoch": 0.40038010769718085, |
|
"grad_norm": 1.0185670852661133, |
|
"learning_rate": 9.79980994615141e-05, |
|
"loss": 0.1345, |
|
"step": 5056 |
|
}, |
|
{ |
|
"epoch": 0.4504276211593285, |
|
"grad_norm": 1.164562702178955, |
|
"learning_rate": 9.774786189420336e-05, |
|
"loss": 0.1364, |
|
"step": 5688 |
|
}, |
|
{ |
|
"epoch": 0.5004751346214761, |
|
"grad_norm": 1.1677042245864868, |
|
"learning_rate": 9.749762432689263e-05, |
|
"loss": 0.1277, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.5505226480836237, |
|
"grad_norm": 3.836665153503418, |
|
"learning_rate": 9.724738675958188e-05, |
|
"loss": 0.1254, |
|
"step": 6952 |
|
}, |
|
{ |
|
"epoch": 0.6005701615457713, |
|
"grad_norm": 2.3849477767944336, |
|
"learning_rate": 9.699714919227115e-05, |
|
"loss": 0.1188, |
|
"step": 7584 |
|
}, |
|
{ |
|
"epoch": 0.6506176750079189, |
|
"grad_norm": 3.4638051986694336, |
|
"learning_rate": 9.674691162496041e-05, |
|
"loss": 0.1159, |
|
"step": 8216 |
|
}, |
|
{ |
|
"epoch": 0.7006651884700665, |
|
"grad_norm": 0.2078377604484558, |
|
"learning_rate": 9.649667405764967e-05, |
|
"loss": 0.1097, |
|
"step": 8848 |
|
}, |
|
{ |
|
"epoch": 0.7507127019322142, |
|
"grad_norm": 2.513908624649048, |
|
"learning_rate": 9.624643649033894e-05, |
|
"loss": 0.1073, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 0.8007602153943617, |
|
"grad_norm": 1.9936473369598389, |
|
"learning_rate": 9.599619892302819e-05, |
|
"loss": 0.1006, |
|
"step": 10112 |
|
}, |
|
{ |
|
"epoch": 0.8508077288565093, |
|
"grad_norm": 1.9626480340957642, |
|
"learning_rate": 9.574596135571746e-05, |
|
"loss": 0.1019, |
|
"step": 10744 |
|
}, |
|
{ |
|
"epoch": 0.900855242318657, |
|
"grad_norm": 1.7610087394714355, |
|
"learning_rate": 9.549572378840672e-05, |
|
"loss": 0.0898, |
|
"step": 11376 |
|
}, |
|
{ |
|
"epoch": 0.9509027557808045, |
|
"grad_norm": 1.6743286848068237, |
|
"learning_rate": 9.524548622109598e-05, |
|
"loss": 0.0912, |
|
"step": 12008 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_gen_len": 12.34557138, |
|
"eval_loss": 0.06976257264614105, |
|
"eval_runtime": 538.4523, |
|
"eval_samples_per_second": 98.843, |
|
"eval_steps_per_second": 1.545, |
|
"eval_wordacc": 0.8983691, |
|
"eval_wordacc_oov": 0.84212982, |
|
"step": 12628 |
|
}, |
|
{ |
|
"epoch": 1.0009502692429522, |
|
"grad_norm": 0.813258707523346, |
|
"learning_rate": 9.499524865378525e-05, |
|
"loss": 0.0908, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 1.0509977827050998, |
|
"grad_norm": 1.4410091638565063, |
|
"learning_rate": 9.474501108647451e-05, |
|
"loss": 0.0864, |
|
"step": 13272 |
|
}, |
|
{ |
|
"epoch": 1.1010452961672474, |
|
"grad_norm": 0.1239413172006607, |
|
"learning_rate": 9.449477351916377e-05, |
|
"loss": 0.0819, |
|
"step": 13904 |
|
}, |
|
{ |
|
"epoch": 1.151092809629395, |
|
"grad_norm": 0.41668111085891724, |
|
"learning_rate": 9.424453595185304e-05, |
|
"loss": 0.0872, |
|
"step": 14536 |
|
}, |
|
{ |
|
"epoch": 1.2011403230915425, |
|
"grad_norm": 0.2707739472389221, |
|
"learning_rate": 9.399429838454229e-05, |
|
"loss": 0.082, |
|
"step": 15168 |
|
}, |
|
{ |
|
"epoch": 1.2511878365536901, |
|
"grad_norm": 0.5811319351196289, |
|
"learning_rate": 9.374406081723154e-05, |
|
"loss": 0.0847, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 1.3012353500158378, |
|
"grad_norm": 1.7169886827468872, |
|
"learning_rate": 9.349382324992082e-05, |
|
"loss": 0.0817, |
|
"step": 16432 |
|
}, |
|
{ |
|
"epoch": 1.3512828634779854, |
|
"grad_norm": 0.11871356517076492, |
|
"learning_rate": 9.324358568261008e-05, |
|
"loss": 0.0884, |
|
"step": 17064 |
|
}, |
|
{ |
|
"epoch": 1.401330376940133, |
|
"grad_norm": 1.614180326461792, |
|
"learning_rate": 9.299334811529935e-05, |
|
"loss": 0.0781, |
|
"step": 17696 |
|
}, |
|
{ |
|
"epoch": 1.4513778904022807, |
|
"grad_norm": 0.5827309489250183, |
|
"learning_rate": 9.27431105479886e-05, |
|
"loss": 0.0722, |
|
"step": 18328 |
|
}, |
|
{ |
|
"epoch": 1.5014254038644284, |
|
"grad_norm": 0.09628592431545258, |
|
"learning_rate": 9.249287298067785e-05, |
|
"loss": 0.0815, |
|
"step": 18960 |
|
}, |
|
{ |
|
"epoch": 1.551472917326576, |
|
"grad_norm": 1.4028997421264648, |
|
"learning_rate": 9.224263541336713e-05, |
|
"loss": 0.0758, |
|
"step": 19592 |
|
}, |
|
{ |
|
"epoch": 1.6015204307887236, |
|
"grad_norm": 0.37902089953422546, |
|
"learning_rate": 9.199239784605639e-05, |
|
"loss": 0.0772, |
|
"step": 20224 |
|
}, |
|
{ |
|
"epoch": 1.651567944250871, |
|
"grad_norm": 0.5193475484848022, |
|
"learning_rate": 9.174216027874564e-05, |
|
"loss": 0.0731, |
|
"step": 20856 |
|
}, |
|
{ |
|
"epoch": 1.7016154577130187, |
|
"grad_norm": 1.5854244232177734, |
|
"learning_rate": 9.149192271143491e-05, |
|
"loss": 0.0704, |
|
"step": 21488 |
|
}, |
|
{ |
|
"epoch": 1.7516629711751663, |
|
"grad_norm": 2.4236505031585693, |
|
"learning_rate": 9.124168514412418e-05, |
|
"loss": 0.0733, |
|
"step": 22120 |
|
}, |
|
{ |
|
"epoch": 1.8017104846373138, |
|
"grad_norm": 0.610543429851532, |
|
"learning_rate": 9.099144757681343e-05, |
|
"loss": 0.0712, |
|
"step": 22752 |
|
}, |
|
{ |
|
"epoch": 1.8517579980994614, |
|
"grad_norm": 0.8024447560310364, |
|
"learning_rate": 9.07412100095027e-05, |
|
"loss": 0.0758, |
|
"step": 23384 |
|
}, |
|
{ |
|
"epoch": 1.901805511561609, |
|
"grad_norm": 0.1649412214756012, |
|
"learning_rate": 9.049097244219195e-05, |
|
"loss": 0.0736, |
|
"step": 24016 |
|
}, |
|
{ |
|
"epoch": 1.9518530250237567, |
|
"grad_norm": 0.15900301933288574, |
|
"learning_rate": 9.024073487488122e-05, |
|
"loss": 0.0746, |
|
"step": 24648 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_gen_len": 12.34418098, |
|
"eval_loss": 0.057007092982530594, |
|
"eval_runtime": 547.57, |
|
"eval_samples_per_second": 97.197, |
|
"eval_steps_per_second": 1.519, |
|
"eval_wordacc": 0.91236707, |
|
"eval_wordacc_oov": 0.85835753, |
|
"step": 25256 |
|
}, |
|
{ |
|
"epoch": 2.0019005384859043, |
|
"grad_norm": 1.2868945598602295, |
|
"learning_rate": 8.999049730757049e-05, |
|
"loss": 0.0676, |
|
"step": 25280 |
|
}, |
|
{ |
|
"epoch": 2.051948051948052, |
|
"grad_norm": 0.5984334945678711, |
|
"learning_rate": 8.974025974025974e-05, |
|
"loss": 0.0627, |
|
"step": 25912 |
|
}, |
|
{ |
|
"epoch": 2.1019955654101996, |
|
"grad_norm": 0.043390534818172455, |
|
"learning_rate": 8.949002217294901e-05, |
|
"loss": 0.0583, |
|
"step": 26544 |
|
}, |
|
{ |
|
"epoch": 2.1520430788723472, |
|
"grad_norm": 0.544601321220398, |
|
"learning_rate": 8.923978460563826e-05, |
|
"loss": 0.0614, |
|
"step": 27176 |
|
}, |
|
{ |
|
"epoch": 2.202090592334495, |
|
"grad_norm": 0.22555088996887207, |
|
"learning_rate": 8.898954703832753e-05, |
|
"loss": 0.0713, |
|
"step": 27808 |
|
}, |
|
{ |
|
"epoch": 2.2521381057966425, |
|
"grad_norm": 2.1508424282073975, |
|
"learning_rate": 8.87393094710168e-05, |
|
"loss": 0.071, |
|
"step": 28440 |
|
}, |
|
{ |
|
"epoch": 2.30218561925879, |
|
"grad_norm": 0.2992984354496002, |
|
"learning_rate": 8.848907190370605e-05, |
|
"loss": 0.0655, |
|
"step": 29072 |
|
}, |
|
{ |
|
"epoch": 2.3522331327209374, |
|
"grad_norm": 1.4744491577148438, |
|
"learning_rate": 8.823883433639532e-05, |
|
"loss": 0.0697, |
|
"step": 29704 |
|
}, |
|
{ |
|
"epoch": 2.402280646183085, |
|
"grad_norm": 0.6834865808486938, |
|
"learning_rate": 8.798859676908457e-05, |
|
"loss": 0.0661, |
|
"step": 30336 |
|
}, |
|
{ |
|
"epoch": 2.4523281596452327, |
|
"grad_norm": 0.06805714964866638, |
|
"learning_rate": 8.773835920177384e-05, |
|
"loss": 0.0672, |
|
"step": 30968 |
|
}, |
|
{ |
|
"epoch": 2.5023756731073803, |
|
"grad_norm": 1.880346655845642, |
|
"learning_rate": 8.748812163446311e-05, |
|
"loss": 0.063, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 2.552423186569528, |
|
"grad_norm": 0.5247331857681274, |
|
"learning_rate": 8.723788406715236e-05, |
|
"loss": 0.0621, |
|
"step": 32232 |
|
}, |
|
{ |
|
"epoch": 2.6024707000316756, |
|
"grad_norm": 0.15831807255744934, |
|
"learning_rate": 8.698764649984163e-05, |
|
"loss": 0.0653, |
|
"step": 32864 |
|
}, |
|
{ |
|
"epoch": 2.652518213493823, |
|
"grad_norm": 0.04121825844049454, |
|
"learning_rate": 8.673740893253088e-05, |
|
"loss": 0.0642, |
|
"step": 33496 |
|
}, |
|
{ |
|
"epoch": 2.702565726955971, |
|
"grad_norm": 0.3549499809741974, |
|
"learning_rate": 8.648717136522015e-05, |
|
"loss": 0.0679, |
|
"step": 34128 |
|
}, |
|
{ |
|
"epoch": 2.7526132404181185, |
|
"grad_norm": 0.16000640392303467, |
|
"learning_rate": 8.623693379790942e-05, |
|
"loss": 0.0597, |
|
"step": 34760 |
|
}, |
|
{ |
|
"epoch": 2.802660753880266, |
|
"grad_norm": 0.6360165476799011, |
|
"learning_rate": 8.598669623059867e-05, |
|
"loss": 0.0666, |
|
"step": 35392 |
|
}, |
|
{ |
|
"epoch": 2.852708267342414, |
|
"grad_norm": 0.4808698296546936, |
|
"learning_rate": 8.573645866328793e-05, |
|
"loss": 0.0654, |
|
"step": 36024 |
|
}, |
|
{ |
|
"epoch": 2.9027557808045614, |
|
"grad_norm": 1.7070688009262085, |
|
"learning_rate": 8.54862210959772e-05, |
|
"loss": 0.0645, |
|
"step": 36656 |
|
}, |
|
{ |
|
"epoch": 2.952803294266709, |
|
"grad_norm": 0.3017909824848175, |
|
"learning_rate": 8.523598352866646e-05, |
|
"loss": 0.0622, |
|
"step": 37288 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_gen_len": 12.35115178, |
|
"eval_loss": 0.04925922676920891, |
|
"eval_runtime": 519.5392, |
|
"eval_samples_per_second": 102.441, |
|
"eval_steps_per_second": 1.601, |
|
"eval_wordacc": 0.9194506, |
|
"eval_wordacc_oov": 0.87168187, |
|
"step": 37884 |
|
}, |
|
{ |
|
"epoch": 3.0028508077288567, |
|
"grad_norm": 0.5821070671081543, |
|
"learning_rate": 8.498574596135573e-05, |
|
"loss": 0.0661, |
|
"step": 37920 |
|
}, |
|
{ |
|
"epoch": 3.052898321191004, |
|
"grad_norm": 0.4475654363632202, |
|
"learning_rate": 8.473550839404498e-05, |
|
"loss": 0.0574, |
|
"step": 38552 |
|
}, |
|
{ |
|
"epoch": 3.1029458346531515, |
|
"grad_norm": 1.4051363468170166, |
|
"learning_rate": 8.448527082673424e-05, |
|
"loss": 0.0614, |
|
"step": 39184 |
|
}, |
|
{ |
|
"epoch": 3.152993348115299, |
|
"grad_norm": 0.5270406603813171, |
|
"learning_rate": 8.423503325942352e-05, |
|
"loss": 0.059, |
|
"step": 39816 |
|
}, |
|
{ |
|
"epoch": 3.203040861577447, |
|
"grad_norm": 0.2516399621963501, |
|
"learning_rate": 8.398479569211277e-05, |
|
"loss": 0.0568, |
|
"step": 40448 |
|
}, |
|
{ |
|
"epoch": 3.2530883750395945, |
|
"grad_norm": 2.4771833419799805, |
|
"learning_rate": 8.373455812480203e-05, |
|
"loss": 0.055, |
|
"step": 41080 |
|
}, |
|
{ |
|
"epoch": 3.303135888501742, |
|
"grad_norm": 0.39709779620170593, |
|
"learning_rate": 8.348432055749129e-05, |
|
"loss": 0.0591, |
|
"step": 41712 |
|
}, |
|
{ |
|
"epoch": 3.3531834019638898, |
|
"grad_norm": 0.2029147893190384, |
|
"learning_rate": 8.323408299018055e-05, |
|
"loss": 0.0572, |
|
"step": 42344 |
|
}, |
|
{ |
|
"epoch": 3.4032309154260374, |
|
"grad_norm": 0.1706971675157547, |
|
"learning_rate": 8.298384542286983e-05, |
|
"loss": 0.0618, |
|
"step": 42976 |
|
}, |
|
{ |
|
"epoch": 3.453278428888185, |
|
"grad_norm": 0.14079004526138306, |
|
"learning_rate": 8.273360785555908e-05, |
|
"loss": 0.0583, |
|
"step": 43608 |
|
}, |
|
{ |
|
"epoch": 3.5033259423503327, |
|
"grad_norm": 0.8446473479270935, |
|
"learning_rate": 8.248337028824834e-05, |
|
"loss": 0.0582, |
|
"step": 44240 |
|
}, |
|
{ |
|
"epoch": 3.5533734558124803, |
|
"grad_norm": 1.3202866315841675, |
|
"learning_rate": 8.22331327209376e-05, |
|
"loss": 0.0594, |
|
"step": 44872 |
|
}, |
|
{ |
|
"epoch": 3.603420969274628, |
|
"grad_norm": 0.1289588212966919, |
|
"learning_rate": 8.198289515362686e-05, |
|
"loss": 0.0668, |
|
"step": 45504 |
|
}, |
|
{ |
|
"epoch": 3.653468482736775, |
|
"grad_norm": 0.7715129852294922, |
|
"learning_rate": 8.173265758631612e-05, |
|
"loss": 0.056, |
|
"step": 46136 |
|
}, |
|
{ |
|
"epoch": 3.703515996198923, |
|
"grad_norm": 0.8882943391799927, |
|
"learning_rate": 8.148242001900539e-05, |
|
"loss": 0.0585, |
|
"step": 46768 |
|
}, |
|
{ |
|
"epoch": 3.7535635096610704, |
|
"grad_norm": 0.5238478183746338, |
|
"learning_rate": 8.123218245169465e-05, |
|
"loss": 0.0523, |
|
"step": 47400 |
|
}, |
|
{ |
|
"epoch": 3.803611023123218, |
|
"grad_norm": 0.7119426727294922, |
|
"learning_rate": 8.098194488438391e-05, |
|
"loss": 0.0566, |
|
"step": 48032 |
|
}, |
|
{ |
|
"epoch": 3.8536585365853657, |
|
"grad_norm": 1.7843942642211914, |
|
"learning_rate": 8.073170731707318e-05, |
|
"loss": 0.0577, |
|
"step": 48664 |
|
}, |
|
{ |
|
"epoch": 3.9037060500475134, |
|
"grad_norm": 0.09263037890195847, |
|
"learning_rate": 8.048146974976244e-05, |
|
"loss": 0.0602, |
|
"step": 49296 |
|
}, |
|
{ |
|
"epoch": 3.953753563509661, |
|
"grad_norm": 0.8220856785774231, |
|
"learning_rate": 8.02312321824517e-05, |
|
"loss": 0.0584, |
|
"step": 49928 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_gen_len": 12.34404945, |
|
"eval_loss": 0.04651999473571777, |
|
"eval_runtime": 525.6595, |
|
"eval_samples_per_second": 101.248, |
|
"eval_steps_per_second": 1.583, |
|
"eval_wordacc": 0.92211867, |
|
"eval_wordacc_oov": 0.87494815, |
|
"step": 50512 |
|
}, |
|
{ |
|
"epoch": 4.003801076971809, |
|
"grad_norm": 0.09420084208250046, |
|
"learning_rate": 7.998099461514096e-05, |
|
"loss": 0.0559, |
|
"step": 50560 |
|
}, |
|
{ |
|
"epoch": 4.053848590433956, |
|
"grad_norm": 0.6216241121292114, |
|
"learning_rate": 7.973075704783022e-05, |
|
"loss": 0.0507, |
|
"step": 51192 |
|
}, |
|
{ |
|
"epoch": 4.103896103896104, |
|
"grad_norm": 0.6072413921356201, |
|
"learning_rate": 7.948051948051949e-05, |
|
"loss": 0.0508, |
|
"step": 51824 |
|
}, |
|
{ |
|
"epoch": 4.153943617358252, |
|
"grad_norm": 0.3261624872684479, |
|
"learning_rate": 7.923028191320875e-05, |
|
"loss": 0.0584, |
|
"step": 52456 |
|
}, |
|
{ |
|
"epoch": 4.203991130820399, |
|
"grad_norm": 3.996793508529663, |
|
"learning_rate": 7.898004434589801e-05, |
|
"loss": 0.0542, |
|
"step": 53088 |
|
}, |
|
{ |
|
"epoch": 4.254038644282547, |
|
"grad_norm": 0.09758679568767548, |
|
"learning_rate": 7.872980677858727e-05, |
|
"loss": 0.0543, |
|
"step": 53720 |
|
}, |
|
{ |
|
"epoch": 4.3040861577446945, |
|
"grad_norm": 0.8341479897499084, |
|
"learning_rate": 7.847956921127652e-05, |
|
"loss": 0.06, |
|
"step": 54352 |
|
}, |
|
{ |
|
"epoch": 4.354133671206842, |
|
"grad_norm": 0.3256041407585144, |
|
"learning_rate": 7.82293316439658e-05, |
|
"loss": 0.0547, |
|
"step": 54984 |
|
}, |
|
{ |
|
"epoch": 4.40418118466899, |
|
"grad_norm": 0.8251773715019226, |
|
"learning_rate": 7.797909407665506e-05, |
|
"loss": 0.0563, |
|
"step": 55616 |
|
}, |
|
{ |
|
"epoch": 4.454228698131137, |
|
"grad_norm": 0.23588858544826508, |
|
"learning_rate": 7.772885650934432e-05, |
|
"loss": 0.0566, |
|
"step": 56248 |
|
}, |
|
{ |
|
"epoch": 4.504276211593285, |
|
"grad_norm": 0.8622148633003235, |
|
"learning_rate": 7.747861894203358e-05, |
|
"loss": 0.0511, |
|
"step": 56880 |
|
}, |
|
{ |
|
"epoch": 4.554323725055433, |
|
"grad_norm": 0.29321447014808655, |
|
"learning_rate": 7.722838137472284e-05, |
|
"loss": 0.0502, |
|
"step": 57512 |
|
}, |
|
{ |
|
"epoch": 4.60437123851758, |
|
"grad_norm": 1.1889938116073608, |
|
"learning_rate": 7.697814380741211e-05, |
|
"loss": 0.0525, |
|
"step": 58144 |
|
}, |
|
{ |
|
"epoch": 4.654418751979728, |
|
"grad_norm": 0.5421351790428162, |
|
"learning_rate": 7.672790624010137e-05, |
|
"loss": 0.0509, |
|
"step": 58776 |
|
}, |
|
{ |
|
"epoch": 4.704466265441875, |
|
"grad_norm": 0.4371638000011444, |
|
"learning_rate": 7.647766867279062e-05, |
|
"loss": 0.0548, |
|
"step": 59408 |
|
}, |
|
{ |
|
"epoch": 4.754513778904023, |
|
"grad_norm": 0.06439998745918274, |
|
"learning_rate": 7.622743110547989e-05, |
|
"loss": 0.0553, |
|
"step": 60040 |
|
}, |
|
{ |
|
"epoch": 4.80456129236617, |
|
"grad_norm": 0.8337986469268799, |
|
"learning_rate": 7.597719353816916e-05, |
|
"loss": 0.0539, |
|
"step": 60672 |
|
}, |
|
{ |
|
"epoch": 4.854608805828318, |
|
"grad_norm": 0.5551128387451172, |
|
"learning_rate": 7.572695597085841e-05, |
|
"loss": 0.0531, |
|
"step": 61304 |
|
}, |
|
{ |
|
"epoch": 4.904656319290465, |
|
"grad_norm": 0.842311680316925, |
|
"learning_rate": 7.547671840354768e-05, |
|
"loss": 0.053, |
|
"step": 61936 |
|
}, |
|
{ |
|
"epoch": 4.954703832752613, |
|
"grad_norm": 0.6434153318405151, |
|
"learning_rate": 7.522648083623693e-05, |
|
"loss": 0.0497, |
|
"step": 62568 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_gen_len": 12.35519146, |
|
"eval_loss": 0.04361514747142792, |
|
"eval_runtime": 525.1326, |
|
"eval_samples_per_second": 101.35, |
|
"eval_steps_per_second": 1.584, |
|
"eval_wordacc": 0.92741723, |
|
"eval_wordacc_oov": 0.88205102, |
|
"step": 63140 |
|
}, |
|
{ |
|
"epoch": 5.004751346214761, |
|
"grad_norm": 0.11094748228788376, |
|
"learning_rate": 7.49762432689262e-05, |
|
"loss": 0.0509, |
|
"step": 63200 |
|
}, |
|
{ |
|
"epoch": 5.054798859676908, |
|
"grad_norm": 0.46878868341445923, |
|
"learning_rate": 7.472600570161547e-05, |
|
"loss": 0.0445, |
|
"step": 63832 |
|
}, |
|
{ |
|
"epoch": 5.104846373139056, |
|
"grad_norm": 1.7899694442749023, |
|
"learning_rate": 7.447576813430472e-05, |
|
"loss": 0.045, |
|
"step": 64464 |
|
}, |
|
{ |
|
"epoch": 5.1548938866012035, |
|
"grad_norm": 0.03743477538228035, |
|
"learning_rate": 7.422553056699399e-05, |
|
"loss": 0.0538, |
|
"step": 65096 |
|
}, |
|
{ |
|
"epoch": 5.204941400063351, |
|
"grad_norm": 0.9636221528053284, |
|
"learning_rate": 7.397529299968324e-05, |
|
"loss": 0.0444, |
|
"step": 65728 |
|
}, |
|
{ |
|
"epoch": 5.254988913525499, |
|
"grad_norm": 0.1024821326136589, |
|
"learning_rate": 7.372505543237251e-05, |
|
"loss": 0.0543, |
|
"step": 66360 |
|
}, |
|
{ |
|
"epoch": 5.305036426987646, |
|
"grad_norm": 0.6220707297325134, |
|
"learning_rate": 7.347481786506178e-05, |
|
"loss": 0.0537, |
|
"step": 66992 |
|
}, |
|
{ |
|
"epoch": 5.355083940449794, |
|
"grad_norm": 0.13747639954090118, |
|
"learning_rate": 7.322458029775103e-05, |
|
"loss": 0.0443, |
|
"step": 67624 |
|
}, |
|
{ |
|
"epoch": 5.405131453911942, |
|
"grad_norm": 0.25481894612312317, |
|
"learning_rate": 7.29743427304403e-05, |
|
"loss": 0.0544, |
|
"step": 68256 |
|
}, |
|
{ |
|
"epoch": 5.455178967374089, |
|
"grad_norm": 0.40640395879745483, |
|
"learning_rate": 7.272410516312955e-05, |
|
"loss": 0.0501, |
|
"step": 68888 |
|
}, |
|
{ |
|
"epoch": 5.505226480836237, |
|
"grad_norm": 0.017994888126850128, |
|
"learning_rate": 7.247386759581882e-05, |
|
"loss": 0.0497, |
|
"step": 69520 |
|
}, |
|
{ |
|
"epoch": 5.555273994298385, |
|
"grad_norm": 0.7399475574493408, |
|
"learning_rate": 7.222363002850809e-05, |
|
"loss": 0.0503, |
|
"step": 70152 |
|
}, |
|
{ |
|
"epoch": 5.605321507760532, |
|
"grad_norm": 0.1965421885251999, |
|
"learning_rate": 7.197339246119734e-05, |
|
"loss": 0.0494, |
|
"step": 70784 |
|
}, |
|
{ |
|
"epoch": 5.65536902122268, |
|
"grad_norm": 0.603735625743866, |
|
"learning_rate": 7.172315489388661e-05, |
|
"loss": 0.0492, |
|
"step": 71416 |
|
}, |
|
{ |
|
"epoch": 5.705416534684828, |
|
"grad_norm": 0.4098168909549713, |
|
"learning_rate": 7.147291732657586e-05, |
|
"loss": 0.0496, |
|
"step": 72048 |
|
}, |
|
{ |
|
"epoch": 5.755464048146975, |
|
"grad_norm": 0.026117555797100067, |
|
"learning_rate": 7.122267975926513e-05, |
|
"loss": 0.05, |
|
"step": 72680 |
|
}, |
|
{ |
|
"epoch": 5.805511561609123, |
|
"grad_norm": 0.03871222585439682, |
|
"learning_rate": 7.09724421919544e-05, |
|
"loss": 0.0508, |
|
"step": 73312 |
|
}, |
|
{ |
|
"epoch": 5.8555590750712705, |
|
"grad_norm": 1.100329041481018, |
|
"learning_rate": 7.072220462464365e-05, |
|
"loss": 0.0503, |
|
"step": 73944 |
|
}, |
|
{ |
|
"epoch": 5.905606588533418, |
|
"grad_norm": 0.8121901154518127, |
|
"learning_rate": 7.04719670573329e-05, |
|
"loss": 0.0468, |
|
"step": 74576 |
|
}, |
|
{ |
|
"epoch": 5.955654101995566, |
|
"grad_norm": 0.6457042694091797, |
|
"learning_rate": 7.022172949002219e-05, |
|
"loss": 0.0502, |
|
"step": 75208 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_gen_len": 12.35192214, |
|
"eval_loss": 0.041068777441978455, |
|
"eval_runtime": 522.1947, |
|
"eval_samples_per_second": 101.92, |
|
"eval_steps_per_second": 1.593, |
|
"eval_wordacc": 0.93108113, |
|
"eval_wordacc_oov": 0.88583575, |
|
"step": 75768 |
|
}, |
|
{ |
|
"epoch": 6.005701615457713, |
|
"grad_norm": 0.20237529277801514, |
|
"learning_rate": 6.997149192271144e-05, |
|
"loss": 0.0496, |
|
"step": 75840 |
|
}, |
|
{ |
|
"epoch": 6.055749128919861, |
|
"grad_norm": 0.10558341443538666, |
|
"learning_rate": 6.97212543554007e-05, |
|
"loss": 0.0422, |
|
"step": 76472 |
|
}, |
|
{ |
|
"epoch": 6.105796642382008, |
|
"grad_norm": 0.38815170526504517, |
|
"learning_rate": 6.947101678808996e-05, |
|
"loss": 0.0468, |
|
"step": 77104 |
|
}, |
|
{ |
|
"epoch": 6.1558441558441555, |
|
"grad_norm": 0.042554233223199844, |
|
"learning_rate": 6.922077922077921e-05, |
|
"loss": 0.0446, |
|
"step": 77736 |
|
}, |
|
{ |
|
"epoch": 6.205891669306303, |
|
"grad_norm": 1.2894216775894165, |
|
"learning_rate": 6.89705416534685e-05, |
|
"loss": 0.0462, |
|
"step": 78368 |
|
}, |
|
{ |
|
"epoch": 6.255939182768451, |
|
"grad_norm": 0.7259889841079712, |
|
"learning_rate": 6.872030408615775e-05, |
|
"loss": 0.045, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 6.305986696230598, |
|
"grad_norm": 0.5286532044410706, |
|
"learning_rate": 6.8470066518847e-05, |
|
"loss": 0.0446, |
|
"step": 79632 |
|
}, |
|
{ |
|
"epoch": 6.356034209692746, |
|
"grad_norm": 0.20671215653419495, |
|
"learning_rate": 6.821982895153627e-05, |
|
"loss": 0.0442, |
|
"step": 80264 |
|
}, |
|
{ |
|
"epoch": 6.406081723154894, |
|
"grad_norm": 0.2522045969963074, |
|
"learning_rate": 6.796959138422552e-05, |
|
"loss": 0.0455, |
|
"step": 80896 |
|
}, |
|
{ |
|
"epoch": 6.456129236617041, |
|
"grad_norm": 0.09235669672489166, |
|
"learning_rate": 6.771935381691479e-05, |
|
"loss": 0.0463, |
|
"step": 81528 |
|
}, |
|
{ |
|
"epoch": 6.506176750079189, |
|
"grad_norm": 1.3369249105453491, |
|
"learning_rate": 6.746911624960406e-05, |
|
"loss": 0.0467, |
|
"step": 82160 |
|
}, |
|
{ |
|
"epoch": 6.556224263541337, |
|
"grad_norm": 0.1361207365989685, |
|
"learning_rate": 6.721887868229331e-05, |
|
"loss": 0.0464, |
|
"step": 82792 |
|
}, |
|
{ |
|
"epoch": 6.606271777003484, |
|
"grad_norm": 0.057785116136074066, |
|
"learning_rate": 6.696864111498258e-05, |
|
"loss": 0.0471, |
|
"step": 83424 |
|
}, |
|
{ |
|
"epoch": 6.656319290465632, |
|
"grad_norm": 0.5493625998497009, |
|
"learning_rate": 6.671840354767185e-05, |
|
"loss": 0.0493, |
|
"step": 84056 |
|
}, |
|
{ |
|
"epoch": 6.7063668039277795, |
|
"grad_norm": 0.08289259672164917, |
|
"learning_rate": 6.64681659803611e-05, |
|
"loss": 0.0491, |
|
"step": 84688 |
|
}, |
|
{ |
|
"epoch": 6.756414317389927, |
|
"grad_norm": 0.08197712898254395, |
|
"learning_rate": 6.621792841305037e-05, |
|
"loss": 0.0487, |
|
"step": 85320 |
|
}, |
|
{ |
|
"epoch": 6.806461830852075, |
|
"grad_norm": 0.1574297547340393, |
|
"learning_rate": 6.596769084573962e-05, |
|
"loss": 0.042, |
|
"step": 85952 |
|
}, |
|
{ |
|
"epoch": 6.856509344314222, |
|
"grad_norm": 0.07021531462669373, |
|
"learning_rate": 6.571745327842889e-05, |
|
"loss": 0.0469, |
|
"step": 86584 |
|
}, |
|
{ |
|
"epoch": 6.90655685777637, |
|
"grad_norm": 0.5342025756835938, |
|
"learning_rate": 6.546721571111816e-05, |
|
"loss": 0.0458, |
|
"step": 87216 |
|
}, |
|
{ |
|
"epoch": 6.956604371238518, |
|
"grad_norm": 0.754435658454895, |
|
"learning_rate": 6.521697814380741e-05, |
|
"loss": 0.0428, |
|
"step": 87848 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_gen_len": 12.34435008, |
|
"eval_loss": 0.039582036435604095, |
|
"eval_runtime": 513.2393, |
|
"eval_samples_per_second": 103.698, |
|
"eval_steps_per_second": 1.621, |
|
"eval_wordacc": 0.93356131, |
|
"eval_wordacc_oov": 0.88780589, |
|
"step": 88396 |
|
}, |
|
{ |
|
"epoch": 7.006651884700665, |
|
"grad_norm": 1.7553069591522217, |
|
"learning_rate": 6.496674057649668e-05, |
|
"loss": 0.0429, |
|
"step": 88480 |
|
}, |
|
{ |
|
"epoch": 7.056699398162813, |
|
"grad_norm": 0.15994039177894592, |
|
"learning_rate": 6.471650300918593e-05, |
|
"loss": 0.0451, |
|
"step": 89112 |
|
}, |
|
{ |
|
"epoch": 7.106746911624961, |
|
"grad_norm": 0.21786805987358093, |
|
"learning_rate": 6.44662654418752e-05, |
|
"loss": 0.0415, |
|
"step": 89744 |
|
}, |
|
{ |
|
"epoch": 7.156794425087108, |
|
"grad_norm": 0.07405902445316315, |
|
"learning_rate": 6.421602787456447e-05, |
|
"loss": 0.0408, |
|
"step": 90376 |
|
}, |
|
{ |
|
"epoch": 7.206841938549256, |
|
"grad_norm": 0.1853848397731781, |
|
"learning_rate": 6.396579030725372e-05, |
|
"loss": 0.0415, |
|
"step": 91008 |
|
}, |
|
{ |
|
"epoch": 7.256889452011404, |
|
"grad_norm": 0.41366642713546753, |
|
"learning_rate": 6.371555273994299e-05, |
|
"loss": 0.046, |
|
"step": 91640 |
|
}, |
|
{ |
|
"epoch": 7.306936965473551, |
|
"grad_norm": 0.2615118622779846, |
|
"learning_rate": 6.346531517263224e-05, |
|
"loss": 0.0413, |
|
"step": 92272 |
|
}, |
|
{ |
|
"epoch": 7.356984478935699, |
|
"grad_norm": 0.06805741786956787, |
|
"learning_rate": 6.321507760532151e-05, |
|
"loss": 0.0436, |
|
"step": 92904 |
|
}, |
|
{ |
|
"epoch": 7.407031992397846, |
|
"grad_norm": 1.3070762157440186, |
|
"learning_rate": 6.296484003801078e-05, |
|
"loss": 0.0464, |
|
"step": 93536 |
|
}, |
|
{ |
|
"epoch": 7.457079505859994, |
|
"grad_norm": 0.11481507122516632, |
|
"learning_rate": 6.271460247070003e-05, |
|
"loss": 0.0442, |
|
"step": 94168 |
|
}, |
|
{ |
|
"epoch": 7.507127019322141, |
|
"grad_norm": 0.9575181603431702, |
|
"learning_rate": 6.246436490338929e-05, |
|
"loss": 0.043, |
|
"step": 94800 |
|
}, |
|
{ |
|
"epoch": 7.5571745327842885, |
|
"grad_norm": 0.2559140920639038, |
|
"learning_rate": 6.221412733607856e-05, |
|
"loss": 0.0375, |
|
"step": 95432 |
|
}, |
|
{ |
|
"epoch": 7.607222046246436, |
|
"grad_norm": 0.043967317789793015, |
|
"learning_rate": 6.196388976876782e-05, |
|
"loss": 0.0412, |
|
"step": 96064 |
|
}, |
|
{ |
|
"epoch": 7.657269559708584, |
|
"grad_norm": 0.5306654572486877, |
|
"learning_rate": 6.171365220145709e-05, |
|
"loss": 0.0449, |
|
"step": 96696 |
|
}, |
|
{ |
|
"epoch": 7.7073170731707314, |
|
"grad_norm": 0.03713352233171463, |
|
"learning_rate": 6.146341463414634e-05, |
|
"loss": 0.0443, |
|
"step": 97328 |
|
}, |
|
{ |
|
"epoch": 7.757364586632879, |
|
"grad_norm": 0.6375567317008972, |
|
"learning_rate": 6.12131770668356e-05, |
|
"loss": 0.0441, |
|
"step": 97960 |
|
}, |
|
{ |
|
"epoch": 7.807412100095027, |
|
"grad_norm": 0.9398515224456787, |
|
"learning_rate": 6.096293949952487e-05, |
|
"loss": 0.0421, |
|
"step": 98592 |
|
}, |
|
{ |
|
"epoch": 7.857459613557174, |
|
"grad_norm": 0.6615290641784668, |
|
"learning_rate": 6.071270193221413e-05, |
|
"loss": 0.0468, |
|
"step": 99224 |
|
}, |
|
{ |
|
"epoch": 7.907507127019322, |
|
"grad_norm": 0.48468518257141113, |
|
"learning_rate": 6.0462464364903394e-05, |
|
"loss": 0.0427, |
|
"step": 99856 |
|
}, |
|
{ |
|
"epoch": 7.95755464048147, |
|
"grad_norm": 0.967910647392273, |
|
"learning_rate": 6.0212226797592654e-05, |
|
"loss": 0.0416, |
|
"step": 100488 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_gen_len": 12.34714967, |
|
"eval_loss": 0.0372321754693985, |
|
"eval_runtime": 517.1292, |
|
"eval_samples_per_second": 102.918, |
|
"eval_steps_per_second": 1.609, |
|
"eval_wordacc": 0.93393709, |
|
"eval_wordacc_oov": 0.88868727, |
|
"step": 101024 |
|
}, |
|
{ |
|
"epoch": 8.007602153943617, |
|
"grad_norm": 0.5545419454574585, |
|
"learning_rate": 5.9961989230281915e-05, |
|
"loss": 0.0447, |
|
"step": 101120 |
|
}, |
|
{ |
|
"epoch": 8.057649667405766, |
|
"grad_norm": 1.7096141576766968, |
|
"learning_rate": 5.971175166297118e-05, |
|
"loss": 0.038, |
|
"step": 101752 |
|
}, |
|
{ |
|
"epoch": 8.107697180867913, |
|
"grad_norm": 0.06642602384090424, |
|
"learning_rate": 5.946151409566044e-05, |
|
"loss": 0.0395, |
|
"step": 102384 |
|
}, |
|
{ |
|
"epoch": 8.15774469433006, |
|
"grad_norm": 0.15396763384342194, |
|
"learning_rate": 5.9211276528349704e-05, |
|
"loss": 0.0411, |
|
"step": 103016 |
|
}, |
|
{ |
|
"epoch": 8.207792207792208, |
|
"grad_norm": 1.0655204057693481, |
|
"learning_rate": 5.8961038961038965e-05, |
|
"loss": 0.0416, |
|
"step": 103648 |
|
}, |
|
{ |
|
"epoch": 8.257839721254355, |
|
"grad_norm": 0.42243492603302, |
|
"learning_rate": 5.871080139372822e-05, |
|
"loss": 0.0445, |
|
"step": 104280 |
|
}, |
|
{ |
|
"epoch": 8.307887234716503, |
|
"grad_norm": 0.05212310701608658, |
|
"learning_rate": 5.846056382641749e-05, |
|
"loss": 0.0392, |
|
"step": 104912 |
|
}, |
|
{ |
|
"epoch": 8.35793474817865, |
|
"grad_norm": 0.6314841508865356, |
|
"learning_rate": 5.8210326259106754e-05, |
|
"loss": 0.0392, |
|
"step": 105544 |
|
}, |
|
{ |
|
"epoch": 8.407982261640798, |
|
"grad_norm": 0.4014628827571869, |
|
"learning_rate": 5.796008869179601e-05, |
|
"loss": 0.0387, |
|
"step": 106176 |
|
}, |
|
{ |
|
"epoch": 8.458029775102945, |
|
"grad_norm": 0.13107645511627197, |
|
"learning_rate": 5.770985112448527e-05, |
|
"loss": 0.0425, |
|
"step": 106808 |
|
}, |
|
{ |
|
"epoch": 8.508077288565094, |
|
"grad_norm": 0.1390117108821869, |
|
"learning_rate": 5.745961355717454e-05, |
|
"loss": 0.038, |
|
"step": 107440 |
|
}, |
|
{ |
|
"epoch": 8.55812480202724, |
|
"grad_norm": 0.1780831664800644, |
|
"learning_rate": 5.72093759898638e-05, |
|
"loss": 0.0397, |
|
"step": 108072 |
|
}, |
|
{ |
|
"epoch": 8.608172315489389, |
|
"grad_norm": 0.8216772675514221, |
|
"learning_rate": 5.695913842255306e-05, |
|
"loss": 0.0377, |
|
"step": 108704 |
|
}, |
|
{ |
|
"epoch": 8.658219828951536, |
|
"grad_norm": 0.040023334324359894, |
|
"learning_rate": 5.670890085524232e-05, |
|
"loss": 0.0395, |
|
"step": 109336 |
|
}, |
|
{ |
|
"epoch": 8.708267342413684, |
|
"grad_norm": 0.7334257364273071, |
|
"learning_rate": 5.645866328793158e-05, |
|
"loss": 0.0401, |
|
"step": 109968 |
|
}, |
|
{ |
|
"epoch": 8.758314855875831, |
|
"grad_norm": 0.09213205426931381, |
|
"learning_rate": 5.620842572062085e-05, |
|
"loss": 0.038, |
|
"step": 110600 |
|
}, |
|
{ |
|
"epoch": 8.80836236933798, |
|
"grad_norm": 0.8264344930648804, |
|
"learning_rate": 5.595818815331011e-05, |
|
"loss": 0.0404, |
|
"step": 111232 |
|
}, |
|
{ |
|
"epoch": 8.858409882800126, |
|
"grad_norm": 0.4497428238391876, |
|
"learning_rate": 5.570795058599937e-05, |
|
"loss": 0.0417, |
|
"step": 111864 |
|
}, |
|
{ |
|
"epoch": 8.908457396262275, |
|
"grad_norm": 0.02390374056994915, |
|
"learning_rate": 5.545771301868863e-05, |
|
"loss": 0.0351, |
|
"step": 112496 |
|
}, |
|
{ |
|
"epoch": 8.958504909724422, |
|
"grad_norm": 0.1745648831129074, |
|
"learning_rate": 5.520747545137789e-05, |
|
"loss": 0.042, |
|
"step": 113128 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_gen_len": 12.3485025, |
|
"eval_loss": 0.03648155927658081, |
|
"eval_runtime": 514.9505, |
|
"eval_samples_per_second": 103.354, |
|
"eval_steps_per_second": 1.616, |
|
"eval_wordacc": 0.93964902, |
|
"eval_wordacc_oov": 0.89444214, |
|
"step": 113652 |
|
}, |
|
{ |
|
"epoch": 9.00855242318657, |
|
"grad_norm": 0.7463224530220032, |
|
"learning_rate": 5.4957237884067156e-05, |
|
"loss": 0.0393, |
|
"step": 113760 |
|
}, |
|
{ |
|
"epoch": 9.058599936648717, |
|
"grad_norm": 0.2374447137117386, |
|
"learning_rate": 5.470700031675642e-05, |
|
"loss": 0.0349, |
|
"step": 114392 |
|
}, |
|
{ |
|
"epoch": 9.108647450110865, |
|
"grad_norm": 1.1859426498413086, |
|
"learning_rate": 5.445676274944568e-05, |
|
"loss": 0.039, |
|
"step": 115024 |
|
}, |
|
{ |
|
"epoch": 9.158694963573012, |
|
"grad_norm": 0.03181586042046547, |
|
"learning_rate": 5.420652518213494e-05, |
|
"loss": 0.0354, |
|
"step": 115656 |
|
}, |
|
{ |
|
"epoch": 9.20874247703516, |
|
"grad_norm": 0.0614316463470459, |
|
"learning_rate": 5.3956287614824206e-05, |
|
"loss": 0.0337, |
|
"step": 116288 |
|
}, |
|
{ |
|
"epoch": 9.258789990497307, |
|
"grad_norm": 0.031755685806274414, |
|
"learning_rate": 5.370605004751347e-05, |
|
"loss": 0.0346, |
|
"step": 116920 |
|
}, |
|
{ |
|
"epoch": 9.308837503959456, |
|
"grad_norm": 0.4263891875743866, |
|
"learning_rate": 5.345581248020273e-05, |
|
"loss": 0.0365, |
|
"step": 117552 |
|
}, |
|
{ |
|
"epoch": 9.358885017421603, |
|
"grad_norm": 0.7516904473304749, |
|
"learning_rate": 5.320557491289199e-05, |
|
"loss": 0.0394, |
|
"step": 118184 |
|
}, |
|
{ |
|
"epoch": 9.408932530883751, |
|
"grad_norm": 0.547282338142395, |
|
"learning_rate": 5.295533734558125e-05, |
|
"loss": 0.037, |
|
"step": 118816 |
|
}, |
|
{ |
|
"epoch": 9.458980044345898, |
|
"grad_norm": 0.048598043620586395, |
|
"learning_rate": 5.2705099778270516e-05, |
|
"loss": 0.0396, |
|
"step": 119448 |
|
}, |
|
{ |
|
"epoch": 9.509027557808047, |
|
"grad_norm": 0.019015343859791756, |
|
"learning_rate": 5.245486221095978e-05, |
|
"loss": 0.0377, |
|
"step": 120080 |
|
}, |
|
{ |
|
"epoch": 9.559075071270193, |
|
"grad_norm": 0.42539820075035095, |
|
"learning_rate": 5.220462464364904e-05, |
|
"loss": 0.0361, |
|
"step": 120712 |
|
}, |
|
{ |
|
"epoch": 9.60912258473234, |
|
"grad_norm": 0.5751402974128723, |
|
"learning_rate": 5.19543870763383e-05, |
|
"loss": 0.0407, |
|
"step": 121344 |
|
}, |
|
{ |
|
"epoch": 9.659170098194489, |
|
"grad_norm": 1.380823016166687, |
|
"learning_rate": 5.170414950902755e-05, |
|
"loss": 0.0388, |
|
"step": 121976 |
|
}, |
|
{ |
|
"epoch": 9.709217611656635, |
|
"grad_norm": 0.6849233508110046, |
|
"learning_rate": 5.145391194171683e-05, |
|
"loss": 0.0387, |
|
"step": 122608 |
|
}, |
|
{ |
|
"epoch": 9.759265125118784, |
|
"grad_norm": 1.4156357049942017, |
|
"learning_rate": 5.120367437440609e-05, |
|
"loss": 0.0373, |
|
"step": 123240 |
|
}, |
|
{ |
|
"epoch": 9.80931263858093, |
|
"grad_norm": 0.18062171339988708, |
|
"learning_rate": 5.095343680709535e-05, |
|
"loss": 0.0333, |
|
"step": 123872 |
|
}, |
|
{ |
|
"epoch": 9.85936015204308, |
|
"grad_norm": 0.6566870212554932, |
|
"learning_rate": 5.07031992397846e-05, |
|
"loss": 0.0367, |
|
"step": 124504 |
|
}, |
|
{ |
|
"epoch": 9.909407665505226, |
|
"grad_norm": 0.2741030156612396, |
|
"learning_rate": 5.0452961672473876e-05, |
|
"loss": 0.0426, |
|
"step": 125136 |
|
}, |
|
{ |
|
"epoch": 9.959455178967374, |
|
"grad_norm": 0.7864658236503601, |
|
"learning_rate": 5.020272410516314e-05, |
|
"loss": 0.0376, |
|
"step": 125768 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_gen_len": 12.34852129, |
|
"eval_loss": 0.03532838076353073, |
|
"eval_runtime": 515.5211, |
|
"eval_samples_per_second": 103.239, |
|
"eval_steps_per_second": 1.614, |
|
"eval_wordacc": 0.94117094, |
|
"eval_wordacc_oov": 0.89620489, |
|
"step": 126280 |
|
}, |
|
{ |
|
"epoch": 10.009502692429521, |
|
"grad_norm": 0.08346331119537354, |
|
"learning_rate": 4.995248653785239e-05, |
|
"loss": 0.0358, |
|
"step": 126400 |
|
}, |
|
{ |
|
"epoch": 10.05955020589167, |
|
"grad_norm": 0.1552925556898117, |
|
"learning_rate": 4.970224897054165e-05, |
|
"loss": 0.0286, |
|
"step": 127032 |
|
}, |
|
{ |
|
"epoch": 10.109597719353816, |
|
"grad_norm": 0.017832357436418533, |
|
"learning_rate": 4.945201140323092e-05, |
|
"loss": 0.0342, |
|
"step": 127664 |
|
}, |
|
{ |
|
"epoch": 10.159645232815965, |
|
"grad_norm": 0.7822960019111633, |
|
"learning_rate": 4.920177383592018e-05, |
|
"loss": 0.0342, |
|
"step": 128296 |
|
}, |
|
{ |
|
"epoch": 10.209692746278112, |
|
"grad_norm": 0.8483818173408508, |
|
"learning_rate": 4.895153626860944e-05, |
|
"loss": 0.036, |
|
"step": 128928 |
|
}, |
|
{ |
|
"epoch": 10.25974025974026, |
|
"grad_norm": 0.12484422326087952, |
|
"learning_rate": 4.87012987012987e-05, |
|
"loss": 0.0353, |
|
"step": 129560 |
|
}, |
|
{ |
|
"epoch": 10.309787773202407, |
|
"grad_norm": 1.0866436958312988, |
|
"learning_rate": 4.845106113398797e-05, |
|
"loss": 0.0336, |
|
"step": 130192 |
|
}, |
|
{ |
|
"epoch": 10.359835286664556, |
|
"grad_norm": 2.065387487411499, |
|
"learning_rate": 4.820082356667723e-05, |
|
"loss": 0.036, |
|
"step": 130824 |
|
}, |
|
{ |
|
"epoch": 10.409882800126702, |
|
"grad_norm": 0.13544411957263947, |
|
"learning_rate": 4.795058599936649e-05, |
|
"loss": 0.0353, |
|
"step": 131456 |
|
}, |
|
{ |
|
"epoch": 10.45993031358885, |
|
"grad_norm": 0.8644410967826843, |
|
"learning_rate": 4.770034843205575e-05, |
|
"loss": 0.0366, |
|
"step": 132088 |
|
}, |
|
{ |
|
"epoch": 10.509977827050998, |
|
"grad_norm": 0.044758204370737076, |
|
"learning_rate": 4.745011086474501e-05, |
|
"loss": 0.0337, |
|
"step": 132720 |
|
}, |
|
{ |
|
"epoch": 10.560025340513146, |
|
"grad_norm": 0.07954395562410355, |
|
"learning_rate": 4.719987329743428e-05, |
|
"loss": 0.038, |
|
"step": 133352 |
|
}, |
|
{ |
|
"epoch": 10.610072853975293, |
|
"grad_norm": 0.15263523161411285, |
|
"learning_rate": 4.694963573012354e-05, |
|
"loss": 0.037, |
|
"step": 133984 |
|
}, |
|
{ |
|
"epoch": 10.660120367437441, |
|
"grad_norm": 0.4090266227722168, |
|
"learning_rate": 4.66993981628128e-05, |
|
"loss": 0.0341, |
|
"step": 134616 |
|
}, |
|
{ |
|
"epoch": 10.710167880899588, |
|
"grad_norm": 0.3766542077064514, |
|
"learning_rate": 4.644916059550206e-05, |
|
"loss": 0.0395, |
|
"step": 135248 |
|
}, |
|
{ |
|
"epoch": 10.760215394361737, |
|
"grad_norm": 0.044228482991456985, |
|
"learning_rate": 4.619892302819132e-05, |
|
"loss": 0.0354, |
|
"step": 135880 |
|
}, |
|
{ |
|
"epoch": 10.810262907823883, |
|
"grad_norm": 3.4171650409698486, |
|
"learning_rate": 4.594868546088059e-05, |
|
"loss": 0.0344, |
|
"step": 136512 |
|
}, |
|
{ |
|
"epoch": 10.86031042128603, |
|
"grad_norm": 0.1112111434340477, |
|
"learning_rate": 4.569844789356984e-05, |
|
"loss": 0.0361, |
|
"step": 137144 |
|
}, |
|
{ |
|
"epoch": 10.910357934748179, |
|
"grad_norm": 0.09063247591257095, |
|
"learning_rate": 4.544821032625911e-05, |
|
"loss": 0.0325, |
|
"step": 137776 |
|
}, |
|
{ |
|
"epoch": 10.960405448210325, |
|
"grad_norm": 0.2144654095172882, |
|
"learning_rate": 4.519797275894837e-05, |
|
"loss": 0.031, |
|
"step": 138408 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_gen_len": 12.35186577, |
|
"eval_loss": 0.03388630226254463, |
|
"eval_runtime": 519.1861, |
|
"eval_samples_per_second": 102.51, |
|
"eval_steps_per_second": 1.603, |
|
"eval_wordacc": 0.94387659, |
|
"eval_wordacc_oov": 0.90081916, |
|
"step": 138908 |
|
}, |
|
{ |
|
"epoch": 11.010452961672474, |
|
"grad_norm": 0.9230628609657288, |
|
"learning_rate": 4.494773519163763e-05, |
|
"loss": 0.0336, |
|
"step": 139040 |
|
}, |
|
{ |
|
"epoch": 11.06050047513462, |
|
"grad_norm": 0.04931863397359848, |
|
"learning_rate": 4.469749762432689e-05, |
|
"loss": 0.0332, |
|
"step": 139672 |
|
}, |
|
{ |
|
"epoch": 11.11054798859677, |
|
"grad_norm": 0.43372172117233276, |
|
"learning_rate": 4.4447260057016154e-05, |
|
"loss": 0.0303, |
|
"step": 140304 |
|
}, |
|
{ |
|
"epoch": 11.160595502058916, |
|
"grad_norm": 1.8216339349746704, |
|
"learning_rate": 4.419702248970542e-05, |
|
"loss": 0.0327, |
|
"step": 140936 |
|
}, |
|
{ |
|
"epoch": 11.210643015521065, |
|
"grad_norm": 0.4168206453323364, |
|
"learning_rate": 4.394678492239468e-05, |
|
"loss": 0.0299, |
|
"step": 141568 |
|
}, |
|
{ |
|
"epoch": 11.260690528983211, |
|
"grad_norm": 0.2748374938964844, |
|
"learning_rate": 4.369654735508394e-05, |
|
"loss": 0.031, |
|
"step": 142200 |
|
}, |
|
{ |
|
"epoch": 11.31073804244536, |
|
"grad_norm": 1.0721484422683716, |
|
"learning_rate": 4.34463097877732e-05, |
|
"loss": 0.0338, |
|
"step": 142832 |
|
}, |
|
{ |
|
"epoch": 11.360785555907507, |
|
"grad_norm": 0.6880238056182861, |
|
"learning_rate": 4.319607222046247e-05, |
|
"loss": 0.0306, |
|
"step": 143464 |
|
}, |
|
{ |
|
"epoch": 11.410833069369655, |
|
"grad_norm": 0.6363319158554077, |
|
"learning_rate": 4.294583465315173e-05, |
|
"loss": 0.0324, |
|
"step": 144096 |
|
}, |
|
{ |
|
"epoch": 11.460880582831802, |
|
"grad_norm": 0.9793805480003357, |
|
"learning_rate": 4.2695597085840985e-05, |
|
"loss": 0.0352, |
|
"step": 144728 |
|
}, |
|
{ |
|
"epoch": 11.51092809629395, |
|
"grad_norm": 0.05823446065187454, |
|
"learning_rate": 4.244535951853025e-05, |
|
"loss": 0.0302, |
|
"step": 145360 |
|
}, |
|
{ |
|
"epoch": 11.560975609756097, |
|
"grad_norm": 0.44829338788986206, |
|
"learning_rate": 4.2195121951219514e-05, |
|
"loss": 0.0317, |
|
"step": 145992 |
|
}, |
|
{ |
|
"epoch": 11.611023123218246, |
|
"grad_norm": 0.21787405014038086, |
|
"learning_rate": 4.194488438390878e-05, |
|
"loss": 0.0327, |
|
"step": 146624 |
|
}, |
|
{ |
|
"epoch": 11.661070636680392, |
|
"grad_norm": 0.09327682852745056, |
|
"learning_rate": 4.1694646816598035e-05, |
|
"loss": 0.0343, |
|
"step": 147256 |
|
}, |
|
{ |
|
"epoch": 11.711118150142541, |
|
"grad_norm": 0.4916613698005676, |
|
"learning_rate": 4.14444092492873e-05, |
|
"loss": 0.0358, |
|
"step": 147888 |
|
}, |
|
{ |
|
"epoch": 11.761165663604688, |
|
"grad_norm": 0.04805804416537285, |
|
"learning_rate": 4.119417168197656e-05, |
|
"loss": 0.0343, |
|
"step": 148520 |
|
}, |
|
{ |
|
"epoch": 11.811213177066836, |
|
"grad_norm": 0.019468722864985466, |
|
"learning_rate": 4.0943934114665824e-05, |
|
"loss": 0.0298, |
|
"step": 149152 |
|
}, |
|
{ |
|
"epoch": 11.861260690528983, |
|
"grad_norm": 0.09931553900241852, |
|
"learning_rate": 4.0693696547355085e-05, |
|
"loss": 0.0304, |
|
"step": 149784 |
|
}, |
|
{ |
|
"epoch": 11.911308203991132, |
|
"grad_norm": 0.03299326449632645, |
|
"learning_rate": 4.0443458980044345e-05, |
|
"loss": 0.0355, |
|
"step": 150416 |
|
}, |
|
{ |
|
"epoch": 11.961355717453278, |
|
"grad_norm": 0.11701735109090805, |
|
"learning_rate": 4.019322141273361e-05, |
|
"loss": 0.0298, |
|
"step": 151048 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_gen_len": 12.34792003, |
|
"eval_loss": 0.03365420550107956, |
|
"eval_runtime": 512.0738, |
|
"eval_samples_per_second": 103.934, |
|
"eval_steps_per_second": 1.625, |
|
"eval_wordacc": 0.94539852, |
|
"eval_wordacc_oov": 0.90128577, |
|
"step": 151536 |
|
}, |
|
{ |
|
"epoch": 12.011403230915427, |
|
"grad_norm": 0.14153118431568146, |
|
"learning_rate": 3.9942983845422874e-05, |
|
"loss": 0.0294, |
|
"step": 151680 |
|
}, |
|
{ |
|
"epoch": 12.061450744377574, |
|
"grad_norm": 0.8496013283729553, |
|
"learning_rate": 3.9692746278112134e-05, |
|
"loss": 0.0302, |
|
"step": 152312 |
|
}, |
|
{ |
|
"epoch": 12.111498257839722, |
|
"grad_norm": 0.3620108664035797, |
|
"learning_rate": 3.9442508710801395e-05, |
|
"loss": 0.0293, |
|
"step": 152944 |
|
}, |
|
{ |
|
"epoch": 12.161545771301869, |
|
"grad_norm": 0.8358076810836792, |
|
"learning_rate": 3.9192271143490656e-05, |
|
"loss": 0.0279, |
|
"step": 153576 |
|
}, |
|
{ |
|
"epoch": 12.211593284764016, |
|
"grad_norm": 2.183274745941162, |
|
"learning_rate": 3.894203357617992e-05, |
|
"loss": 0.0298, |
|
"step": 154208 |
|
}, |
|
{ |
|
"epoch": 12.261640798226164, |
|
"grad_norm": 0.11177387088537216, |
|
"learning_rate": 3.869179600886918e-05, |
|
"loss": 0.0279, |
|
"step": 154840 |
|
}, |
|
{ |
|
"epoch": 12.311688311688311, |
|
"grad_norm": 0.16532179713249207, |
|
"learning_rate": 3.8441558441558445e-05, |
|
"loss": 0.0283, |
|
"step": 155472 |
|
}, |
|
{ |
|
"epoch": 12.36173582515046, |
|
"grad_norm": 0.2667466104030609, |
|
"learning_rate": 3.8191320874247705e-05, |
|
"loss": 0.0324, |
|
"step": 156104 |
|
}, |
|
{ |
|
"epoch": 12.411783338612606, |
|
"grad_norm": 1.4449559450149536, |
|
"learning_rate": 3.794108330693697e-05, |
|
"loss": 0.0327, |
|
"step": 156736 |
|
}, |
|
{ |
|
"epoch": 12.461830852074755, |
|
"grad_norm": 1.9251255989074707, |
|
"learning_rate": 3.769084573962623e-05, |
|
"loss": 0.0293, |
|
"step": 157368 |
|
}, |
|
{ |
|
"epoch": 12.511878365536901, |
|
"grad_norm": 1.981763243675232, |
|
"learning_rate": 3.744060817231549e-05, |
|
"loss": 0.0296, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 12.56192587899905, |
|
"grad_norm": 0.5154635906219482, |
|
"learning_rate": 3.7190370605004755e-05, |
|
"loss": 0.0326, |
|
"step": 158632 |
|
}, |
|
{ |
|
"epoch": 12.611973392461197, |
|
"grad_norm": 0.7047850489616394, |
|
"learning_rate": 3.6940133037694016e-05, |
|
"loss": 0.0313, |
|
"step": 159264 |
|
}, |
|
{ |
|
"epoch": 12.662020905923345, |
|
"grad_norm": 0.05169570446014404, |
|
"learning_rate": 3.6689895470383276e-05, |
|
"loss": 0.0311, |
|
"step": 159896 |
|
}, |
|
{ |
|
"epoch": 12.712068419385492, |
|
"grad_norm": 0.3161393105983734, |
|
"learning_rate": 3.643965790307254e-05, |
|
"loss": 0.0295, |
|
"step": 160528 |
|
}, |
|
{ |
|
"epoch": 12.76211593284764, |
|
"grad_norm": 0.4407559633255005, |
|
"learning_rate": 3.6189420335761805e-05, |
|
"loss": 0.0307, |
|
"step": 161160 |
|
}, |
|
{ |
|
"epoch": 12.812163446309787, |
|
"grad_norm": 2.090928792953491, |
|
"learning_rate": 3.5939182768451065e-05, |
|
"loss": 0.0288, |
|
"step": 161792 |
|
}, |
|
{ |
|
"epoch": 12.862210959771936, |
|
"grad_norm": 0.039744604378938675, |
|
"learning_rate": 3.568894520114032e-05, |
|
"loss": 0.0285, |
|
"step": 162424 |
|
}, |
|
{ |
|
"epoch": 12.912258473234083, |
|
"grad_norm": 0.48494115471839905, |
|
"learning_rate": 3.543870763382959e-05, |
|
"loss": 0.0302, |
|
"step": 163056 |
|
}, |
|
{ |
|
"epoch": 12.962305986696231, |
|
"grad_norm": 0.38117051124572754, |
|
"learning_rate": 3.518847006651885e-05, |
|
"loss": 0.0302, |
|
"step": 163688 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_gen_len": 12.34829582, |
|
"eval_loss": 0.032208241522312164, |
|
"eval_runtime": 516.3051, |
|
"eval_samples_per_second": 103.082, |
|
"eval_steps_per_second": 1.611, |
|
"eval_wordacc": 0.94701439, |
|
"eval_wordacc_oov": 0.90429282, |
|
"step": 164164 |
|
}, |
|
{ |
|
"epoch": 13.012353500158378, |
|
"grad_norm": 0.5338648557662964, |
|
"learning_rate": 3.4938232499208115e-05, |
|
"loss": 0.0273, |
|
"step": 164320 |
|
}, |
|
{ |
|
"epoch": 13.062401013620526, |
|
"grad_norm": 0.5787109136581421, |
|
"learning_rate": 3.468799493189737e-05, |
|
"loss": 0.0278, |
|
"step": 164952 |
|
}, |
|
{ |
|
"epoch": 13.112448527082673, |
|
"grad_norm": 3.402815818786621, |
|
"learning_rate": 3.4437757364586636e-05, |
|
"loss": 0.0309, |
|
"step": 165584 |
|
}, |
|
{ |
|
"epoch": 13.162496040544822, |
|
"grad_norm": 0.3387065827846527, |
|
"learning_rate": 3.41875197972759e-05, |
|
"loss": 0.0266, |
|
"step": 166216 |
|
}, |
|
{ |
|
"epoch": 13.212543554006968, |
|
"grad_norm": 0.6726759076118469, |
|
"learning_rate": 3.393728222996516e-05, |
|
"loss": 0.0257, |
|
"step": 166848 |
|
}, |
|
{ |
|
"epoch": 13.262591067469117, |
|
"grad_norm": 0.039511628448963165, |
|
"learning_rate": 3.368704466265442e-05, |
|
"loss": 0.0285, |
|
"step": 167480 |
|
}, |
|
{ |
|
"epoch": 13.312638580931264, |
|
"grad_norm": 0.16470862925052643, |
|
"learning_rate": 3.343680709534368e-05, |
|
"loss": 0.0265, |
|
"step": 168112 |
|
}, |
|
{ |
|
"epoch": 13.362686094393412, |
|
"grad_norm": 1.1072640419006348, |
|
"learning_rate": 3.318656952803295e-05, |
|
"loss": 0.0289, |
|
"step": 168744 |
|
}, |
|
{ |
|
"epoch": 13.412733607855559, |
|
"grad_norm": 0.4822203516960144, |
|
"learning_rate": 3.293633196072221e-05, |
|
"loss": 0.0282, |
|
"step": 169376 |
|
}, |
|
{ |
|
"epoch": 13.462781121317708, |
|
"grad_norm": 0.29173898696899414, |
|
"learning_rate": 3.268609439341147e-05, |
|
"loss": 0.025, |
|
"step": 170008 |
|
}, |
|
{ |
|
"epoch": 13.512828634779854, |
|
"grad_norm": 0.12304585427045822, |
|
"learning_rate": 3.243585682610073e-05, |
|
"loss": 0.0286, |
|
"step": 170640 |
|
}, |
|
{ |
|
"epoch": 13.562876148242001, |
|
"grad_norm": 0.02250618301331997, |
|
"learning_rate": 3.218561925878999e-05, |
|
"loss": 0.0261, |
|
"step": 171272 |
|
}, |
|
{ |
|
"epoch": 13.61292366170415, |
|
"grad_norm": 1.214078664779663, |
|
"learning_rate": 3.193538169147926e-05, |
|
"loss": 0.0265, |
|
"step": 171904 |
|
}, |
|
{ |
|
"epoch": 13.662971175166296, |
|
"grad_norm": 1.080644965171814, |
|
"learning_rate": 3.168514412416851e-05, |
|
"loss": 0.0264, |
|
"step": 172536 |
|
}, |
|
{ |
|
"epoch": 13.713018688628445, |
|
"grad_norm": 0.47105857729911804, |
|
"learning_rate": 3.143490655685778e-05, |
|
"loss": 0.03, |
|
"step": 173168 |
|
}, |
|
{ |
|
"epoch": 13.763066202090592, |
|
"grad_norm": 0.8579492568969727, |
|
"learning_rate": 3.118466898954704e-05, |
|
"loss": 0.03, |
|
"step": 173800 |
|
}, |
|
{ |
|
"epoch": 13.81311371555274, |
|
"grad_norm": 1.1287840604782104, |
|
"learning_rate": 3.0934431422236307e-05, |
|
"loss": 0.0274, |
|
"step": 174432 |
|
}, |
|
{ |
|
"epoch": 13.863161229014887, |
|
"grad_norm": 1.2223260402679443, |
|
"learning_rate": 3.068419385492556e-05, |
|
"loss": 0.0275, |
|
"step": 175064 |
|
}, |
|
{ |
|
"epoch": 13.913208742477035, |
|
"grad_norm": 0.6263837814331055, |
|
"learning_rate": 3.0433956287614825e-05, |
|
"loss": 0.03, |
|
"step": 175696 |
|
}, |
|
{ |
|
"epoch": 13.963256255939182, |
|
"grad_norm": 2.3187966346740723, |
|
"learning_rate": 3.018371872030409e-05, |
|
"loss": 0.0277, |
|
"step": 176328 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_gen_len": 12.35064447, |
|
"eval_loss": 0.031636711210012436, |
|
"eval_runtime": 514.9724, |
|
"eval_samples_per_second": 103.349, |
|
"eval_steps_per_second": 1.616, |
|
"eval_wordacc": 0.94785991, |
|
"eval_wordacc_oov": 0.9040336, |
|
"step": 176792 |
|
}, |
|
{ |
|
"epoch": 14.01330376940133, |
|
"grad_norm": 0.09013538807630539, |
|
"learning_rate": 2.993348115299335e-05, |
|
"loss": 0.0291, |
|
"step": 176960 |
|
}, |
|
{ |
|
"epoch": 14.063351282863477, |
|
"grad_norm": 0.7201142907142639, |
|
"learning_rate": 2.9683243585682614e-05, |
|
"loss": 0.0216, |
|
"step": 177592 |
|
}, |
|
{ |
|
"epoch": 14.113398796325626, |
|
"grad_norm": 0.2142077535390854, |
|
"learning_rate": 2.943300601837187e-05, |
|
"loss": 0.0264, |
|
"step": 178224 |
|
}, |
|
{ |
|
"epoch": 14.163446309787773, |
|
"grad_norm": 0.618773341178894, |
|
"learning_rate": 2.918276845106114e-05, |
|
"loss": 0.0248, |
|
"step": 178856 |
|
}, |
|
{ |
|
"epoch": 14.213493823249921, |
|
"grad_norm": 1.0433090925216675, |
|
"learning_rate": 2.8932530883750396e-05, |
|
"loss": 0.0286, |
|
"step": 179488 |
|
}, |
|
{ |
|
"epoch": 14.263541336712068, |
|
"grad_norm": 0.11434757709503174, |
|
"learning_rate": 2.8682293316439656e-05, |
|
"loss": 0.0253, |
|
"step": 180120 |
|
}, |
|
{ |
|
"epoch": 14.313588850174217, |
|
"grad_norm": 1.1828645467758179, |
|
"learning_rate": 2.843205574912892e-05, |
|
"loss": 0.0245, |
|
"step": 180752 |
|
}, |
|
{ |
|
"epoch": 14.363636363636363, |
|
"grad_norm": 0.3703102171421051, |
|
"learning_rate": 2.818181818181818e-05, |
|
"loss": 0.0261, |
|
"step": 181384 |
|
}, |
|
{ |
|
"epoch": 14.413683877098512, |
|
"grad_norm": 0.20109039545059204, |
|
"learning_rate": 2.7931580614507445e-05, |
|
"loss": 0.0255, |
|
"step": 182016 |
|
}, |
|
{ |
|
"epoch": 14.463731390560659, |
|
"grad_norm": 0.627756655216217, |
|
"learning_rate": 2.7681343047196706e-05, |
|
"loss": 0.0228, |
|
"step": 182648 |
|
}, |
|
{ |
|
"epoch": 14.513778904022807, |
|
"grad_norm": 0.04185875132679939, |
|
"learning_rate": 2.743110547988597e-05, |
|
"loss": 0.0237, |
|
"step": 183280 |
|
}, |
|
{ |
|
"epoch": 14.563826417484954, |
|
"grad_norm": 0.10042019188404083, |
|
"learning_rate": 2.718086791257523e-05, |
|
"loss": 0.0272, |
|
"step": 183912 |
|
}, |
|
{ |
|
"epoch": 14.613873930947102, |
|
"grad_norm": 1.0942548513412476, |
|
"learning_rate": 2.693063034526449e-05, |
|
"loss": 0.0242, |
|
"step": 184544 |
|
}, |
|
{ |
|
"epoch": 14.66392144440925, |
|
"grad_norm": 0.7228376269340515, |
|
"learning_rate": 2.6680392777953756e-05, |
|
"loss": 0.0252, |
|
"step": 185176 |
|
}, |
|
{ |
|
"epoch": 14.713968957871398, |
|
"grad_norm": 0.6470568776130676, |
|
"learning_rate": 2.6430155210643016e-05, |
|
"loss": 0.0269, |
|
"step": 185808 |
|
}, |
|
{ |
|
"epoch": 14.764016471333544, |
|
"grad_norm": 0.4635055363178253, |
|
"learning_rate": 2.617991764333228e-05, |
|
"loss": 0.0243, |
|
"step": 186440 |
|
}, |
|
{ |
|
"epoch": 14.814063984795691, |
|
"grad_norm": 0.43463265895843506, |
|
"learning_rate": 2.592968007602154e-05, |
|
"loss": 0.0265, |
|
"step": 187072 |
|
}, |
|
{ |
|
"epoch": 14.86411149825784, |
|
"grad_norm": 0.7284203767776489, |
|
"learning_rate": 2.5679442508710805e-05, |
|
"loss": 0.0269, |
|
"step": 187704 |
|
}, |
|
{ |
|
"epoch": 14.914159011719988, |
|
"grad_norm": 1.462714433670044, |
|
"learning_rate": 2.5429204941400066e-05, |
|
"loss": 0.0236, |
|
"step": 188336 |
|
}, |
|
{ |
|
"epoch": 14.964206525182135, |
|
"grad_norm": 0.029578888788819313, |
|
"learning_rate": 2.5178967374089323e-05, |
|
"loss": 0.0277, |
|
"step": 188968 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_gen_len": 12.35139604, |
|
"eval_loss": 0.032311730086803436, |
|
"eval_runtime": 516.3173, |
|
"eval_samples_per_second": 103.08, |
|
"eval_steps_per_second": 1.611, |
|
"eval_wordacc": 0.94876179, |
|
"eval_wordacc_oov": 0.90299668, |
|
"step": 189420 |
|
}, |
|
{ |
|
"epoch": 15.014254038644282, |
|
"grad_norm": 0.3032616376876831, |
|
"learning_rate": 2.4928729806778587e-05, |
|
"loss": 0.0275, |
|
"step": 189600 |
|
}, |
|
{ |
|
"epoch": 15.06430155210643, |
|
"grad_norm": 0.08022745698690414, |
|
"learning_rate": 2.467849223946785e-05, |
|
"loss": 0.0212, |
|
"step": 190232 |
|
}, |
|
{ |
|
"epoch": 15.114349065568577, |
|
"grad_norm": 0.06174493208527565, |
|
"learning_rate": 2.4428254672157112e-05, |
|
"loss": 0.0233, |
|
"step": 190864 |
|
}, |
|
{ |
|
"epoch": 15.164396579030726, |
|
"grad_norm": 0.7364438772201538, |
|
"learning_rate": 2.4178017104846373e-05, |
|
"loss": 0.0214, |
|
"step": 191496 |
|
}, |
|
{ |
|
"epoch": 15.214444092492872, |
|
"grad_norm": 0.022581493481993675, |
|
"learning_rate": 2.3927779537535637e-05, |
|
"loss": 0.0236, |
|
"step": 192128 |
|
}, |
|
{ |
|
"epoch": 15.26449160595502, |
|
"grad_norm": 0.03313857316970825, |
|
"learning_rate": 2.3677541970224898e-05, |
|
"loss": 0.0246, |
|
"step": 192760 |
|
}, |
|
{ |
|
"epoch": 15.314539119417168, |
|
"grad_norm": 0.7156698703765869, |
|
"learning_rate": 2.3427304402914162e-05, |
|
"loss": 0.0263, |
|
"step": 193392 |
|
}, |
|
{ |
|
"epoch": 15.364586632879316, |
|
"grad_norm": 0.626669704914093, |
|
"learning_rate": 2.3177066835603423e-05, |
|
"loss": 0.0219, |
|
"step": 194024 |
|
}, |
|
{ |
|
"epoch": 15.414634146341463, |
|
"grad_norm": 1.3695429563522339, |
|
"learning_rate": 2.2926829268292687e-05, |
|
"loss": 0.0244, |
|
"step": 194656 |
|
}, |
|
{ |
|
"epoch": 15.464681659803611, |
|
"grad_norm": 0.03239896893501282, |
|
"learning_rate": 2.2676591700981947e-05, |
|
"loss": 0.0217, |
|
"step": 195288 |
|
}, |
|
{ |
|
"epoch": 15.514729173265758, |
|
"grad_norm": 0.4434269666671753, |
|
"learning_rate": 2.2426354133671208e-05, |
|
"loss": 0.026, |
|
"step": 195920 |
|
}, |
|
{ |
|
"epoch": 15.564776686727907, |
|
"grad_norm": 0.11541499197483063, |
|
"learning_rate": 2.217611656636047e-05, |
|
"loss": 0.0251, |
|
"step": 196552 |
|
}, |
|
{ |
|
"epoch": 15.614824200190053, |
|
"grad_norm": 0.11521150171756744, |
|
"learning_rate": 2.1925878999049733e-05, |
|
"loss": 0.0234, |
|
"step": 197184 |
|
}, |
|
{ |
|
"epoch": 15.664871713652202, |
|
"grad_norm": 0.10318135470151901, |
|
"learning_rate": 2.1675641431738994e-05, |
|
"loss": 0.0239, |
|
"step": 197816 |
|
}, |
|
{ |
|
"epoch": 15.714919227114349, |
|
"grad_norm": 0.181732639670372, |
|
"learning_rate": 2.1425403864428258e-05, |
|
"loss": 0.0231, |
|
"step": 198448 |
|
}, |
|
{ |
|
"epoch": 15.764966740576497, |
|
"grad_norm": 0.43891534209251404, |
|
"learning_rate": 2.117516629711752e-05, |
|
"loss": 0.0258, |
|
"step": 199080 |
|
}, |
|
{ |
|
"epoch": 15.815014254038644, |
|
"grad_norm": 0.2669106125831604, |
|
"learning_rate": 2.0924928729806782e-05, |
|
"loss": 0.0258, |
|
"step": 199712 |
|
}, |
|
{ |
|
"epoch": 15.865061767500793, |
|
"grad_norm": 0.787148118019104, |
|
"learning_rate": 2.067469116249604e-05, |
|
"loss": 0.023, |
|
"step": 200344 |
|
}, |
|
{ |
|
"epoch": 15.91510928096294, |
|
"grad_norm": 0.09489845484495163, |
|
"learning_rate": 2.0424453595185304e-05, |
|
"loss": 0.0233, |
|
"step": 200976 |
|
}, |
|
{ |
|
"epoch": 15.965156794425088, |
|
"grad_norm": 0.07817026227712631, |
|
"learning_rate": 2.0174216027874565e-05, |
|
"loss": 0.0245, |
|
"step": 201608 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_gen_len": 12.35013716, |
|
"eval_loss": 0.03139541670680046, |
|
"eval_runtime": 516.248, |
|
"eval_samples_per_second": 103.094, |
|
"eval_steps_per_second": 1.612, |
|
"eval_wordacc": 0.95127955, |
|
"eval_wordacc_oov": 0.90724803, |
|
"step": 202048 |
|
}, |
|
{ |
|
"epoch": 16.015204307887235, |
|
"grad_norm": 0.026963738724589348, |
|
"learning_rate": 1.992397846056383e-05, |
|
"loss": 0.0209, |
|
"step": 202240 |
|
}, |
|
{ |
|
"epoch": 16.06525182134938, |
|
"grad_norm": 0.031288400292396545, |
|
"learning_rate": 1.967374089325309e-05, |
|
"loss": 0.0211, |
|
"step": 202872 |
|
}, |
|
{ |
|
"epoch": 16.11529933481153, |
|
"grad_norm": 0.04831352084875107, |
|
"learning_rate": 1.9423503325942354e-05, |
|
"loss": 0.0216, |
|
"step": 203504 |
|
}, |
|
{ |
|
"epoch": 16.16534684827368, |
|
"grad_norm": 0.3481796383857727, |
|
"learning_rate": 1.9173265758631614e-05, |
|
"loss": 0.0215, |
|
"step": 204136 |
|
}, |
|
{ |
|
"epoch": 16.215394361735825, |
|
"grad_norm": 0.21324285864830017, |
|
"learning_rate": 1.8923028191320875e-05, |
|
"loss": 0.0218, |
|
"step": 204768 |
|
}, |
|
{ |
|
"epoch": 16.265441875197972, |
|
"grad_norm": 1.6428892612457275, |
|
"learning_rate": 1.8672790624010136e-05, |
|
"loss": 0.021, |
|
"step": 205400 |
|
}, |
|
{ |
|
"epoch": 16.31548938866012, |
|
"grad_norm": 0.2430795133113861, |
|
"learning_rate": 1.84225530566994e-05, |
|
"loss": 0.02, |
|
"step": 206032 |
|
}, |
|
{ |
|
"epoch": 16.36553690212227, |
|
"grad_norm": 0.5367847681045532, |
|
"learning_rate": 1.817231548938866e-05, |
|
"loss": 0.0209, |
|
"step": 206664 |
|
}, |
|
{ |
|
"epoch": 16.415584415584416, |
|
"grad_norm": 0.08580495417118073, |
|
"learning_rate": 1.7922077922077925e-05, |
|
"loss": 0.0204, |
|
"step": 207296 |
|
}, |
|
{ |
|
"epoch": 16.465631929046562, |
|
"grad_norm": 0.6704065203666687, |
|
"learning_rate": 1.7671840354767185e-05, |
|
"loss": 0.0218, |
|
"step": 207928 |
|
}, |
|
{ |
|
"epoch": 16.51567944250871, |
|
"grad_norm": 1.014145016670227, |
|
"learning_rate": 1.742160278745645e-05, |
|
"loss": 0.0216, |
|
"step": 208560 |
|
}, |
|
{ |
|
"epoch": 16.56572695597086, |
|
"grad_norm": 0.040132321417331696, |
|
"learning_rate": 1.7171365220145707e-05, |
|
"loss": 0.0249, |
|
"step": 209192 |
|
}, |
|
{ |
|
"epoch": 16.615774469433006, |
|
"grad_norm": 1.3288291692733765, |
|
"learning_rate": 1.692112765283497e-05, |
|
"loss": 0.0216, |
|
"step": 209824 |
|
}, |
|
{ |
|
"epoch": 16.665821982895153, |
|
"grad_norm": 0.6535305976867676, |
|
"learning_rate": 1.667089008552423e-05, |
|
"loss": 0.0221, |
|
"step": 210456 |
|
}, |
|
{ |
|
"epoch": 16.7158694963573, |
|
"grad_norm": 0.516942024230957, |
|
"learning_rate": 1.6420652518213496e-05, |
|
"loss": 0.0209, |
|
"step": 211088 |
|
}, |
|
{ |
|
"epoch": 16.76591700981945, |
|
"grad_norm": 0.6040933728218079, |
|
"learning_rate": 1.6170414950902756e-05, |
|
"loss": 0.0231, |
|
"step": 211720 |
|
}, |
|
{ |
|
"epoch": 16.815964523281597, |
|
"grad_norm": 0.07999061793088913, |
|
"learning_rate": 1.592017738359202e-05, |
|
"loss": 0.023, |
|
"step": 212352 |
|
}, |
|
{ |
|
"epoch": 16.866012036743744, |
|
"grad_norm": 2.897904872894287, |
|
"learning_rate": 1.566993981628128e-05, |
|
"loss": 0.0227, |
|
"step": 212984 |
|
}, |
|
{ |
|
"epoch": 16.91605955020589, |
|
"grad_norm": 0.2583966553211212, |
|
"learning_rate": 1.5419702248970542e-05, |
|
"loss": 0.0218, |
|
"step": 213616 |
|
}, |
|
{ |
|
"epoch": 16.96610706366804, |
|
"grad_norm": 1.3506840467453003, |
|
"learning_rate": 1.5169464681659804e-05, |
|
"loss": 0.0235, |
|
"step": 214248 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_gen_len": 12.35109541, |
|
"eval_loss": 0.03128550201654434, |
|
"eval_runtime": 515.496, |
|
"eval_samples_per_second": 103.244, |
|
"eval_steps_per_second": 1.614, |
|
"eval_wordacc": 0.95197475, |
|
"eval_wordacc_oov": 0.90714434, |
|
"step": 214676 |
|
}, |
|
{ |
|
"epoch": 17.016154577130187, |
|
"grad_norm": 0.28379184007644653, |
|
"learning_rate": 1.4919227114349067e-05, |
|
"loss": 0.0196, |
|
"step": 214880 |
|
}, |
|
{ |
|
"epoch": 17.066202090592334, |
|
"grad_norm": 0.2609073221683502, |
|
"learning_rate": 1.4668989547038327e-05, |
|
"loss": 0.0199, |
|
"step": 215512 |
|
}, |
|
{ |
|
"epoch": 17.11624960405448, |
|
"grad_norm": 0.22402538359165192, |
|
"learning_rate": 1.441875197972759e-05, |
|
"loss": 0.0203, |
|
"step": 216144 |
|
}, |
|
{ |
|
"epoch": 17.16629711751663, |
|
"grad_norm": 0.020612630993127823, |
|
"learning_rate": 1.4168514412416852e-05, |
|
"loss": 0.0196, |
|
"step": 216776 |
|
}, |
|
{ |
|
"epoch": 17.216344630978778, |
|
"grad_norm": 0.036708053201436996, |
|
"learning_rate": 1.3918276845106115e-05, |
|
"loss": 0.0164, |
|
"step": 217408 |
|
}, |
|
{ |
|
"epoch": 17.266392144440925, |
|
"grad_norm": 1.0852469205856323, |
|
"learning_rate": 1.3668039277795375e-05, |
|
"loss": 0.0197, |
|
"step": 218040 |
|
}, |
|
{ |
|
"epoch": 17.31643965790307, |
|
"grad_norm": 0.08419329673051834, |
|
"learning_rate": 1.3417801710484638e-05, |
|
"loss": 0.0223, |
|
"step": 218672 |
|
}, |
|
{ |
|
"epoch": 17.366487171365222, |
|
"grad_norm": 0.8586103320121765, |
|
"learning_rate": 1.31675641431739e-05, |
|
"loss": 0.0217, |
|
"step": 219304 |
|
}, |
|
{ |
|
"epoch": 17.41653468482737, |
|
"grad_norm": 0.41822636127471924, |
|
"learning_rate": 1.2917326575863162e-05, |
|
"loss": 0.0203, |
|
"step": 219936 |
|
}, |
|
{ |
|
"epoch": 17.466582198289515, |
|
"grad_norm": 0.16271276772022247, |
|
"learning_rate": 1.2667089008552425e-05, |
|
"loss": 0.0211, |
|
"step": 220568 |
|
}, |
|
{ |
|
"epoch": 17.516629711751662, |
|
"grad_norm": 0.7023501992225647, |
|
"learning_rate": 1.2416851441241686e-05, |
|
"loss": 0.0197, |
|
"step": 221200 |
|
}, |
|
{ |
|
"epoch": 17.56667722521381, |
|
"grad_norm": 0.038701217621564865, |
|
"learning_rate": 1.2166613873930948e-05, |
|
"loss": 0.022, |
|
"step": 221832 |
|
}, |
|
{ |
|
"epoch": 17.61672473867596, |
|
"grad_norm": 0.03481602668762207, |
|
"learning_rate": 1.191637630662021e-05, |
|
"loss": 0.0178, |
|
"step": 222464 |
|
}, |
|
{ |
|
"epoch": 17.666772252138106, |
|
"grad_norm": 0.3818272352218628, |
|
"learning_rate": 1.1666138739309473e-05, |
|
"loss": 0.0204, |
|
"step": 223096 |
|
}, |
|
{ |
|
"epoch": 17.716819765600253, |
|
"grad_norm": 1.1725687980651855, |
|
"learning_rate": 1.1415901171998734e-05, |
|
"loss": 0.0214, |
|
"step": 223728 |
|
}, |
|
{ |
|
"epoch": 17.7668672790624, |
|
"grad_norm": 0.5618239641189575, |
|
"learning_rate": 1.1165663604687996e-05, |
|
"loss": 0.0198, |
|
"step": 224360 |
|
}, |
|
{ |
|
"epoch": 17.81691479252455, |
|
"grad_norm": 0.11917304992675781, |
|
"learning_rate": 1.0915426037377258e-05, |
|
"loss": 0.0203, |
|
"step": 224992 |
|
}, |
|
{ |
|
"epoch": 17.866962305986696, |
|
"grad_norm": 0.1154344379901886, |
|
"learning_rate": 1.066518847006652e-05, |
|
"loss": 0.0217, |
|
"step": 225624 |
|
}, |
|
{ |
|
"epoch": 17.917009819448843, |
|
"grad_norm": 0.12502028048038483, |
|
"learning_rate": 1.0414950902755781e-05, |
|
"loss": 0.0217, |
|
"step": 226256 |
|
}, |
|
{ |
|
"epoch": 17.96705733291099, |
|
"grad_norm": 0.27321675419807434, |
|
"learning_rate": 1.0164713335445044e-05, |
|
"loss": 0.0206, |
|
"step": 226888 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_gen_len": 12.35015595, |
|
"eval_loss": 0.030980365350842476, |
|
"eval_runtime": 547.6731, |
|
"eval_samples_per_second": 97.178, |
|
"eval_steps_per_second": 1.519, |
|
"eval_wordacc": 0.95308331, |
|
"eval_wordacc_oov": 0.90844048, |
|
"step": 227304 |
|
}, |
|
{ |
|
"epoch": 18.01710484637314, |
|
"grad_norm": 0.15305259823799133, |
|
"learning_rate": 9.914475768134306e-06, |
|
"loss": 0.018, |
|
"step": 227520 |
|
}, |
|
{ |
|
"epoch": 18.067152359835287, |
|
"grad_norm": 2.5299997329711914, |
|
"learning_rate": 9.664238200823567e-06, |
|
"loss": 0.0183, |
|
"step": 228152 |
|
}, |
|
{ |
|
"epoch": 18.117199873297434, |
|
"grad_norm": 0.2780356705188751, |
|
"learning_rate": 9.41400063351283e-06, |
|
"loss": 0.0179, |
|
"step": 228784 |
|
}, |
|
{ |
|
"epoch": 18.16724738675958, |
|
"grad_norm": 0.06621097773313522, |
|
"learning_rate": 9.163763066202092e-06, |
|
"loss": 0.0196, |
|
"step": 229416 |
|
}, |
|
{ |
|
"epoch": 18.21729490022173, |
|
"grad_norm": 0.05437783896923065, |
|
"learning_rate": 8.913525498891354e-06, |
|
"loss": 0.0186, |
|
"step": 230048 |
|
}, |
|
{ |
|
"epoch": 18.267342413683878, |
|
"grad_norm": 0.62905353307724, |
|
"learning_rate": 8.663287931580615e-06, |
|
"loss": 0.0202, |
|
"step": 230680 |
|
}, |
|
{ |
|
"epoch": 18.317389927146024, |
|
"grad_norm": 0.03628231957554817, |
|
"learning_rate": 8.413050364269877e-06, |
|
"loss": 0.0186, |
|
"step": 231312 |
|
}, |
|
{ |
|
"epoch": 18.36743744060817, |
|
"grad_norm": 0.1532977819442749, |
|
"learning_rate": 8.16281279695914e-06, |
|
"loss": 0.0197, |
|
"step": 231944 |
|
}, |
|
{ |
|
"epoch": 18.41748495407032, |
|
"grad_norm": 0.15254627168178558, |
|
"learning_rate": 7.9125752296484e-06, |
|
"loss": 0.0177, |
|
"step": 232576 |
|
}, |
|
{ |
|
"epoch": 18.467532467532468, |
|
"grad_norm": 1.0562399625778198, |
|
"learning_rate": 7.662337662337663e-06, |
|
"loss": 0.0172, |
|
"step": 233208 |
|
}, |
|
{ |
|
"epoch": 18.517579980994615, |
|
"grad_norm": 0.4453679323196411, |
|
"learning_rate": 7.412100095026925e-06, |
|
"loss": 0.0197, |
|
"step": 233840 |
|
}, |
|
{ |
|
"epoch": 18.56762749445676, |
|
"grad_norm": 0.26340222358703613, |
|
"learning_rate": 7.161862527716187e-06, |
|
"loss": 0.018, |
|
"step": 234472 |
|
}, |
|
{ |
|
"epoch": 18.617675007918912, |
|
"grad_norm": 0.052670177072286606, |
|
"learning_rate": 6.911624960405448e-06, |
|
"loss": 0.0184, |
|
"step": 235104 |
|
}, |
|
{ |
|
"epoch": 18.66772252138106, |
|
"grad_norm": 0.026885686442255974, |
|
"learning_rate": 6.661387393094711e-06, |
|
"loss": 0.0183, |
|
"step": 235736 |
|
}, |
|
{ |
|
"epoch": 18.717770034843205, |
|
"grad_norm": 0.34494689106941223, |
|
"learning_rate": 6.411149825783973e-06, |
|
"loss": 0.0192, |
|
"step": 236368 |
|
}, |
|
{ |
|
"epoch": 18.767817548305352, |
|
"grad_norm": 0.0610111765563488, |
|
"learning_rate": 6.160912258473235e-06, |
|
"loss": 0.0179, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 18.817865061767503, |
|
"grad_norm": 0.02205130271613598, |
|
"learning_rate": 5.910674691162496e-06, |
|
"loss": 0.0218, |
|
"step": 237632 |
|
}, |
|
{ |
|
"epoch": 18.86791257522965, |
|
"grad_norm": 0.046364884823560715, |
|
"learning_rate": 5.660437123851759e-06, |
|
"loss": 0.0167, |
|
"step": 238264 |
|
}, |
|
{ |
|
"epoch": 18.917960088691796, |
|
"grad_norm": 0.8253034949302673, |
|
"learning_rate": 5.41019955654102e-06, |
|
"loss": 0.0206, |
|
"step": 238896 |
|
}, |
|
{ |
|
"epoch": 18.968007602153943, |
|
"grad_norm": 0.07394740730524063, |
|
"learning_rate": 5.159961989230283e-06, |
|
"loss": 0.0178, |
|
"step": 239528 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_gen_len": 12.35068205, |
|
"eval_loss": 0.03073255531489849, |
|
"eval_runtime": 534.9855, |
|
"eval_samples_per_second": 99.483, |
|
"eval_steps_per_second": 1.555, |
|
"eval_wordacc": 0.95447371, |
|
"eval_wordacc_oov": 0.9093737, |
|
"step": 239932 |
|
}, |
|
{ |
|
"epoch": 19.01805511561609, |
|
"grad_norm": 0.05381210148334503, |
|
"learning_rate": 4.909724421919544e-06, |
|
"loss": 0.017, |
|
"step": 240160 |
|
}, |
|
{ |
|
"epoch": 19.06810262907824, |
|
"grad_norm": 0.05161185935139656, |
|
"learning_rate": 4.659486854608807e-06, |
|
"loss": 0.0184, |
|
"step": 240792 |
|
}, |
|
{ |
|
"epoch": 19.118150142540387, |
|
"grad_norm": 0.054974284023046494, |
|
"learning_rate": 4.409249287298068e-06, |
|
"loss": 0.0165, |
|
"step": 241424 |
|
}, |
|
{ |
|
"epoch": 19.168197656002533, |
|
"grad_norm": 0.4919557571411133, |
|
"learning_rate": 4.15901171998733e-06, |
|
"loss": 0.0171, |
|
"step": 242056 |
|
}, |
|
{ |
|
"epoch": 19.21824516946468, |
|
"grad_norm": 0.8104053735733032, |
|
"learning_rate": 3.908774152676592e-06, |
|
"loss": 0.0161, |
|
"step": 242688 |
|
}, |
|
{ |
|
"epoch": 19.26829268292683, |
|
"grad_norm": 0.04452453926205635, |
|
"learning_rate": 3.6585365853658537e-06, |
|
"loss": 0.017, |
|
"step": 243320 |
|
}, |
|
{ |
|
"epoch": 19.318340196388977, |
|
"grad_norm": 2.039726972579956, |
|
"learning_rate": 3.408299018055116e-06, |
|
"loss": 0.0188, |
|
"step": 243952 |
|
}, |
|
{ |
|
"epoch": 19.368387709851124, |
|
"grad_norm": 0.05839849263429642, |
|
"learning_rate": 3.1580614507443777e-06, |
|
"loss": 0.0174, |
|
"step": 244584 |
|
}, |
|
{ |
|
"epoch": 19.41843522331327, |
|
"grad_norm": 0.06209117919206619, |
|
"learning_rate": 2.9078238834336396e-06, |
|
"loss": 0.0177, |
|
"step": 245216 |
|
}, |
|
{ |
|
"epoch": 19.46848273677542, |
|
"grad_norm": 0.023634545505046844, |
|
"learning_rate": 2.6575863161229016e-06, |
|
"loss": 0.0187, |
|
"step": 245848 |
|
}, |
|
{ |
|
"epoch": 19.518530250237568, |
|
"grad_norm": 0.4597359001636505, |
|
"learning_rate": 2.4073487488121636e-06, |
|
"loss": 0.0157, |
|
"step": 246480 |
|
}, |
|
{ |
|
"epoch": 19.568577763699714, |
|
"grad_norm": 0.03822094574570656, |
|
"learning_rate": 2.1571111815014256e-06, |
|
"loss": 0.0191, |
|
"step": 247112 |
|
}, |
|
{ |
|
"epoch": 19.61862527716186, |
|
"grad_norm": 0.028475474566221237, |
|
"learning_rate": 1.9068736141906876e-06, |
|
"loss": 0.0184, |
|
"step": 247744 |
|
}, |
|
{ |
|
"epoch": 19.66867279062401, |
|
"grad_norm": 2.0368409156799316, |
|
"learning_rate": 1.6566360468799495e-06, |
|
"loss": 0.016, |
|
"step": 248376 |
|
}, |
|
{ |
|
"epoch": 19.71872030408616, |
|
"grad_norm": 0.6395711898803711, |
|
"learning_rate": 1.4063984795692113e-06, |
|
"loss": 0.0182, |
|
"step": 249008 |
|
}, |
|
{ |
|
"epoch": 19.768767817548305, |
|
"grad_norm": 0.11478591710329056, |
|
"learning_rate": 1.1561609122584733e-06, |
|
"loss": 0.0172, |
|
"step": 249640 |
|
}, |
|
{ |
|
"epoch": 19.818815331010452, |
|
"grad_norm": 2.0897409915924072, |
|
"learning_rate": 9.059233449477353e-07, |
|
"loss": 0.0178, |
|
"step": 250272 |
|
}, |
|
{ |
|
"epoch": 19.868862844472602, |
|
"grad_norm": 1.6272854804992676, |
|
"learning_rate": 6.556857776369973e-07, |
|
"loss": 0.0165, |
|
"step": 250904 |
|
}, |
|
{ |
|
"epoch": 19.91891035793475, |
|
"grad_norm": 0.018388045951724052, |
|
"learning_rate": 4.0544821032625913e-07, |
|
"loss": 0.0153, |
|
"step": 251536 |
|
}, |
|
{ |
|
"epoch": 19.968957871396896, |
|
"grad_norm": 0.097270667552948, |
|
"learning_rate": 1.5521064301552106e-07, |
|
"loss": 0.016, |
|
"step": 252168 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_gen_len": 12.35156514, |
|
"eval_loss": 0.030767865478992462, |
|
"eval_runtime": 535.8673, |
|
"eval_samples_per_second": 99.319, |
|
"eval_steps_per_second": 1.553, |
|
"eval_wordacc": 0.95458645, |
|
"eval_wordacc_oov": 0.90963293, |
|
"step": 252560 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"step": 252560, |
|
"total_flos": 1486519029399552.0, |
|
"train_loss": 0.04684994914012881, |
|
"train_runtime": 19459.2196, |
|
"train_samples_per_second": 103.832, |
|
"train_steps_per_second": 12.979 |
|
} |
|
], |
|
"logging_steps": 632, |
|
"max_steps": 252560, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 20, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1486519029399552.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|