|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9980609102315501, |
|
"eval_steps": 500, |
|
"global_step": 7000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.999954848325688e-06, |
|
"loss": 0.7277, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.999819394208827e-06, |
|
"loss": 0.6455, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.999593640367633e-06, |
|
"loss": 0.5428, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.9992775913324e-06, |
|
"loss": 0.3486, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.998871253445414e-06, |
|
"loss": 0.3524, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.998374634860822e-06, |
|
"loss": 0.2701, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.997787745544478e-06, |
|
"loss": 0.2106, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.997110597273727e-06, |
|
"loss": 0.2596, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.996343203637181e-06, |
|
"loss": 0.2015, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.995485580034448e-06, |
|
"loss": 0.2462, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.99453774367581e-06, |
|
"loss": 0.2172, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.993499713581887e-06, |
|
"loss": 0.2655, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.992371510583257e-06, |
|
"loss": 0.1823, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.991153157320033e-06, |
|
"loss": 0.227, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.989844678241412e-06, |
|
"loss": 0.2243, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.988446099605179e-06, |
|
"loss": 0.1787, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.986957449477188e-06, |
|
"loss": 0.2211, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.98537875773079e-06, |
|
"loss": 0.1683, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.983710056046243e-06, |
|
"loss": 0.1879, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.98195137791007e-06, |
|
"loss": 0.1809, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.980102758614384e-06, |
|
"loss": 0.2204, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.978164235256191e-06, |
|
"loss": 0.189, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.976135846736634e-06, |
|
"loss": 0.1389, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.97401763376022e-06, |
|
"loss": 0.1231, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.971809638833998e-06, |
|
"loss": 0.2018, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.96951190626671e-06, |
|
"loss": 0.1271, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.967124482167896e-06, |
|
"loss": 0.1364, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.964647414446981e-06, |
|
"loss": 0.1658, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.9620807528123e-06, |
|
"loss": 0.1479, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.959424548770104e-06, |
|
"loss": 0.1353, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.95667885562353e-06, |
|
"loss": 0.1761, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.953843728471533e-06, |
|
"loss": 0.1391, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.950919224207774e-06, |
|
"loss": 0.1776, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.94790540151948e-06, |
|
"loss": 0.2089, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.94480232088627e-06, |
|
"loss": 0.1368, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.941610044578937e-06, |
|
"loss": 0.1501, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.938328636658202e-06, |
|
"loss": 0.1807, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.934958162973425e-06, |
|
"loss": 0.13, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.931498691161287e-06, |
|
"loss": 0.1437, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.92795029064443e-06, |
|
"loss": 0.0828, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.924313032630063e-06, |
|
"loss": 0.1041, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.920586990108539e-06, |
|
"loss": 0.2663, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.916772237851882e-06, |
|
"loss": 0.2055, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.912868852412294e-06, |
|
"loss": 0.151, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.908876912120614e-06, |
|
"loss": 0.1638, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.904796497084747e-06, |
|
"loss": 0.1399, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.900627689188059e-06, |
|
"loss": 0.1615, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.89637057208773e-06, |
|
"loss": 0.1209, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.892025231213077e-06, |
|
"loss": 0.1605, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.887591753763842e-06, |
|
"loss": 0.124, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.13451002538204193, |
|
"eval_runtime": 446.6893, |
|
"eval_samples_per_second": 52.338, |
|
"eval_steps_per_second": 8.724, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.883070228708436e-06, |
|
"loss": 0.1045, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.878460746782163e-06, |
|
"loss": 0.2208, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.873763400485387e-06, |
|
"loss": 0.1483, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.868978284081688e-06, |
|
"loss": 0.1322, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.864105493595962e-06, |
|
"loss": 0.1227, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.859145126812497e-06, |
|
"loss": 0.1371, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.854097283273012e-06, |
|
"loss": 0.1909, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.848962064274655e-06, |
|
"loss": 0.0993, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 8.843739572867977e-06, |
|
"loss": 0.1053, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.838429913854859e-06, |
|
"loss": 0.1401, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.83303319378641e-06, |
|
"loss": 0.09, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.827549520960827e-06, |
|
"loss": 0.1667, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.821979005421223e-06, |
|
"loss": 0.1069, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.816321758953426e-06, |
|
"loss": 0.1659, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.810577895083726e-06, |
|
"loss": 0.1309, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.804747529076592e-06, |
|
"loss": 0.16, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.79883077793238e-06, |
|
"loss": 0.1041, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.792827760384965e-06, |
|
"loss": 0.1344, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.786738596899362e-06, |
|
"loss": 0.0922, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.780563409669322e-06, |
|
"loss": 0.0813, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.774302322614858e-06, |
|
"loss": 0.1795, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.767955461379783e-06, |
|
"loss": 0.1409, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 8.761522953329163e-06, |
|
"loss": 0.1362, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.75500492754678e-06, |
|
"loss": 0.1227, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.74840151483254e-06, |
|
"loss": 0.0643, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.741712847699832e-06, |
|
"loss": 0.1895, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.734939060372892e-06, |
|
"loss": 0.0959, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.72808028878409e-06, |
|
"loss": 0.1089, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.721136670571216e-06, |
|
"loss": 0.0959, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.714108345074704e-06, |
|
"loss": 0.1264, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.706995453334855e-06, |
|
"loss": 0.1078, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.699798138088988e-06, |
|
"loss": 0.1278, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.692516543768581e-06, |
|
"loss": 0.1337, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.685150816496384e-06, |
|
"loss": 0.0815, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.677701104083467e-06, |
|
"loss": 0.1033, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.670167556026274e-06, |
|
"loss": 0.1412, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.662550323503606e-06, |
|
"loss": 0.197, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.654849559373595e-06, |
|
"loss": 0.1118, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.647065418170644e-06, |
|
"loss": 0.1132, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.639198056102309e-06, |
|
"loss": 0.0928, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.631247631046176e-06, |
|
"loss": 0.0723, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.623214302546691e-06, |
|
"loss": 0.126, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.615098231811958e-06, |
|
"loss": 0.1363, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.606899581710505e-06, |
|
"loss": 0.1243, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.598618516768005e-06, |
|
"loss": 0.0382, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.590255203163995e-06, |
|
"loss": 0.1295, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.581809808728524e-06, |
|
"loss": 0.1145, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.573282502938791e-06, |
|
"loss": 0.0915, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.564673456915746e-06, |
|
"loss": 0.1918, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.55598284342065e-06, |
|
"loss": 0.1732, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.12965914607048035, |
|
"eval_runtime": 448.5256, |
|
"eval_samples_per_second": 52.124, |
|
"eval_steps_per_second": 8.688, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.547210836851614e-06, |
|
"loss": 0.0833, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.538357613240099e-06, |
|
"loss": 0.1597, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.52942335024738e-06, |
|
"loss": 0.1041, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.520408227160983e-06, |
|
"loss": 0.1819, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.511312424891083e-06, |
|
"loss": 0.1112, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.502136125966886e-06, |
|
"loss": 0.1372, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.49287951453295e-06, |
|
"loss": 0.1596, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.483542776345496e-06, |
|
"loss": 0.1558, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.47412609876869e-06, |
|
"loss": 0.069, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.464629670770866e-06, |
|
"loss": 0.1468, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.455053682920747e-06, |
|
"loss": 0.106, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.445398327383616e-06, |
|
"loss": 0.1562, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.435663797917459e-06, |
|
"loss": 0.1404, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.425850289869074e-06, |
|
"loss": 0.1198, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.415958000170163e-06, |
|
"loss": 0.107, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.405987127333367e-06, |
|
"loss": 0.0911, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.395937871448282e-06, |
|
"loss": 0.066, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.385810434177462e-06, |
|
"loss": 0.0988, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.375605018752344e-06, |
|
"loss": 0.1648, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.365321829969196e-06, |
|
"loss": 0.1124, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.35496107418499e-06, |
|
"loss": 0.1251, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.34452295931327e-06, |
|
"loss": 0.1167, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.334007694819972e-06, |
|
"loss": 0.1238, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.323415491719233e-06, |
|
"loss": 0.1127, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.312746562569141e-06, |
|
"loss": 0.0823, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.302001121467481e-06, |
|
"loss": 0.145, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.291179384047432e-06, |
|
"loss": 0.1257, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.280281567473247e-06, |
|
"loss": 0.1014, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.269307890435886e-06, |
|
"loss": 0.1289, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.258258573148637e-06, |
|
"loss": 0.104, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.247133837342685e-06, |
|
"loss": 0.1231, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.235933906262679e-06, |
|
"loss": 0.1594, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.22465900466223e-06, |
|
"loss": 0.0999, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.213309358799427e-06, |
|
"loss": 0.061, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.201885196432273e-06, |
|
"loss": 0.0618, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.190386746814125e-06, |
|
"loss": 0.1607, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.1788142406891e-06, |
|
"loss": 0.0726, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.167167910287428e-06, |
|
"loss": 0.0594, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.155447989320808e-06, |
|
"loss": 0.1341, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.14365471297771e-06, |
|
"loss": 0.1253, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.131788317918651e-06, |
|
"loss": 0.0865, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.119849042271458e-06, |
|
"loss": 0.1227, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.10783712562648e-06, |
|
"loss": 0.1024, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.095752809031783e-06, |
|
"loss": 0.1495, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.083596334988313e-06, |
|
"loss": 0.1098, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.071367947445025e-06, |
|
"loss": 0.1044, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.059067891794002e-06, |
|
"loss": 0.0757, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.046696414865507e-06, |
|
"loss": 0.0991, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.034253764923052e-06, |
|
"loss": 0.141, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.021740191658405e-06, |
|
"loss": 0.1023, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.11160164326429367, |
|
"eval_runtime": 448.3846, |
|
"eval_samples_per_second": 52.141, |
|
"eval_steps_per_second": 8.691, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.009155946186583e-06, |
|
"loss": 0.0782, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.996501281040809e-06, |
|
"loss": 0.1132, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.983776450167446e-06, |
|
"loss": 0.042, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.970981708920902e-06, |
|
"loss": 0.0896, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.958117314058508e-06, |
|
"loss": 0.1038, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.945183523735359e-06, |
|
"loss": 0.052, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.932180597499137e-06, |
|
"loss": 0.1224, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.919108796284908e-06, |
|
"loss": 0.1473, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.905968382409874e-06, |
|
"loss": 0.1499, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.89275961956812e-06, |
|
"loss": 0.0728, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.879482772825316e-06, |
|
"loss": 0.1159, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.8661381086134e-06, |
|
"loss": 0.1409, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.85272589472523e-06, |
|
"loss": 0.1065, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.839246400309213e-06, |
|
"loss": 0.1386, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.8256998958639e-06, |
|
"loss": 0.0995, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.812086653232559e-06, |
|
"loss": 0.0947, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.798406945597726e-06, |
|
"loss": 0.1507, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.784661047475712e-06, |
|
"loss": 0.0693, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.7708492347111e-06, |
|
"loss": 0.1194, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.756971784471213e-06, |
|
"loss": 0.1297, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.743028975240545e-06, |
|
"loss": 0.0933, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.729021086815177e-06, |
|
"loss": 0.1029, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.71494840029716e-06, |
|
"loss": 0.0572, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.700811198088876e-06, |
|
"loss": 0.0754, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.68660976388737e-06, |
|
"loss": 0.0957, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.672344382678653e-06, |
|
"loss": 0.1321, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.658015340731994e-06, |
|
"loss": 0.0711, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.643622925594161e-06, |
|
"loss": 0.1211, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.629167426083665e-06, |
|
"loss": 0.0726, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.614649132284948e-06, |
|
"loss": 0.0481, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.600068335542579e-06, |
|
"loss": 0.1569, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.585425328455395e-06, |
|
"loss": 0.046, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.5707204048706365e-06, |
|
"loss": 0.0909, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.5559538598780434e-06, |
|
"loss": 0.1273, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.541125989803943e-06, |
|
"loss": 0.0636, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.526237092205292e-06, |
|
"loss": 0.1124, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.511287465863719e-06, |
|
"loss": 0.1033, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.496277410779513e-06, |
|
"loss": 0.1055, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.481207228165619e-06, |
|
"loss": 0.0778, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.466077220441581e-06, |
|
"loss": 0.0965, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.45088769122748e-06, |
|
"loss": 0.0561, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.435638945337836e-06, |
|
"loss": 0.1318, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.420331288775502e-06, |
|
"loss": 0.0561, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.404965028725507e-06, |
|
"loss": 0.0997, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.3895404735489035e-06, |
|
"loss": 0.135, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.374057932776579e-06, |
|
"loss": 0.0591, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.358517717103039e-06, |
|
"loss": 0.0574, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.342920138380175e-06, |
|
"loss": 0.0589, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.327265509611007e-06, |
|
"loss": 0.0685, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.311554144943401e-06, |
|
"loss": 0.0386, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.14258363842964172, |
|
"eval_runtime": 447.0432, |
|
"eval_samples_per_second": 52.297, |
|
"eval_steps_per_second": 8.717, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.295786359663766e-06, |
|
"loss": 0.0679, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.2799624701907275e-06, |
|
"loss": 0.1962, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.264082794068776e-06, |
|
"loss": 0.0871, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.248147649961896e-06, |
|
"loss": 0.1062, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.232157357647171e-06, |
|
"loss": 0.0415, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.216112238008363e-06, |
|
"loss": 0.1604, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.200012613029481e-06, |
|
"loss": 0.0401, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.1838588057883136e-06, |
|
"loss": 0.0683, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.167651140449947e-06, |
|
"loss": 0.0891, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.1513899422602585e-06, |
|
"loss": 0.1354, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.135075537539394e-06, |
|
"loss": 0.0904, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.118708253675213e-06, |
|
"loss": 0.1138, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.1022884191167256e-06, |
|
"loss": 0.0917, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.085816363367499e-06, |
|
"loss": 0.0465, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.0692924169790395e-06, |
|
"loss": 0.0826, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.05271691154417e-06, |
|
"loss": 0.0623, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.036090179690365e-06, |
|
"loss": 0.1098, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.019412555073087e-06, |
|
"loss": 0.0722, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.0026843723690755e-06, |
|
"loss": 0.1178, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.9859059672696475e-06, |
|
"loss": 0.0345, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.969077676473951e-06, |
|
"loss": 0.0445, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.95219983768221e-06, |
|
"loss": 0.1169, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.9352727895889486e-06, |
|
"loss": 0.0563, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.918296871876196e-06, |
|
"loss": 0.0492, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.901272425206663e-06, |
|
"loss": 0.1094, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.884199791216922e-06, |
|
"loss": 0.0515, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.867079312510525e-06, |
|
"loss": 0.1382, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.849911332651154e-06, |
|
"loss": 0.1041, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.832696196155716e-06, |
|
"loss": 0.0913, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.815434248487425e-06, |
|
"loss": 0.0794, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.798125836048876e-06, |
|
"loss": 0.0864, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.780771306175093e-06, |
|
"loss": 0.1276, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.763371007126555e-06, |
|
"loss": 0.1068, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.745925288082212e-06, |
|
"loss": 0.1093, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.728434499132473e-06, |
|
"loss": 0.099, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.710898991272189e-06, |
|
"loss": 0.0841, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.693319116393596e-06, |
|
"loss": 0.0863, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.6756952272792685e-06, |
|
"loss": 0.0726, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.658027677595027e-06, |
|
"loss": 0.1138, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.6403168218828494e-06, |
|
"loss": 0.102, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.622563015553752e-06, |
|
"loss": 0.085, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.604766614880662e-06, |
|
"loss": 0.0646, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.586927976991262e-06, |
|
"loss": 0.0487, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.569047459860825e-06, |
|
"loss": 0.0929, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.551125422305035e-06, |
|
"loss": 0.097, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.533162223972784e-06, |
|
"loss": 0.0976, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.515158225338948e-06, |
|
"loss": 0.1232, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.497113787697166e-06, |
|
"loss": 0.0646, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.479029273152582e-06, |
|
"loss": 0.0285, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.460905044614577e-06, |
|
"loss": 0.0716, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.11192344129085541, |
|
"eval_runtime": 452.1105, |
|
"eval_samples_per_second": 51.711, |
|
"eval_steps_per_second": 8.62, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.442741465789492e-06, |
|
"loss": 0.0429, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.424538901173323e-06, |
|
"loss": 0.0803, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.406297716044413e-06, |
|
"loss": 0.0858, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.388018276456118e-06, |
|
"loss": 0.0605, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.3697009492294605e-06, |
|
"loss": 0.0681, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.351346101945772e-06, |
|
"loss": 0.0607, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.332954102939312e-06, |
|
"loss": 0.1236, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.314525321289882e-06, |
|
"loss": 0.0361, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.296060126815411e-06, |
|
"loss": 0.0547, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.277558890064545e-06, |
|
"loss": 0.1043, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.259021982309201e-06, |
|
"loss": 0.084, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.240449775537121e-06, |
|
"loss": 0.058, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.221842642444411e-06, |
|
"loss": 0.0862, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.20320095642805e-06, |
|
"loss": 0.1117, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.184525091578414e-06, |
|
"loss": 0.1552, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.165815422671752e-06, |
|
"loss": 0.0642, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.147072325162679e-06, |
|
"loss": 0.048, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.128296175176628e-06, |
|
"loss": 0.0966, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.109487349502319e-06, |
|
"loss": 0.0853, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.090646225584179e-06, |
|
"loss": 0.0884, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.071773181514785e-06, |
|
"loss": 0.0887, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.052868596027264e-06, |
|
"loss": 0.0736, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.033932848487701e-06, |
|
"loss": 0.1141, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.014966318887518e-06, |
|
"loss": 0.0481, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 5.995969387835857e-06, |
|
"loss": 0.0891, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 5.976942436551939e-06, |
|
"loss": 0.0881, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 5.95788584685741e-06, |
|
"loss": 0.0967, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.938800001168685e-06, |
|
"loss": 0.0753, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.919685282489269e-06, |
|
"loss": 0.0671, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.900542074402076e-06, |
|
"loss": 0.1066, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.881370761061724e-06, |
|
"loss": 0.0888, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.862171727186837e-06, |
|
"loss": 0.0783, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.842945358052311e-06, |
|
"loss": 0.0945, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 5.823692039481596e-06, |
|
"loss": 0.103, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.804412157838945e-06, |
|
"loss": 0.0861, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.785106100021663e-06, |
|
"loss": 0.0627, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.765774253452345e-06, |
|
"loss": 0.0828, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.746417006071094e-06, |
|
"loss": 0.0629, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.727034746327746e-06, |
|
"loss": 0.1149, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.707627863174073e-06, |
|
"loss": 0.0865, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.688196746055968e-06, |
|
"loss": 0.0598, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.668741784905642e-06, |
|
"loss": 0.0564, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.649263370133794e-06, |
|
"loss": 0.0993, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.629761892621775e-06, |
|
"loss": 0.0692, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.610237743713742e-06, |
|
"loss": 0.0787, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.590691315208815e-06, |
|
"loss": 0.045, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.571122999353203e-06, |
|
"loss": 0.0221, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.551533188832339e-06, |
|
"loss": 0.065, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.531922276763001e-06, |
|
"loss": 0.0351, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.512290656685415e-06, |
|
"loss": 0.0655, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.12717388570308685, |
|
"eval_runtime": 447.1731, |
|
"eval_samples_per_second": 52.282, |
|
"eval_steps_per_second": 8.715, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.4926387225553675e-06, |
|
"loss": 0.0666, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.472966868736292e-06, |
|
"loss": 0.0775, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.453275489991362e-06, |
|
"loss": 0.0906, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.433564981475563e-06, |
|
"loss": 0.0459, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.413835738727768e-06, |
|
"loss": 0.1028, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.394088157662793e-06, |
|
"loss": 0.1357, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.374322634563464e-06, |
|
"loss": 0.0751, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.354539566072651e-06, |
|
"loss": 0.1197, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.334739349185315e-06, |
|
"loss": 0.1061, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.314922381240544e-06, |
|
"loss": 0.0739, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.295089059913573e-06, |
|
"loss": 0.0794, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.275239783207811e-06, |
|
"loss": 0.0515, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.255374949446845e-06, |
|
"loss": 0.0413, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.235494957266457e-06, |
|
"loss": 0.0763, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.215600205606617e-06, |
|
"loss": 0.0887, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.1956910937034785e-06, |
|
"loss": 0.0472, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.175768021081368e-06, |
|
"loss": 0.0852, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.15583138754477e-06, |
|
"loss": 0.0981, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.135881593170299e-06, |
|
"loss": 0.0239, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.1159190382986725e-06, |
|
"loss": 0.0749, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.095944123526683e-06, |
|
"loss": 0.0929, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.075957249699149e-06, |
|
"loss": 0.0315, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.05595881790088e-06, |
|
"loss": 0.0884, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.03594922944862e-06, |
|
"loss": 0.0381, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.015928885883002e-06, |
|
"loss": 0.0473, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.995898188960485e-06, |
|
"loss": 0.0673, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.975857540645292e-06, |
|
"loss": 0.061, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.955807343101347e-06, |
|
"loss": 0.1168, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.935747998684199e-06, |
|
"loss": 0.0772, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.9156799099329535e-06, |
|
"loss": 0.0439, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.8956034795621925e-06, |
|
"loss": 0.0715, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.8755191104538905e-06, |
|
"loss": 0.0766, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.855427205649333e-06, |
|
"loss": 0.0304, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.835328168341029e-06, |
|
"loss": 0.0465, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.815222401864617e-06, |
|
"loss": 0.1278, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.795110309690772e-06, |
|
"loss": 0.0688, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.7749922954171075e-06, |
|
"loss": 0.0719, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.754868762760083e-06, |
|
"loss": 0.0588, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.734740115546896e-06, |
|
"loss": 0.0889, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.714606757707377e-06, |
|
"loss": 0.0488, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.69446909326589e-06, |
|
"loss": 0.0752, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.674327526333218e-06, |
|
"loss": 0.0838, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.654182461098462e-06, |
|
"loss": 0.0794, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.63403430182092e-06, |
|
"loss": 0.0691, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.61388345282198e-06, |
|
"loss": 0.0814, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.59373031847701e-06, |
|
"loss": 0.04, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.573575303207232e-06, |
|
"loss": 0.1552, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.5534188114716195e-06, |
|
"loss": 0.0742, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.533261247758775e-06, |
|
"loss": 0.0525, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.5131030165788055e-06, |
|
"loss": 0.0725, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.09580323845148087, |
|
"eval_runtime": 449.6777, |
|
"eval_samples_per_second": 51.991, |
|
"eval_steps_per_second": 8.666, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.49294452245522e-06, |
|
"loss": 0.0529, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.4727861699168e-06, |
|
"loss": 0.053, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.45262836348949e-06, |
|
"loss": 0.0941, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.4324715076882714e-06, |
|
"loss": 0.0586, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.412316007009046e-06, |
|
"loss": 0.0536, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.392162265920531e-06, |
|
"loss": 0.0935, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.372010688856124e-06, |
|
"loss": 0.0655, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.351861680205804e-06, |
|
"loss": 0.085, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.331715644308002e-06, |
|
"loss": 0.0822, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.3115729854415e-06, |
|
"loss": 0.0196, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.291434107817306e-06, |
|
"loss": 0.0476, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.271299415570548e-06, |
|
"loss": 0.0322, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.251169312752372e-06, |
|
"loss": 0.1195, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.231044203321815e-06, |
|
"loss": 0.0776, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.2109244911377206e-06, |
|
"loss": 0.0563, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.190810579950615e-06, |
|
"loss": 0.1301, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.17070287339462e-06, |
|
"loss": 0.0751, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.150601774979341e-06, |
|
"loss": 0.0594, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.130507688081777e-06, |
|
"loss": 0.0534, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.110421015938227e-06, |
|
"loss": 0.0672, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.090342161636191e-06, |
|
"loss": 0.0722, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.070271528106287e-06, |
|
"loss": 0.0633, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.050209518114164e-06, |
|
"loss": 0.1139, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.030156534252416e-06, |
|
"loss": 0.0358, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.010112978932511e-06, |
|
"loss": 0.0972, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.990079254376704e-06, |
|
"loss": 0.0397, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.97005576260998e-06, |
|
"loss": 0.0627, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.9500429054519696e-06, |
|
"loss": 0.0617, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.930041084508903e-06, |
|
"loss": 0.0681, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.910050701165534e-06, |
|
"loss": 0.0434, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.890072156577101e-06, |
|
"loss": 0.0437, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.870105851661263e-06, |
|
"loss": 0.0673, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.850152187090063e-06, |
|
"loss": 0.0627, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.8302115632818865e-06, |
|
"loss": 0.0969, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.810284380393421e-06, |
|
"loss": 0.0379, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.790371038311635e-06, |
|
"loss": 0.086, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.7704719366457428e-06, |
|
"loss": 0.0953, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.75058747471919e-06, |
|
"loss": 0.027, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.730718051561647e-06, |
|
"loss": 0.0792, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.710864065900986e-06, |
|
"loss": 0.0777, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.6910259161552983e-06, |
|
"loss": 0.1068, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.67120400042488e-06, |
|
"loss": 0.0759, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.6513987164842605e-06, |
|
"loss": 0.0391, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.631610461774208e-06, |
|
"loss": 0.0546, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.6118396333937564e-06, |
|
"loss": 0.0817, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.5920866280922445e-06, |
|
"loss": 0.1293, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.572351842261341e-06, |
|
"loss": 0.0762, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.5526356719271034e-06, |
|
"loss": 0.0385, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.532938512742018e-06, |
|
"loss": 0.0444, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.5132607599770728e-06, |
|
"loss": 0.0759, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.09747572988271713, |
|
"eval_runtime": 449.16, |
|
"eval_samples_per_second": 52.05, |
|
"eval_steps_per_second": 8.676, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.493602808513816e-06, |
|
"loss": 0.0295, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.4739650528364323e-06, |
|
"loss": 0.0367, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.4543478870238354e-06, |
|
"loss": 0.047, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.4347517047417474e-06, |
|
"loss": 0.0243, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.4151768992348124e-06, |
|
"loss": 0.1458, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.3956238633186902e-06, |
|
"loss": 0.11, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.376092989372188e-06, |
|
"loss": 0.0348, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.3565846693293755e-06, |
|
"loss": 0.0407, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.337099294671723e-06, |
|
"loss": 0.0623, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.3176372564202514e-06, |
|
"loss": 0.0472, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.298198945127675e-06, |
|
"loss": 0.0584, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.2787847508705756e-06, |
|
"loss": 0.0522, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.2593950632415603e-06, |
|
"loss": 0.0596, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.24003027134146e-06, |
|
"loss": 0.0293, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.2206907637715064e-06, |
|
"loss": 0.0649, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.201376928625543e-06, |
|
"loss": 0.0601, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.182089153482235e-06, |
|
"loss": 0.0341, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.162827825397289e-06, |
|
"loss": 0.0459, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.1435933308956882e-06, |
|
"loss": 0.0224, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.124386055963936e-06, |
|
"loss": 0.0288, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.105206386042308e-06, |
|
"loss": 0.0997, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.0860547060171217e-06, |
|
"loss": 0.0825, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.0669314002130036e-06, |
|
"loss": 0.0409, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.047836852385192e-06, |
|
"loss": 0.0429, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.028771445711819e-06, |
|
"loss": 0.0614, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.0097355627862352e-06, |
|
"loss": 0.0406, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.99072958560932e-06, |
|
"loss": 0.0391, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.971753895581829e-06, |
|
"loss": 0.0983, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.952808873496727e-06, |
|
"loss": 0.0237, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.9338948995315536e-06, |
|
"loss": 0.0713, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.9150123532407973e-06, |
|
"loss": 0.0285, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.8961616135482686e-06, |
|
"loss": 0.0724, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.8773430587395077e-06, |
|
"loss": 0.0603, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.858557066454183e-06, |
|
"loss": 0.09, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.83980401367852e-06, |
|
"loss": 0.0203, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.8210842767377328e-06, |
|
"loss": 0.0579, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.8023982312884707e-06, |
|
"loss": 0.0249, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.783746252311286e-06, |
|
"loss": 0.0508, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.7651287141031027e-06, |
|
"loss": 0.0213, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.746545990269708e-06, |
|
"loss": 0.086, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.727998453718255e-06, |
|
"loss": 0.0348, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.709486476649779e-06, |
|
"loss": 0.1003, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.6910104305517307e-06, |
|
"loss": 0.0035, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.672570686190513e-06, |
|
"loss": 0.0776, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.6541676136040553e-06, |
|
"loss": 0.0812, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.635801582094372e-06, |
|
"loss": 0.0662, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.6174729602201664e-06, |
|
"loss": 0.0873, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.5991821157894182e-06, |
|
"loss": 0.0223, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.5809294158520197e-06, |
|
"loss": 0.0421, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.5627152266923973e-06, |
|
"loss": 0.061, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.127749502658844, |
|
"eval_runtime": 447.4087, |
|
"eval_samples_per_second": 52.254, |
|
"eval_steps_per_second": 8.71, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.5445399138221693e-06, |
|
"loss": 0.0794, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.5264038419728033e-06, |
|
"loss": 0.0787, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.5083073750883057e-06, |
|
"loss": 0.0568, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.4902508763179086e-06, |
|
"loss": 0.0291, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.472234708008795e-06, |
|
"loss": 0.0688, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.4542592316988106e-06, |
|
"loss": 0.0574, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.436324808109227e-06, |
|
"loss": 0.0695, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.4184317971374866e-06, |
|
"loss": 0.1057, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.4005805578499946e-06, |
|
"loss": 0.0234, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.3827714484748997e-06, |
|
"loss": 0.0852, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.3650048263949187e-06, |
|
"loss": 0.0613, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.3472810481401524e-06, |
|
"loss": 0.0783, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.329600469380944e-06, |
|
"loss": 0.0732, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.3119634449207286e-06, |
|
"loss": 0.0728, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.294370328688919e-06, |
|
"loss": 0.1218, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.2768214737338107e-06, |
|
"loss": 0.0373, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.2593172322154817e-06, |
|
"loss": 0.0424, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.24185795539874e-06, |
|
"loss": 0.0805, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.224443993646062e-06, |
|
"loss": 0.0652, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.207075696410577e-06, |
|
"loss": 0.0688, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.1897534122290383e-06, |
|
"loss": 0.0305, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.1724774887148365e-06, |
|
"loss": 0.0273, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.1552482725510296e-06, |
|
"loss": 0.051, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.1380661094833717e-06, |
|
"loss": 0.0281, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.120931344313393e-06, |
|
"loss": 0.0893, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.1038443208914606e-06, |
|
"loss": 0.0333, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.0868053821098957e-06, |
|
"loss": 0.0247, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.0698148698960826e-06, |
|
"loss": 0.0781, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.0528731252056063e-06, |
|
"loss": 0.0861, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.035980488015422e-06, |
|
"loss": 0.0862, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.0191372973170156e-06, |
|
"loss": 0.012, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.0023438911096174e-06, |
|
"loss": 0.0289, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.9856006063934057e-06, |
|
"loss": 0.0076, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.9689077791627564e-06, |
|
"loss": 0.0455, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.952265744399489e-06, |
|
"loss": 0.1003, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.935674836066151e-06, |
|
"loss": 0.1002, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.9191353870993173e-06, |
|
"loss": 0.098, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.902647729402903e-06, |
|
"loss": 0.0318, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8862121938415108e-06, |
|
"loss": 0.0779, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8698291102337817e-06, |
|
"loss": 0.013, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8534988073457876e-06, |
|
"loss": 0.0663, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8372216128844242e-06, |
|
"loss": 0.0701, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.820997853490837e-06, |
|
"loss": 0.027, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8048278547338745e-06, |
|
"loss": 0.0587, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.788711941103541e-06, |
|
"loss": 0.1021, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7726504360044994e-06, |
|
"loss": 0.0984, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7566436617495665e-06, |
|
"loss": 0.0771, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7406919395532607e-06, |
|
"loss": 0.0779, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7247955895253417e-06, |
|
"loss": 0.0265, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7089549306643985e-06, |
|
"loss": 0.0804, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.10553926229476929, |
|
"eval_runtime": 447.3537, |
|
"eval_samples_per_second": 52.261, |
|
"eval_steps_per_second": 8.711, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.6931702808514368e-06, |
|
"loss": 0.1135, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.6774419568435118e-06, |
|
"loss": 0.0521, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.6617702742673597e-06, |
|
"loss": 0.048, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.6461555476130759e-06, |
|
"loss": 0.0534, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.630598090227791e-06, |
|
"loss": 0.0355, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.615098214309398e-06, |
|
"loss": 0.0472, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5996562309002698e-06, |
|
"loss": 0.0735, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5842724498810353e-06, |
|
"loss": 0.0217, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5689471799643443e-06, |
|
"loss": 0.0754, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5536807286886871e-06, |
|
"loss": 0.0841, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.53847340241221e-06, |
|
"loss": 0.0436, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5233255063065798e-06, |
|
"loss": 0.0282, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5082373443508486e-06, |
|
"loss": 0.036, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.4932092193253592e-06, |
|
"loss": 0.0263, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.4782414328056736e-06, |
|
"loss": 0.1591, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.4633342851565094e-06, |
|
"loss": 0.0599, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.4484880755257256e-06, |
|
"loss": 0.1064, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.4337031018383067e-06, |
|
"loss": 0.0376, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.418979660790396e-06, |
|
"loss": 0.0918, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.4043180478433327e-06, |
|
"loss": 0.057, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.3897185572177244e-06, |
|
"loss": 0.0387, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.375181481887551e-06, |
|
"loss": 0.0826, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.360707113574272e-06, |
|
"loss": 0.043, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.3462957427409852e-06, |
|
"loss": 0.0494, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.3319476585865862e-06, |
|
"loss": 0.0351, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.317663149039977e-06, |
|
"loss": 0.0078, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.3034425007542776e-06, |
|
"loss": 0.0252, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.289285999101078e-06, |
|
"loss": 0.0471, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.2751939281647143e-06, |
|
"loss": 0.0454, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2611665707365612e-06, |
|
"loss": 0.0639, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2472042083093647e-06, |
|
"loss": 0.1117, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2333071210715855e-06, |
|
"loss": 0.0734, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2194755879017843e-06, |
|
"loss": 0.0301, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2057098863630184e-06, |
|
"loss": 0.0352, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.1920102926972738e-06, |
|
"loss": 0.0521, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.1783770818199275e-06, |
|
"loss": 0.0261, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.1648105273142198e-06, |
|
"loss": 0.0568, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.151310901425776e-06, |
|
"loss": 0.0203, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.1378784750571335e-06, |
|
"loss": 0.083, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.1245135177623087e-06, |
|
"loss": 0.0551, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.1112162977413925e-06, |
|
"loss": 0.0085, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.0979870818351596e-06, |
|
"loss": 0.0411, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.0848261355197223e-06, |
|
"loss": 0.0857, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.0717337229011941e-06, |
|
"loss": 0.0282, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.058710106710401e-06, |
|
"loss": 0.0527, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.0457555482975958e-06, |
|
"loss": 0.0702, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.0328703076272267e-06, |
|
"loss": 0.0824, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.0200546432727093e-06, |
|
"loss": 0.0826, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.007308812411247e-06, |
|
"loss": 0.0542, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.94633070818662e-07, |
|
"loss": 0.0886, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.10989828407764435, |
|
"eval_runtime": 447.8187, |
|
"eval_samples_per_second": 52.206, |
|
"eval_steps_per_second": 8.702, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.8202767286427e-07, |
|
"loss": 0.0689, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.69492871505768e-07, |
|
"loss": 0.0212, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.570289182841664e-07, |
|
"loss": 0.0596, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.44636063318732e-07, |
|
"loss": 0.0136, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.323145553019798e-07, |
|
"loss": 0.0335, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.200646414946699e-07, |
|
"loss": 0.041, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.078865677208571e-07, |
|
"loss": 0.0312, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.957805783629492e-07, |
|
"loss": 0.0638, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.837469163568095e-07, |
|
"loss": 0.0227, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.717858231868755e-07, |
|
"loss": 0.0649, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.59897538881317e-07, |
|
"loss": 0.0137, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.480823020072212e-07, |
|
"loss": 0.0243, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.363403496657983e-07, |
|
"loss": 0.0273, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.246719174876306e-07, |
|
"loss": 0.0572, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.13077239627939e-07, |
|
"loss": 0.0629, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.015565487618886e-07, |
|
"loss": 0.0406, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.901100760799157e-07, |
|
"loss": 0.0314, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.78738051283088e-07, |
|
"loss": 0.0787, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.674407025785011e-07, |
|
"loss": 0.1002, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.5621825667469e-07, |
|
"loss": 0.1132, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.4507093877709e-07, |
|
"loss": 0.0525, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.33998972583506e-07, |
|
"loss": 0.035, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.230025802796343e-07, |
|
"loss": 0.0541, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.120819825345952e-07, |
|
"loss": 0.0394, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.01237398496509e-07, |
|
"loss": 0.0442, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.904690457881011e-07, |
|
"loss": 0.066, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.797771405023263e-07, |
|
"loss": 0.032, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.69161897198043e-07, |
|
"loss": 0.0657, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.586235288956968e-07, |
|
"loss": 0.0427, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.481622470730558e-07, |
|
"loss": 0.0726, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.377782616609594e-07, |
|
"loss": 0.0229, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.274717810391067e-07, |
|
"loss": 0.0292, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.172430120318808e-07, |
|
"loss": 0.0189, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.070921599041899e-07, |
|
"loss": 0.0462, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 5.970194283573546e-07, |
|
"loss": 0.0274, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.870250195250159e-07, |
|
"loss": 0.0131, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.771091339690828e-07, |
|
"loss": 0.0304, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.672719706757034e-07, |
|
"loss": 0.0352, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.575137270512739e-07, |
|
"loss": 0.0677, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.478345989184799e-07, |
|
"loss": 0.0441, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.382347805123599e-07, |
|
"loss": 0.0603, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.28714464476415e-07, |
|
"loss": 0.0078, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.192738418587362e-07, |
|
"loss": 0.0277, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.099131021081786e-07, |
|
"loss": 0.0442, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.006324330705489e-07, |
|
"loss": 0.0308, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.914320209848486e-07, |
|
"loss": 0.0695, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.823120504795239e-07, |
|
"loss": 0.0592, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.732727045687714e-07, |
|
"loss": 0.0112, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.643141646488575e-07, |
|
"loss": 0.0376, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.554366104944838e-07, |
|
"loss": 0.0591, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.127273827791214, |
|
"eval_runtime": 447.0287, |
|
"eval_samples_per_second": 52.299, |
|
"eval_steps_per_second": 8.718, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.4664022025517597e-07, |
|
"loss": 0.057, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.379251704517115e-07, |
|
"loss": 0.1099, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.2929163597257356e-07, |
|
"loss": 0.0713, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.2073979007044633e-07, |
|
"loss": 0.0428, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.1226980435873386e-07, |
|
"loss": 0.0226, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.038818488081201e-07, |
|
"loss": 0.0673, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.9557609174315347e-07, |
|
"loss": 0.0356, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.873526998388746e-07, |
|
"loss": 0.0083, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.79211838117466e-07, |
|
"loss": 0.0356, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.711536699449439e-07, |
|
"loss": 0.0497, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.631783570278803e-07, |
|
"loss": 0.0685, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.55286059410155e-07, |
|
"loss": 0.0371, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.474769354697472e-07, |
|
"loss": 0.0806, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.3975114191555417e-07, |
|
"loss": 0.0322, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.3210883378425e-07, |
|
"loss": 0.0328, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.245501644371708e-07, |
|
"loss": 0.0421, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.1707528555723923e-07, |
|
"loss": 0.0283, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.0968434714592165e-07, |
|
"loss": 0.1675, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.0237749752021366e-07, |
|
"loss": 0.0674, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.951548833096703e-07, |
|
"loss": 0.0235, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.880166494534562e-07, |
|
"loss": 0.0461, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.8096293919744403e-07, |
|
"loss": 0.0475, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.739938940913348e-07, |
|
"loss": 0.0303, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.67109653985819e-07, |
|
"loss": 0.065, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.6031035702977194e-07, |
|
"loss": 0.0323, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.5359613966747707e-07, |
|
"loss": 0.0829, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.4696713663589493e-07, |
|
"loss": 0.0269, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.4042348096195e-07, |
|
"loss": 0.0394, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.3396530395987127e-07, |
|
"loss": 0.0427, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.2759273522854835e-07, |
|
"loss": 0.051, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.2130590264893507e-07, |
|
"loss": 0.0342, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.1510493238148555e-07, |
|
"loss": 0.0473, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.0898994886361512e-07, |
|
"loss": 0.0465, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.0296107480721294e-07, |
|
"loss": 0.0147, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.970184311961684e-07, |
|
"loss": 0.0332, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.9116213728395466e-07, |
|
"loss": 0.0474, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8539231059122636e-07, |
|
"loss": 0.021, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7970906690346523e-07, |
|
"loss": 0.052, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.741125202686583e-07, |
|
"loss": 0.0221, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.6860278299500536e-07, |
|
"loss": 0.0104, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.6317996564866812e-07, |
|
"loss": 0.0247, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5784417705154953e-07, |
|
"loss": 0.0746, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5259552427911212e-07, |
|
"loss": 0.0246, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.474341126582261e-07, |
|
"loss": 0.0253, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4236004576505963e-07, |
|
"loss": 0.0318, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.373734254229969e-07, |
|
"loss": 0.0498, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.3247435170059792e-07, |
|
"loss": 0.0332, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.2766292290958636e-07, |
|
"loss": 0.0655, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2293923560288096e-07, |
|
"loss": 0.0498, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1830338457265578e-07, |
|
"loss": 0.0424, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.1212323009967804, |
|
"eval_runtime": 445.1384, |
|
"eval_samples_per_second": 52.521, |
|
"eval_steps_per_second": 8.755, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1375546284843791e-07, |
|
"loss": 0.0113, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0929556169523941e-07, |
|
"loss": 0.0747, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0492377061173075e-07, |
|
"loss": 0.0955, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0064017732843828e-07, |
|
"loss": 0.0707, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.64448678059886e-08, |
|
"loss": 0.0268, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.233792623338149e-08, |
|
"loss": 0.0124, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.831943502630074e-08, |
|
"loss": 0.0593, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.438947482545995e-08, |
|
"loss": 0.0237, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.054812449498434e-08, |
|
"loss": 0.0219, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.679546112083052e-08, |
|
"loss": 0.0184, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.313156000923476e-08, |
|
"loss": 0.0831, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 6.955649468520613e-08, |
|
"loss": 0.0344, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 6.607033689104925e-08, |
|
"loss": 0.0463, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 6.267315658492484e-08, |
|
"loss": 0.0162, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 5.9365021939445485e-08, |
|
"loss": 0.0374, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 5.614599934030812e-08, |
|
"loss": 0.0378, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 5.301615338496213e-08, |
|
"loss": 0.0657, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.99755468813109e-08, |
|
"loss": 0.0945, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.702424084645579e-08, |
|
"loss": 0.0539, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.4162294505465664e-08, |
|
"loss": 0.0156, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.138976529019478e-08, |
|
"loss": 0.0373, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.870670883812577e-08, |
|
"loss": 0.0703, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.611317899125399e-08, |
|
"loss": 0.0328, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.3609227795008914e-08, |
|
"loss": 0.0084, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.1194905497207435e-08, |
|
"loss": 0.0554, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.8870260547047223e-08, |
|
"loss": 0.0278, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6635339594132946e-08, |
|
"loss": 0.0248, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.449018748754206e-08, |
|
"loss": 0.0269, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.2434847274923508e-08, |
|
"loss": 0.0208, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.0469360201632426e-08, |
|
"loss": 0.0535, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.8593765709906286e-08, |
|
"loss": 0.042, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.6808101438070045e-08, |
|
"loss": 0.0451, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.5112403219781746e-08, |
|
"loss": 0.0391, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3506705083315086e-08, |
|
"loss": 0.1109, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.199103925087347e-08, |
|
"loss": 0.0667, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0565436137946028e-08, |
|
"loss": 0.0145, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.229924352696095e-09, |
|
"loss": 0.0314, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.984530695387182e-09, |
|
"loss": 0.0173, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 6.8292801578453964e-09, |
|
"loss": 0.0851, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 5.764195922958348e-09, |
|
"loss": 0.0334, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.789299364208022e-09, |
|
"loss": 0.0735, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.904610045245116e-09, |
|
"loss": 0.0734, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.110145719492863e-09, |
|
"loss": 0.0666, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.4059223297938084e-09, |
|
"loss": 0.0156, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7919540080875707e-09, |
|
"loss": 0.062, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.2682530751290666e-09, |
|
"loss": 0.0855, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.348300402397091e-10, |
|
"loss": 0.0876, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.916936010980754e-10, |
|
"loss": 0.0559, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.388506435635485e-10, |
|
"loss": 0.0508, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.630624153992516e-11, |
|
"loss": 0.122, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.0, |
|
"eval_loss": 0.12235034257173538, |
|
"eval_runtime": 448.3547, |
|
"eval_samples_per_second": 52.144, |
|
"eval_steps_per_second": 8.692, |
|
"step": 7000 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 7013, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"total_flos": 0.0, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|