|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9977761304670127, |
|
"eval_steps": 500, |
|
"global_step": 2022, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.014825796886582653, |
|
"grad_norm": 1.3803142660082024, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6862, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.029651593773165306, |
|
"grad_norm": 1.790122423429332, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6305, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04447739065974796, |
|
"grad_norm": 0.8497603342076308, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5946, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05930318754633061, |
|
"grad_norm": 0.7732974567839591, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5858, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07412898443291327, |
|
"grad_norm": 0.8204168744024732, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5662, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.08895478131949593, |
|
"grad_norm": 0.7530005874020861, |
|
"learning_rate": 5e-06, |
|
"loss": 0.564, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.10378057820607858, |
|
"grad_norm": 0.6979206993642474, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5536, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.11860637509266123, |
|
"grad_norm": 0.9504756484437652, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5447, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1334321719792439, |
|
"grad_norm": 0.8937944128313379, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5466, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.14825796886582654, |
|
"grad_norm": 0.6670560083366487, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5354, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.16308376575240918, |
|
"grad_norm": 0.7364925811772932, |
|
"learning_rate": 5e-06, |
|
"loss": 0.535, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.17790956263899185, |
|
"grad_norm": 0.7183776415658619, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5338, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1927353595255745, |
|
"grad_norm": 0.5967155900573925, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5301, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.20756115641215717, |
|
"grad_norm": 0.5747353912092298, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5263, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2223869532987398, |
|
"grad_norm": 0.6159997079939371, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5248, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.23721275018532245, |
|
"grad_norm": 0.5680164570830968, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5279, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.2520385470719051, |
|
"grad_norm": 0.5992456838263306, |
|
"learning_rate": 5e-06, |
|
"loss": 0.526, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.2668643439584878, |
|
"grad_norm": 0.5930479122165856, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5224, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.28169014084507044, |
|
"grad_norm": 0.5912410054870685, |
|
"learning_rate": 5e-06, |
|
"loss": 0.52, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.2965159377316531, |
|
"grad_norm": 0.5894696404162991, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5179, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3113417346182357, |
|
"grad_norm": 0.5643703980877148, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5182, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.32616753150481836, |
|
"grad_norm": 0.5861129018729444, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5104, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.34099332839140106, |
|
"grad_norm": 0.6236624375674511, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5131, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3558191252779837, |
|
"grad_norm": 0.5812434696304949, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5191, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.37064492216456635, |
|
"grad_norm": 0.7181156036216183, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5093, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.385470719051149, |
|
"grad_norm": 0.587369525728773, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5161, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.40029651593773163, |
|
"grad_norm": 0.5460966546871288, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5178, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.41512231282431433, |
|
"grad_norm": 0.5876447069011888, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5142, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.429948109710897, |
|
"grad_norm": 0.816294324192534, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5125, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.4447739065974796, |
|
"grad_norm": 0.5492628600148616, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5078, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.45959970348406226, |
|
"grad_norm": 0.579955709352068, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5045, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.4744255003706449, |
|
"grad_norm": 0.6279455964566482, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5047, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.4892512972572276, |
|
"grad_norm": 0.6390417903003635, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5067, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5040770941438102, |
|
"grad_norm": 0.5816721190754486, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5135, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5189028910303929, |
|
"grad_norm": 0.6458696214772466, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5034, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5337286879169756, |
|
"grad_norm": 0.6957712465947605, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5048, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5485544848035582, |
|
"grad_norm": 0.543098433920077, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4991, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.5633802816901409, |
|
"grad_norm": 0.7663819453316695, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5025, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.5782060785767235, |
|
"grad_norm": 0.5194777261335899, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5072, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.5930318754633062, |
|
"grad_norm": 0.8363946362832739, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4948, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6078576723498889, |
|
"grad_norm": 0.5430505650241441, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5024, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6226834692364714, |
|
"grad_norm": 0.5574129318838634, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5034, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.6375092661230541, |
|
"grad_norm": 0.701668057920261, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4912, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.6523350630096367, |
|
"grad_norm": 0.5952225478810055, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4898, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.6671608598962194, |
|
"grad_norm": 0.585331803107272, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5027, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.6819866567828021, |
|
"grad_norm": 0.5906175115867641, |
|
"learning_rate": 5e-06, |
|
"loss": 0.49, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.6968124536693847, |
|
"grad_norm": 0.5782512807193211, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4962, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.7116382505559674, |
|
"grad_norm": 0.528401047427795, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4979, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.72646404744255, |
|
"grad_norm": 0.7128769222059271, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4982, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7412898443291327, |
|
"grad_norm": 0.5767857437742224, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4891, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.7561156412157154, |
|
"grad_norm": 0.6930280814580314, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5006, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.770941438102298, |
|
"grad_norm": 0.6530314579756452, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4966, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.7857672349888807, |
|
"grad_norm": 0.6639683423874188, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4962, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8005930318754633, |
|
"grad_norm": 0.5556537170587046, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4947, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.815418828762046, |
|
"grad_norm": 0.5740824829726571, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4919, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8302446256486287, |
|
"grad_norm": 0.5307405473088794, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4864, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.8450704225352113, |
|
"grad_norm": 0.5249601710096029, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4926, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.859896219421794, |
|
"grad_norm": 0.5263542668413514, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4877, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.8747220163083765, |
|
"grad_norm": 0.5027236096291062, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4903, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.8895478131949592, |
|
"grad_norm": 0.687053549139694, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4945, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9043736100815419, |
|
"grad_norm": 0.6064169430923966, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4898, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.9191994069681245, |
|
"grad_norm": 0.5127672913798731, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4916, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.9340252038547072, |
|
"grad_norm": 0.5911308230908845, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4876, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.9488510007412898, |
|
"grad_norm": 0.5318485612638623, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4911, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.9636767976278725, |
|
"grad_norm": 0.599135799828323, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4907, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.9785025945144552, |
|
"grad_norm": 0.5017158299240659, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4864, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.9933283914010378, |
|
"grad_norm": 0.638905966589537, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4886, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.9992587101556709, |
|
"eval_loss": 0.489400178194046, |
|
"eval_runtime": 119.2986, |
|
"eval_samples_per_second": 152.332, |
|
"eval_steps_per_second": 0.595, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.0081541882876204, |
|
"grad_norm": 0.8019081623682851, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4775, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.0229799851742032, |
|
"grad_norm": 0.5833948483649493, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4344, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.0378057820607858, |
|
"grad_norm": 0.5717232078953663, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4372, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 0.5320074952877888, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4376, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.0674573758339512, |
|
"grad_norm": 0.5726271946361784, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4314, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.0822831727205338, |
|
"grad_norm": 0.596509516956136, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4347, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.0971089696071163, |
|
"grad_norm": 0.5390515390075241, |
|
"learning_rate": 5e-06, |
|
"loss": 0.434, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.111934766493699, |
|
"grad_norm": 0.6129914776744078, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4354, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.1267605633802817, |
|
"grad_norm": 0.5592764678336093, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4368, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.1415863602668643, |
|
"grad_norm": 0.7060047642916599, |
|
"learning_rate": 5e-06, |
|
"loss": 0.443, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.156412157153447, |
|
"grad_norm": 0.5940229109864722, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4433, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.1712379540400297, |
|
"grad_norm": 0.5491622519034774, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4304, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.1860637509266123, |
|
"grad_norm": 0.6079761033641661, |
|
"learning_rate": 5e-06, |
|
"loss": 0.438, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.200889547813195, |
|
"grad_norm": 0.5693457065422755, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4403, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.2157153446997775, |
|
"grad_norm": 0.6505910795677263, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4375, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.2305411415863603, |
|
"grad_norm": 0.6268213064250889, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4373, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.2453669384729429, |
|
"grad_norm": 0.5919676508645508, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4306, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.2601927353595257, |
|
"grad_norm": 0.5398299551449609, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4353, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.2750185322461083, |
|
"grad_norm": 0.5424618353666834, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4376, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.2898443291326909, |
|
"grad_norm": 0.5634395154305121, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4361, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.3046701260192735, |
|
"grad_norm": 0.5785455634852342, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4367, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.3194959229058563, |
|
"grad_norm": 0.620601502244543, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4367, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.3343217197924389, |
|
"grad_norm": 0.530395060539937, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4347, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.3491475166790214, |
|
"grad_norm": 0.5770452107287607, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4391, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.3639733135656043, |
|
"grad_norm": 0.5858170710249173, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4343, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.3787991104521868, |
|
"grad_norm": 0.5500756623276715, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4346, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.3936249073387694, |
|
"grad_norm": 0.6255449782390206, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4389, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.408450704225352, |
|
"grad_norm": 0.5460524509865728, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4369, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.4232765011119348, |
|
"grad_norm": 0.5924885663735421, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4338, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.4381022979985174, |
|
"grad_norm": 0.5764645405905721, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4367, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.4529280948851, |
|
"grad_norm": 0.5850117598054916, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4362, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.4677538917716828, |
|
"grad_norm": 0.6081832575580323, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4406, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.4825796886582654, |
|
"grad_norm": 0.543363234220218, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4351, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.497405485544848, |
|
"grad_norm": 0.5448664659996519, |
|
"learning_rate": 5e-06, |
|
"loss": 0.436, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.5122312824314306, |
|
"grad_norm": 0.65441524183388, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4372, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.5270570793180134, |
|
"grad_norm": 0.5280772751509661, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4357, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.541882876204596, |
|
"grad_norm": 0.5723443909631686, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4318, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.5567086730911788, |
|
"grad_norm": 0.5085846011938958, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4323, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.5715344699777614, |
|
"grad_norm": 0.638285906027132, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4415, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.586360266864344, |
|
"grad_norm": 0.571218693857694, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4346, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.6011860637509265, |
|
"grad_norm": 0.517292661892654, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4382, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.6160118606375091, |
|
"grad_norm": 0.5102023179424974, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4323, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.630837657524092, |
|
"grad_norm": 0.5721112226520383, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4378, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.6456634544106745, |
|
"grad_norm": 0.5551329559644929, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4348, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.6604892512972573, |
|
"grad_norm": 0.6119221260699137, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4274, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.67531504818384, |
|
"grad_norm": 0.5333241050316515, |
|
"learning_rate": 5e-06, |
|
"loss": 0.432, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.6901408450704225, |
|
"grad_norm": 0.564328179152952, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4391, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.704966641957005, |
|
"grad_norm": 0.5254085440281446, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4359, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.7197924388435877, |
|
"grad_norm": 0.5996709499991413, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4355, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.7346182357301705, |
|
"grad_norm": 0.5493385200435505, |
|
"learning_rate": 5e-06, |
|
"loss": 0.432, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.7494440326167533, |
|
"grad_norm": 0.5906833738744632, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4367, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.7642698295033359, |
|
"grad_norm": 0.5341688856567031, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4298, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.7790956263899185, |
|
"grad_norm": 0.533463029816806, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4344, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.793921423276501, |
|
"grad_norm": 0.5606747279800302, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4362, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.8087472201630836, |
|
"grad_norm": 0.5183566067624092, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4271, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.8235730170496665, |
|
"grad_norm": 0.5326764319923035, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4306, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.838398813936249, |
|
"grad_norm": 0.5967407273470395, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4344, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.8532246108228319, |
|
"grad_norm": 0.569200724691502, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4353, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.8680504077094144, |
|
"grad_norm": 0.5364248806496852, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4364, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.882876204595997, |
|
"grad_norm": 0.5385402357985632, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4379, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.8977020014825796, |
|
"grad_norm": 0.5223344068630751, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4347, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.9125277983691622, |
|
"grad_norm": 0.5617283138858877, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4319, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.927353595255745, |
|
"grad_norm": 0.5893151487107228, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4313, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.9421793921423276, |
|
"grad_norm": 0.5366729507399915, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4399, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.9570051890289104, |
|
"grad_norm": 0.5521187635061144, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4346, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.971830985915493, |
|
"grad_norm": 0.5736814101430955, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4344, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.9866567828020756, |
|
"grad_norm": 0.6153323220981107, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4288, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.47887247800827026, |
|
"eval_runtime": 117.4023, |
|
"eval_samples_per_second": 154.793, |
|
"eval_steps_per_second": 0.605, |
|
"step": 1349 |
|
}, |
|
{ |
|
"epoch": 2.001482579688658, |
|
"grad_norm": 0.8057512126070844, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4465, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.0163083765752408, |
|
"grad_norm": 0.6985686766657238, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3758, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.031134173461824, |
|
"grad_norm": 0.7516592382333935, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3739, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.0459599703484064, |
|
"grad_norm": 0.6560426338450871, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3733, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.060785767234989, |
|
"grad_norm": 0.6554981025712345, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3738, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.0756115641215716, |
|
"grad_norm": 0.5777290379829684, |
|
"learning_rate": 5e-06, |
|
"loss": 0.373, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.090437361008154, |
|
"grad_norm": 0.6039864607065941, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3726, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 0.5737027873462299, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3745, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.1200889547813193, |
|
"grad_norm": 0.5962503173170325, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3793, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.1349147516679023, |
|
"grad_norm": 0.5896368959935617, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3783, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.149740548554485, |
|
"grad_norm": 0.5817443097816994, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3775, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.1645663454410675, |
|
"grad_norm": 0.5749673449125087, |
|
"learning_rate": 5e-06, |
|
"loss": 0.378, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.17939214232765, |
|
"grad_norm": 0.581781908254146, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3694, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.1942179392142327, |
|
"grad_norm": 0.5991853455621899, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3821, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.2090437361008153, |
|
"grad_norm": 0.5786028266566305, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3761, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.223869532987398, |
|
"grad_norm": 0.6524631416486719, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3733, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.238695329873981, |
|
"grad_norm": 0.66274701181455, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3812, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.2535211267605635, |
|
"grad_norm": 0.5890654308630106, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3785, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.268346923647146, |
|
"grad_norm": 0.5995580394178166, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3766, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.2831727205337287, |
|
"grad_norm": 0.5621523912385117, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3819, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.2979985174203112, |
|
"grad_norm": 0.5631391589331654, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3831, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.312824314306894, |
|
"grad_norm": 0.598394877682484, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3776, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.327650111193477, |
|
"grad_norm": 0.5751635376845493, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3743, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.3424759080800595, |
|
"grad_norm": 0.7431168749670549, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3806, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.357301704966642, |
|
"grad_norm": 0.6148778429363163, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3792, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.3721275018532246, |
|
"grad_norm": 0.5904751763294604, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3779, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.386953298739807, |
|
"grad_norm": 0.6160494878206139, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3795, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.40177909562639, |
|
"grad_norm": 0.6601509703827891, |
|
"learning_rate": 5e-06, |
|
"loss": 0.381, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.4166048925129724, |
|
"grad_norm": 0.5557483892389672, |
|
"learning_rate": 5e-06, |
|
"loss": 0.38, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.431430689399555, |
|
"grad_norm": 0.5514157650195483, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3785, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.446256486286138, |
|
"grad_norm": 0.5896166088948557, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3785, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.4610822831727206, |
|
"grad_norm": 0.6187824694109818, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3866, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.475908080059303, |
|
"grad_norm": 0.6196885650354601, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3773, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.4907338769458858, |
|
"grad_norm": 0.6497541102453935, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3792, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.5055596738324684, |
|
"grad_norm": 0.6221830439300937, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3814, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.5203854707190514, |
|
"grad_norm": 0.5575748670805571, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3788, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.535211267605634, |
|
"grad_norm": 0.5820255267213036, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3797, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.5500370644922166, |
|
"grad_norm": 0.5778840920723647, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3836, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.564862861378799, |
|
"grad_norm": 0.6369372874220104, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3822, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.5796886582653817, |
|
"grad_norm": 0.5836272270823646, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3825, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.5945144551519643, |
|
"grad_norm": 0.5893118830611698, |
|
"learning_rate": 5e-06, |
|
"loss": 0.382, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.609340252038547, |
|
"grad_norm": 0.691740548026418, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3828, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.6241660489251295, |
|
"grad_norm": 0.6016891230542475, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3841, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.6389918458117125, |
|
"grad_norm": 0.5709726000324739, |
|
"learning_rate": 5e-06, |
|
"loss": 0.38, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.653817642698295, |
|
"grad_norm": 0.5707418370568915, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3819, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.6686434395848777, |
|
"grad_norm": 0.6121781418312778, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3775, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.6834692364714603, |
|
"grad_norm": 0.5845205444282813, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3794, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.698295033358043, |
|
"grad_norm": 0.6090204118467554, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3749, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.713120830244626, |
|
"grad_norm": 0.6311725348745423, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3875, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.7279466271312085, |
|
"grad_norm": 0.5855235496792751, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3798, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.742772424017791, |
|
"grad_norm": 0.6360580241341309, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3795, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.7575982209043737, |
|
"grad_norm": 0.6949332357026632, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3833, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.7724240177909563, |
|
"grad_norm": 0.5661995725449918, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3815, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.787249814677539, |
|
"grad_norm": 0.6181579391717933, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3814, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.8020756115641214, |
|
"grad_norm": 0.621591189180585, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3839, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.816901408450704, |
|
"grad_norm": 0.6102808235648141, |
|
"learning_rate": 5e-06, |
|
"loss": 0.386, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.8317272053372866, |
|
"grad_norm": 0.5570401466362744, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3829, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.8465530022238696, |
|
"grad_norm": 0.605864801280795, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3791, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.8613787991104522, |
|
"grad_norm": 0.5737361891170868, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3831, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 2.876204595997035, |
|
"grad_norm": 0.5867020788912005, |
|
"learning_rate": 5e-06, |
|
"loss": 0.38, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.8910303928836174, |
|
"grad_norm": 0.557534237372008, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3825, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 2.9058561897702, |
|
"grad_norm": 0.5964701402385094, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3852, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.920681986656783, |
|
"grad_norm": 0.600588952238815, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3876, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 2.9355077835433656, |
|
"grad_norm": 0.5765877577022089, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3851, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.950333580429948, |
|
"grad_norm": 0.6187803453792526, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3818, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.965159377316531, |
|
"grad_norm": 0.5518073320285379, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3834, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.9799851742031134, |
|
"grad_norm": 0.5343617840090896, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3772, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.994810971089696, |
|
"grad_norm": 0.5629835017462416, |
|
"learning_rate": 5e-06, |
|
"loss": 0.382, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.9977761304670127, |
|
"eval_loss": 0.4834764897823334, |
|
"eval_runtime": 117.0978, |
|
"eval_samples_per_second": 155.195, |
|
"eval_steps_per_second": 0.606, |
|
"step": 2022 |
|
}, |
|
{ |
|
"epoch": 2.9977761304670127, |
|
"step": 2022, |
|
"total_flos": 3386087791656960.0, |
|
"train_loss": 0.4438282892800698, |
|
"train_runtime": 17391.1323, |
|
"train_samples_per_second": 59.561, |
|
"train_steps_per_second": 0.116 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2022, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3386087791656960.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|