|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.197845249755142, |
|
"eval_steps": 76, |
|
"global_step": 303, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.000652954619653934, |
|
"grad_norm": 4.757781028747559, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 4.2047, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.000652954619653934, |
|
"eval_loss": NaN, |
|
"eval_runtime": 704.7675, |
|
"eval_samples_per_second": 3.661, |
|
"eval_steps_per_second": 0.915, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.001305909239307868, |
|
"grad_norm": 4.874630928039551, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 4.1653, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0019588638589618022, |
|
"grad_norm": 5.5883331298828125, |
|
"learning_rate": 1.2e-05, |
|
"loss": 4.1635, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.002611818478615736, |
|
"grad_norm": 5.3879313468933105, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 3.8569, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.00326477309826967, |
|
"grad_norm": 4.848824501037598, |
|
"learning_rate": 2e-05, |
|
"loss": 4.0147, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0039177277179236044, |
|
"grad_norm": 5.3052978515625, |
|
"learning_rate": 2.4e-05, |
|
"loss": 3.7956, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.004570682337577538, |
|
"grad_norm": 6.782866954803467, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 4.6771, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.005223636957231472, |
|
"grad_norm": 6.1247735023498535, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 3.8201, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.005876591576885406, |
|
"grad_norm": 5.305412769317627, |
|
"learning_rate": 3.6e-05, |
|
"loss": 3.2011, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.00652954619653934, |
|
"grad_norm": 4.974228858947754, |
|
"learning_rate": 4e-05, |
|
"loss": 3.2488, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.007182500816193275, |
|
"grad_norm": 4.17997407913208, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 2.8667, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.007835455435847209, |
|
"grad_norm": 5.494494915008545, |
|
"learning_rate": 4.8e-05, |
|
"loss": 3.0199, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.008488410055501142, |
|
"grad_norm": 3.663966178894043, |
|
"learning_rate": 5.2000000000000004e-05, |
|
"loss": 2.5652, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.009141364675155077, |
|
"grad_norm": 3.196805477142334, |
|
"learning_rate": 5.6000000000000006e-05, |
|
"loss": 2.6467, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.009794319294809012, |
|
"grad_norm": 3.161735773086548, |
|
"learning_rate": 6e-05, |
|
"loss": 2.4477, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.010447273914462945, |
|
"grad_norm": 4.473456859588623, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 2.4943, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.01110022853411688, |
|
"grad_norm": 5.138904094696045, |
|
"learning_rate": 6.800000000000001e-05, |
|
"loss": 2.51, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.011753183153770812, |
|
"grad_norm": 5.232083320617676, |
|
"learning_rate": 7.2e-05, |
|
"loss": 2.3384, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.012406137773424747, |
|
"grad_norm": 5.67042875289917, |
|
"learning_rate": 7.6e-05, |
|
"loss": 2.3158, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.01305909239307868, |
|
"grad_norm": 6.414549350738525, |
|
"learning_rate": 8e-05, |
|
"loss": 2.2479, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.013712047012732615, |
|
"grad_norm": 6.064276218414307, |
|
"learning_rate": 8.4e-05, |
|
"loss": 2.689, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.01436500163238655, |
|
"grad_norm": 4.989445686340332, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 1.9927, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.015017956252040483, |
|
"grad_norm": 6.970681667327881, |
|
"learning_rate": 9.200000000000001e-05, |
|
"loss": 2.0652, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.015670910871694418, |
|
"grad_norm": 5.712936878204346, |
|
"learning_rate": 9.6e-05, |
|
"loss": 1.7749, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.01632386549134835, |
|
"grad_norm": 6.509803771972656, |
|
"learning_rate": 0.0001, |
|
"loss": 1.6381, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.016976820111002284, |
|
"grad_norm": 9.46315860748291, |
|
"learning_rate": 0.00010400000000000001, |
|
"loss": 2.9097, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.01762977473065622, |
|
"grad_norm": 7.60649299621582, |
|
"learning_rate": 0.00010800000000000001, |
|
"loss": 2.3185, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.018282729350310153, |
|
"grad_norm": 4.304985046386719, |
|
"learning_rate": 0.00011200000000000001, |
|
"loss": 1.8866, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.018935683969964087, |
|
"grad_norm": 4.42153787612915, |
|
"learning_rate": 0.000116, |
|
"loss": 1.9317, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.019588638589618023, |
|
"grad_norm": 4.993203163146973, |
|
"learning_rate": 0.00012, |
|
"loss": 1.963, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.020241593209271956, |
|
"grad_norm": 5.630634784698486, |
|
"learning_rate": 0.000124, |
|
"loss": 2.0134, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.02089454782892589, |
|
"grad_norm": 4.011257648468018, |
|
"learning_rate": 0.00012800000000000002, |
|
"loss": 1.5838, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.021547502448579822, |
|
"grad_norm": 4.430102348327637, |
|
"learning_rate": 0.000132, |
|
"loss": 1.567, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.02220045706823376, |
|
"grad_norm": 4.285562038421631, |
|
"learning_rate": 0.00013600000000000003, |
|
"loss": 1.5668, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.022853411687887692, |
|
"grad_norm": 4.76671028137207, |
|
"learning_rate": 0.00014, |
|
"loss": 1.4504, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.023506366307541625, |
|
"grad_norm": 3.6335926055908203, |
|
"learning_rate": 0.000144, |
|
"loss": 1.7653, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.02415932092719556, |
|
"grad_norm": 3.863640546798706, |
|
"learning_rate": 0.000148, |
|
"loss": 1.7619, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.024812275546849494, |
|
"grad_norm": 3.339837074279785, |
|
"learning_rate": 0.000152, |
|
"loss": 1.9767, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.025465230166503428, |
|
"grad_norm": 3.004544734954834, |
|
"learning_rate": 0.00015600000000000002, |
|
"loss": 1.954, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.02611818478615736, |
|
"grad_norm": 3.1160991191864014, |
|
"learning_rate": 0.00016, |
|
"loss": 2.2944, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.026771139405811297, |
|
"grad_norm": 2.9257686138153076, |
|
"learning_rate": 0.000164, |
|
"loss": 2.1756, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.02742409402546523, |
|
"grad_norm": 3.0116686820983887, |
|
"learning_rate": 0.000168, |
|
"loss": 2.2031, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.028077048645119163, |
|
"grad_norm": 3.1789779663085938, |
|
"learning_rate": 0.000172, |
|
"loss": 1.71, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.0287300032647731, |
|
"grad_norm": 3.2652573585510254, |
|
"learning_rate": 0.00017600000000000002, |
|
"loss": 2.108, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.029382957884427033, |
|
"grad_norm": 3.3591766357421875, |
|
"learning_rate": 0.00018, |
|
"loss": 2.252, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.030035912504080966, |
|
"grad_norm": 4.020256519317627, |
|
"learning_rate": 0.00018400000000000003, |
|
"loss": 2.0209, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.0306888671237349, |
|
"grad_norm": 3.1844100952148438, |
|
"learning_rate": 0.000188, |
|
"loss": 1.7205, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.031341821743388835, |
|
"grad_norm": 4.041150093078613, |
|
"learning_rate": 0.000192, |
|
"loss": 2.0322, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.031994776363042765, |
|
"grad_norm": 3.976959228515625, |
|
"learning_rate": 0.000196, |
|
"loss": 1.5729, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.0326477309826967, |
|
"grad_norm": 3.8643813133239746, |
|
"learning_rate": 0.0002, |
|
"loss": 1.3921, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.03330068560235064, |
|
"grad_norm": 4.523630142211914, |
|
"learning_rate": 0.0001999922905547776, |
|
"loss": 2.0141, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.03395364022200457, |
|
"grad_norm": 3.308136463165283, |
|
"learning_rate": 0.0001999691634078213, |
|
"loss": 1.6375, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.034606594841658504, |
|
"grad_norm": 3.70528244972229, |
|
"learning_rate": 0.00019993062212508053, |
|
"loss": 2.1218, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.03525954946131244, |
|
"grad_norm": 3.909076690673828, |
|
"learning_rate": 0.0001998766726491935, |
|
"loss": 1.7935, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.03591250408096637, |
|
"grad_norm": 3.2719497680664062, |
|
"learning_rate": 0.00019980732329857076, |
|
"loss": 1.4998, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.03656545870062031, |
|
"grad_norm": 4.357359886169434, |
|
"learning_rate": 0.0001997225847661127, |
|
"loss": 1.7119, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.03721841332027424, |
|
"grad_norm": 3.111414909362793, |
|
"learning_rate": 0.00019962247011756081, |
|
"loss": 1.3332, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.03787136793992817, |
|
"grad_norm": 3.3698601722717285, |
|
"learning_rate": 0.00019950699478948309, |
|
"loss": 1.3576, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.03852432255958211, |
|
"grad_norm": 4.0696024894714355, |
|
"learning_rate": 0.00019937617658689384, |
|
"loss": 1.2196, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.039177277179236046, |
|
"grad_norm": 3.984119176864624, |
|
"learning_rate": 0.00019923003568050844, |
|
"loss": 1.4652, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.039830231798889976, |
|
"grad_norm": 3.909024238586426, |
|
"learning_rate": 0.00019906859460363307, |
|
"loss": 1.2951, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.04048318641854391, |
|
"grad_norm": 3.550440788269043, |
|
"learning_rate": 0.0001988918782486906, |
|
"loss": 1.4699, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.04113614103819784, |
|
"grad_norm": 3.341071367263794, |
|
"learning_rate": 0.0001986999138633821, |
|
"loss": 2.1781, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.04178909565785178, |
|
"grad_norm": 2.938025951385498, |
|
"learning_rate": 0.00019849273104648592, |
|
"loss": 1.7088, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.042442050277505715, |
|
"grad_norm": 2.956080913543701, |
|
"learning_rate": 0.00019827036174329353, |
|
"loss": 2.0676, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.043095004897159644, |
|
"grad_norm": 2.349592924118042, |
|
"learning_rate": 0.00019803284024068427, |
|
"loss": 1.7783, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.04374795951681358, |
|
"grad_norm": 2.7854583263397217, |
|
"learning_rate": 0.0001977802031618383, |
|
"loss": 2.222, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.04440091413646752, |
|
"grad_norm": 3.6111695766448975, |
|
"learning_rate": 0.00019751248946059014, |
|
"loss": 2.0701, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.04505386875612145, |
|
"grad_norm": 3.226724863052368, |
|
"learning_rate": 0.00019722974041542203, |
|
"loss": 2.2502, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.045706823375775384, |
|
"grad_norm": 3.4755945205688477, |
|
"learning_rate": 0.0001969319996230995, |
|
"loss": 2.0837, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.04635977799542932, |
|
"grad_norm": 3.1673951148986816, |
|
"learning_rate": 0.0001966193129919491, |
|
"loss": 1.8692, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.04701273261508325, |
|
"grad_norm": 3.5966238975524902, |
|
"learning_rate": 0.00019629172873477995, |
|
"loss": 2.0036, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.047665687234737186, |
|
"grad_norm": 4.344339370727539, |
|
"learning_rate": 0.00019594929736144976, |
|
"loss": 1.963, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.04831864185439112, |
|
"grad_norm": 3.6855549812316895, |
|
"learning_rate": 0.00019559207167107684, |
|
"loss": 1.5932, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.04897159647404505, |
|
"grad_norm": 3.7931909561157227, |
|
"learning_rate": 0.000195220106743899, |
|
"loss": 1.272, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.04962455109369899, |
|
"grad_norm": 3.422001361846924, |
|
"learning_rate": 0.00019483345993278093, |
|
"loss": 1.8281, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.04962455109369899, |
|
"eval_loss": NaN, |
|
"eval_runtime": 704.1079, |
|
"eval_samples_per_second": 3.664, |
|
"eval_steps_per_second": 0.916, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.05027750571335292, |
|
"grad_norm": 3.9176759719848633, |
|
"learning_rate": 0.0001944321908543708, |
|
"loss": 1.9556, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.050930460333006855, |
|
"grad_norm": 3.578507423400879, |
|
"learning_rate": 0.00019401636137990816, |
|
"loss": 2.025, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.05158341495266079, |
|
"grad_norm": 3.641873598098755, |
|
"learning_rate": 0.00019358603562568416, |
|
"loss": 1.667, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.05223636957231472, |
|
"grad_norm": 4.218084335327148, |
|
"learning_rate": 0.0001931412799431554, |
|
"loss": 1.797, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.05288932419196866, |
|
"grad_norm": 3.496741533279419, |
|
"learning_rate": 0.0001926821629087133, |
|
"loss": 1.5076, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.053542278811622594, |
|
"grad_norm": 5.606319427490234, |
|
"learning_rate": 0.00019220875531311045, |
|
"loss": 1.5796, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.054195233431276524, |
|
"grad_norm": 3.143007516860962, |
|
"learning_rate": 0.00019172113015054532, |
|
"loss": 1.4222, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.05484818805093046, |
|
"grad_norm": 3.359457015991211, |
|
"learning_rate": 0.00019121936260740752, |
|
"loss": 0.9389, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.0555011426705844, |
|
"grad_norm": 3.911376476287842, |
|
"learning_rate": 0.00019070353005068484, |
|
"loss": 1.8963, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.056154097290238326, |
|
"grad_norm": 3.542954444885254, |
|
"learning_rate": 0.00019017371201603407, |
|
"loss": 1.4677, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.05680705190989226, |
|
"grad_norm": 3.1694722175598145, |
|
"learning_rate": 0.00018962999019551754, |
|
"loss": 1.4803, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.0574600065295462, |
|
"grad_norm": 2.960282564163208, |
|
"learning_rate": 0.00018907244842500704, |
|
"loss": 1.8923, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.05811296114920013, |
|
"grad_norm": 2.6101479530334473, |
|
"learning_rate": 0.00018850117267125738, |
|
"loss": 1.9243, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.058765915768854066, |
|
"grad_norm": 3.0619993209838867, |
|
"learning_rate": 0.00018791625101865117, |
|
"loss": 2.1384, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.059418870388507995, |
|
"grad_norm": 2.6776371002197266, |
|
"learning_rate": 0.0001873177736556172, |
|
"loss": 1.7285, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.06007182500816193, |
|
"grad_norm": 3.687798023223877, |
|
"learning_rate": 0.00018670583286072443, |
|
"loss": 1.8332, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.06072477962781587, |
|
"grad_norm": 2.632847547531128, |
|
"learning_rate": 0.0001860805229884536, |
|
"loss": 1.8342, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.0613777342474698, |
|
"grad_norm": 3.5173568725585938, |
|
"learning_rate": 0.00018544194045464886, |
|
"loss": 2.004, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.062030688867123734, |
|
"grad_norm": 3.2944045066833496, |
|
"learning_rate": 0.0001847901837216515, |
|
"loss": 1.8861, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.06268364348677767, |
|
"grad_norm": 3.5300235748291016, |
|
"learning_rate": 0.00018412535328311814, |
|
"loss": 1.7608, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.06333659810643161, |
|
"grad_norm": 3.253826856613159, |
|
"learning_rate": 0.0001834475516485257, |
|
"loss": 1.9151, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.06398955272608553, |
|
"grad_norm": 3.243023633956909, |
|
"learning_rate": 0.00018275688332736577, |
|
"loss": 1.5671, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.06464250734573947, |
|
"grad_norm": 3.3818089962005615, |
|
"learning_rate": 0.00018205345481302998, |
|
"loss": 1.4077, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.0652954619653934, |
|
"grad_norm": 3.632511615753174, |
|
"learning_rate": 0.00018133737456639044, |
|
"loss": 1.0454, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.06594841658504734, |
|
"grad_norm": 3.6621978282928467, |
|
"learning_rate": 0.0001806087529990758, |
|
"loss": 2.0844, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.06660137120470128, |
|
"grad_norm": 5.99480676651001, |
|
"learning_rate": 0.0001798677024564473, |
|
"loss": 1.8015, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.06725432582435521, |
|
"grad_norm": 3.063887357711792, |
|
"learning_rate": 0.00017911433720027624, |
|
"loss": 1.7182, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.06790728044400914, |
|
"grad_norm": 3.2303333282470703, |
|
"learning_rate": 0.00017834877339112612, |
|
"loss": 1.6701, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.06856023506366307, |
|
"grad_norm": 7.370791435241699, |
|
"learning_rate": 0.000177571129070442, |
|
"loss": 1.6819, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.06921318968331701, |
|
"grad_norm": 3.4059948921203613, |
|
"learning_rate": 0.00017678152414234968, |
|
"loss": 1.3683, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.06986614430297095, |
|
"grad_norm": 4.130568504333496, |
|
"learning_rate": 0.000175980080355168, |
|
"loss": 1.6074, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.07051909892262488, |
|
"grad_norm": 4.288647651672363, |
|
"learning_rate": 0.00017516692128263648, |
|
"loss": 1.2521, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.07117205354227882, |
|
"grad_norm": 3.245211362838745, |
|
"learning_rate": 0.00017434217230486164, |
|
"loss": 1.3333, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.07182500816193274, |
|
"grad_norm": 3.7068018913269043, |
|
"learning_rate": 0.00017350596058898483, |
|
"loss": 1.287, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.07247796278158668, |
|
"grad_norm": 3.418928623199463, |
|
"learning_rate": 0.0001726584150695744, |
|
"loss": 1.3896, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.07313091740124061, |
|
"grad_norm": 3.3947291374206543, |
|
"learning_rate": 0.00017179966642874543, |
|
"loss": 1.7948, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.07378387202089455, |
|
"grad_norm": 2.770167589187622, |
|
"learning_rate": 0.0001709298470760101, |
|
"loss": 1.8008, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.07443682664054849, |
|
"grad_norm": 2.841723918914795, |
|
"learning_rate": 0.00017004909112786144, |
|
"loss": 1.7773, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.07508978126020241, |
|
"grad_norm": 3.010446071624756, |
|
"learning_rate": 0.00016915753438709417, |
|
"loss": 1.9485, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.07574273587985635, |
|
"grad_norm": 2.804893970489502, |
|
"learning_rate": 0.00016825531432186543, |
|
"loss": 1.8283, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.07639569049951028, |
|
"grad_norm": 2.607825517654419, |
|
"learning_rate": 0.00016734257004449862, |
|
"loss": 1.5881, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.07704864511916422, |
|
"grad_norm": 3.1926357746124268, |
|
"learning_rate": 0.00016641944229003395, |
|
"loss": 1.9909, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.07770159973881816, |
|
"grad_norm": 2.6762239933013916, |
|
"learning_rate": 0.00016548607339452853, |
|
"loss": 1.7493, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.07835455435847209, |
|
"grad_norm": 2.9282002449035645, |
|
"learning_rate": 0.00016454260727310978, |
|
"loss": 1.6987, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.07900750897812601, |
|
"grad_norm": 3.6605136394500732, |
|
"learning_rate": 0.00016358918939778536, |
|
"loss": 2.1916, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.07966046359777995, |
|
"grad_norm": 3.031012535095215, |
|
"learning_rate": 0.00016262596677501297, |
|
"loss": 1.9056, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.08031341821743389, |
|
"grad_norm": 3.2578601837158203, |
|
"learning_rate": 0.0001616530879230335, |
|
"loss": 1.5707, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.08096637283708782, |
|
"grad_norm": 3.2448766231536865, |
|
"learning_rate": 0.00016067070284897137, |
|
"loss": 1.43, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.08161932745674176, |
|
"grad_norm": 3.2631771564483643, |
|
"learning_rate": 0.00015967896302570485, |
|
"loss": 1.0851, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.08227228207639568, |
|
"grad_norm": 3.316664457321167, |
|
"learning_rate": 0.0001586780213685108, |
|
"loss": 1.9519, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.08292523669604962, |
|
"grad_norm": 2.7955403327941895, |
|
"learning_rate": 0.00015766803221148673, |
|
"loss": 1.6003, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.08357819131570356, |
|
"grad_norm": 3.2884178161621094, |
|
"learning_rate": 0.0001566491512837543, |
|
"loss": 1.7108, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.0842311459353575, |
|
"grad_norm": 3.472278356552124, |
|
"learning_rate": 0.00015562153568544752, |
|
"loss": 1.8399, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.08488410055501143, |
|
"grad_norm": 2.900644302368164, |
|
"learning_rate": 0.00015458534386348966, |
|
"loss": 1.6259, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.08553705517466537, |
|
"grad_norm": 3.018883228302002, |
|
"learning_rate": 0.0001535407355871626, |
|
"loss": 1.5246, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.08619000979431929, |
|
"grad_norm": 3.375364303588867, |
|
"learning_rate": 0.00015248787192347196, |
|
"loss": 1.6209, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.08684296441397323, |
|
"grad_norm": 2.9012420177459717, |
|
"learning_rate": 0.00015142691521231267, |
|
"loss": 1.2602, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.08749591903362716, |
|
"grad_norm": 4.421230316162109, |
|
"learning_rate": 0.00015035802904143762, |
|
"loss": 1.5445, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.0881488736532811, |
|
"grad_norm": 3.3842790126800537, |
|
"learning_rate": 0.00014928137822123452, |
|
"loss": 1.2163, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.08880182827293504, |
|
"grad_norm": 3.1379024982452393, |
|
"learning_rate": 0.0001481971287593138, |
|
"loss": 1.6229, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.08945478289258897, |
|
"grad_norm": 3.0401148796081543, |
|
"learning_rate": 0.00014710544783491208, |
|
"loss": 2.1325, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.0901077375122429, |
|
"grad_norm": 3.1100378036499023, |
|
"learning_rate": 0.00014600650377311522, |
|
"loss": 1.8436, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.09076069213189683, |
|
"grad_norm": 2.646742105484009, |
|
"learning_rate": 0.00014490046601890405, |
|
"loss": 2.0225, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.09141364675155077, |
|
"grad_norm": 2.7213666439056396, |
|
"learning_rate": 0.00014378750511102826, |
|
"loss": 1.9278, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.0920666013712047, |
|
"grad_norm": 2.6186165809631348, |
|
"learning_rate": 0.00014266779265571087, |
|
"loss": 2.1003, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.09271955599085864, |
|
"grad_norm": 2.469266653060913, |
|
"learning_rate": 0.00014154150130018866, |
|
"loss": 2.1259, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.09337251061051256, |
|
"grad_norm": 2.95566725730896, |
|
"learning_rate": 0.00014040880470609187, |
|
"loss": 2.0982, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.0940254652301665, |
|
"grad_norm": 3.408046007156372, |
|
"learning_rate": 0.00013926987752266735, |
|
"loss": 2.1553, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.09467841984982044, |
|
"grad_norm": 2.5797595977783203, |
|
"learning_rate": 0.00013812489535984981, |
|
"loss": 1.9652, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.09533137446947437, |
|
"grad_norm": 3.9302403926849365, |
|
"learning_rate": 0.00013697403476118454, |
|
"loss": 1.916, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.09598432908912831, |
|
"grad_norm": 3.0103251934051514, |
|
"learning_rate": 0.0001358174731766064, |
|
"loss": 1.5778, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.09663728370878225, |
|
"grad_norm": 3.0252418518066406, |
|
"learning_rate": 0.00013465538893507907, |
|
"loss": 1.862, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.09729023832843617, |
|
"grad_norm": 3.1504366397857666, |
|
"learning_rate": 0.00013348796121709862, |
|
"loss": 1.7159, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.0979431929480901, |
|
"grad_norm": 2.718940019607544, |
|
"learning_rate": 0.00013231537002706594, |
|
"loss": 1.1477, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.09859614756774404, |
|
"grad_norm": 3.6509456634521484, |
|
"learning_rate": 0.0001311377961655319, |
|
"loss": 1.6706, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.09924910218739798, |
|
"grad_norm": 3.2815134525299072, |
|
"learning_rate": 0.00012995542120132017, |
|
"loss": 1.4804, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.09924910218739798, |
|
"eval_loss": NaN, |
|
"eval_runtime": 699.3675, |
|
"eval_samples_per_second": 3.689, |
|
"eval_steps_per_second": 0.922, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.09990205680705191, |
|
"grad_norm": 2.6223411560058594, |
|
"learning_rate": 0.00012876842744353112, |
|
"loss": 1.4696, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.10055501142670584, |
|
"grad_norm": 3.07543683052063, |
|
"learning_rate": 0.00012757699791343186, |
|
"loss": 1.4233, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.10120796604635977, |
|
"grad_norm": 3.1689348220825195, |
|
"learning_rate": 0.0001263813163162364, |
|
"loss": 1.6883, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.10186092066601371, |
|
"grad_norm": 3.1390719413757324, |
|
"learning_rate": 0.00012518156701278019, |
|
"loss": 1.9252, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.10251387528566765, |
|
"grad_norm": 2.6922318935394287, |
|
"learning_rate": 0.00012397793499109404, |
|
"loss": 1.4566, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.10316682990532158, |
|
"grad_norm": 2.970062255859375, |
|
"learning_rate": 0.00012277060583788064, |
|
"loss": 1.4658, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.10381978452497552, |
|
"grad_norm": 3.3430395126342773, |
|
"learning_rate": 0.00012155976570989949, |
|
"loss": 1.1341, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.10447273914462944, |
|
"grad_norm": 3.108964204788208, |
|
"learning_rate": 0.0001203456013052634, |
|
"loss": 1.4, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.10512569376428338, |
|
"grad_norm": 2.5076346397399902, |
|
"learning_rate": 0.00011912829983465168, |
|
"loss": 1.2977, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.10577864838393732, |
|
"grad_norm": 2.782015085220337, |
|
"learning_rate": 0.00011790804899244452, |
|
"loss": 1.5569, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.10643160300359125, |
|
"grad_norm": 2.4178273677825928, |
|
"learning_rate": 0.00011668503692778239, |
|
"loss": 1.588, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.10708455762324519, |
|
"grad_norm": 2.7303292751312256, |
|
"learning_rate": 0.0001154594522155557, |
|
"loss": 1.9145, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.10773751224289912, |
|
"grad_norm": 2.631848096847534, |
|
"learning_rate": 0.00011423148382732853, |
|
"loss": 1.9198, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.10839046686255305, |
|
"grad_norm": 2.4288294315338135, |
|
"learning_rate": 0.00011300132110220134, |
|
"loss": 1.9154, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.10904342148220698, |
|
"grad_norm": 2.742737293243408, |
|
"learning_rate": 0.00011176915371761702, |
|
"loss": 1.8217, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.10969637610186092, |
|
"grad_norm": 2.437319278717041, |
|
"learning_rate": 0.00011053517166011471, |
|
"loss": 1.7938, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.11034933072151486, |
|
"grad_norm": 2.4741907119750977, |
|
"learning_rate": 0.00010929956519603594, |
|
"loss": 1.7694, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.1110022853411688, |
|
"grad_norm": 3.3339791297912598, |
|
"learning_rate": 0.0001080625248421878, |
|
"loss": 1.9348, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.11165523996082272, |
|
"grad_norm": 3.3256616592407227, |
|
"learning_rate": 0.0001068242413364671, |
|
"loss": 1.8868, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.11230819458047665, |
|
"grad_norm": 2.873408317565918, |
|
"learning_rate": 0.00010558490560845107, |
|
"loss": 1.7327, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.11296114920013059, |
|
"grad_norm": 2.9288012981414795, |
|
"learning_rate": 0.00010434470874995781, |
|
"loss": 1.9178, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.11361410381978453, |
|
"grad_norm": 3.4907095432281494, |
|
"learning_rate": 0.00010310384198558225, |
|
"loss": 1.5658, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.11426705843943846, |
|
"grad_norm": 2.5337088108062744, |
|
"learning_rate": 0.00010186249664321139, |
|
"loss": 0.9634, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1149200130590924, |
|
"grad_norm": 3.862973690032959, |
|
"learning_rate": 0.00010062086412452352, |
|
"loss": 1.7911, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.11557296767874632, |
|
"grad_norm": 3.585923194885254, |
|
"learning_rate": 9.93791358754765e-05, |
|
"loss": 1.9018, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.11622592229840026, |
|
"grad_norm": 2.8271729946136475, |
|
"learning_rate": 9.813750335678866e-05, |
|
"loss": 1.5603, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.1168788769180542, |
|
"grad_norm": 2.844407796859741, |
|
"learning_rate": 9.689615801441774e-05, |
|
"loss": 1.4833, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.11753183153770813, |
|
"grad_norm": 3.0475728511810303, |
|
"learning_rate": 9.56552912500422e-05, |
|
"loss": 1.4969, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.11818478615736207, |
|
"grad_norm": 3.360382318496704, |
|
"learning_rate": 9.441509439154895e-05, |
|
"loss": 1.7167, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.11883774077701599, |
|
"grad_norm": 2.9843080043792725, |
|
"learning_rate": 9.317575866353292e-05, |
|
"loss": 1.4516, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.11949069539666993, |
|
"grad_norm": 3.0425808429718018, |
|
"learning_rate": 9.193747515781224e-05, |
|
"loss": 1.2171, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.12014365001632386, |
|
"grad_norm": 3.03841495513916, |
|
"learning_rate": 9.070043480396404e-05, |
|
"loss": 1.2647, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.1207966046359778, |
|
"grad_norm": 2.7172021865844727, |
|
"learning_rate": 8.94648283398853e-05, |
|
"loss": 1.0993, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.12144955925563174, |
|
"grad_norm": 3.215526580810547, |
|
"learning_rate": 8.823084628238298e-05, |
|
"loss": 1.2345, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.12210251387528567, |
|
"grad_norm": 2.8291847705841064, |
|
"learning_rate": 8.699867889779868e-05, |
|
"loss": 1.7698, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.1227554684949396, |
|
"grad_norm": 2.8613245487213135, |
|
"learning_rate": 8.57685161726715e-05, |
|
"loss": 1.9141, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.12340842311459353, |
|
"grad_norm": 2.358774185180664, |
|
"learning_rate": 8.454054778444431e-05, |
|
"loss": 1.7095, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.12406137773424747, |
|
"grad_norm": 2.2548892498016357, |
|
"learning_rate": 8.331496307221762e-05, |
|
"loss": 1.7769, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.1247143323539014, |
|
"grad_norm": 2.9085566997528076, |
|
"learning_rate": 8.209195100755551e-05, |
|
"loss": 1.9351, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.12536728697355534, |
|
"grad_norm": 2.9980623722076416, |
|
"learning_rate": 8.087170016534835e-05, |
|
"loss": 2.0055, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.12602024159320926, |
|
"grad_norm": 2.8316338062286377, |
|
"learning_rate": 7.965439869473664e-05, |
|
"loss": 1.8518, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.12667319621286321, |
|
"grad_norm": 2.983452081680298, |
|
"learning_rate": 7.84402342901005e-05, |
|
"loss": 1.996, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.12732615083251714, |
|
"grad_norm": 2.925206184387207, |
|
"learning_rate": 7.722939416211937e-05, |
|
"loss": 1.9398, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.12797910545217106, |
|
"grad_norm": 2.671654462814331, |
|
"learning_rate": 7.6022065008906e-05, |
|
"loss": 1.7095, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.128632060071825, |
|
"grad_norm": 3.020463705062866, |
|
"learning_rate": 7.481843298721982e-05, |
|
"loss": 1.82, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.12928501469147893, |
|
"grad_norm": 2.837937593460083, |
|
"learning_rate": 7.361868368376364e-05, |
|
"loss": 1.5275, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.12993796931113288, |
|
"grad_norm": 3.599421739578247, |
|
"learning_rate": 7.242300208656814e-05, |
|
"loss": 1.2762, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.1305909239307868, |
|
"grad_norm": 3.3294129371643066, |
|
"learning_rate": 7.123157255646888e-05, |
|
"loss": 1.0834, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.13124387855044076, |
|
"grad_norm": 3.3265252113342285, |
|
"learning_rate": 7.004457879867986e-05, |
|
"loss": 1.8808, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.13189683317009468, |
|
"grad_norm": 3.3643386363983154, |
|
"learning_rate": 6.886220383446814e-05, |
|
"loss": 1.5511, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.1325497877897486, |
|
"grad_norm": 2.6517298221588135, |
|
"learning_rate": 6.768462997293413e-05, |
|
"loss": 1.5762, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.13320274240940255, |
|
"grad_norm": 3.111353874206543, |
|
"learning_rate": 6.651203878290139e-05, |
|
"loss": 1.4893, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.13385569702905648, |
|
"grad_norm": 2.823064088821411, |
|
"learning_rate": 6.534461106492093e-05, |
|
"loss": 1.658, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.13450865164871043, |
|
"grad_norm": 2.825458526611328, |
|
"learning_rate": 6.418252682339361e-05, |
|
"loss": 1.6023, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.13516160626836435, |
|
"grad_norm": 2.8322906494140625, |
|
"learning_rate": 6.30259652388155e-05, |
|
"loss": 1.4659, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.13581456088801827, |
|
"grad_norm": 2.5433881282806396, |
|
"learning_rate": 6.187510464015022e-05, |
|
"loss": 1.3384, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.13646751550767222, |
|
"grad_norm": 3.240464925765991, |
|
"learning_rate": 6.0730122477332675e-05, |
|
"loss": 0.9005, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.13712047012732614, |
|
"grad_norm": 3.5450103282928467, |
|
"learning_rate": 5.9591195293908174e-05, |
|
"loss": 1.3865, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.1377734247469801, |
|
"grad_norm": 2.67393159866333, |
|
"learning_rate": 5.845849869981137e-05, |
|
"loss": 1.2309, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.13842637936663402, |
|
"grad_norm": 3.3300986289978027, |
|
"learning_rate": 5.733220734428919e-05, |
|
"loss": 1.9498, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.13907933398628794, |
|
"grad_norm": 2.3089957237243652, |
|
"learning_rate": 5.621249488897176e-05, |
|
"loss": 1.7914, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.1397322886059419, |
|
"grad_norm": 2.3583531379699707, |
|
"learning_rate": 5.509953398109594e-05, |
|
"loss": 1.6828, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.1403852432255958, |
|
"grad_norm": 2.0636346340179443, |
|
"learning_rate": 5.399349622688479e-05, |
|
"loss": 1.7596, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.14103819784524976, |
|
"grad_norm": 2.3721492290496826, |
|
"learning_rate": 5.2894552165087916e-05, |
|
"loss": 1.8535, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.14169115246490369, |
|
"grad_norm": 2.338106632232666, |
|
"learning_rate": 5.1802871240686234e-05, |
|
"loss": 1.8166, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.14234410708455764, |
|
"grad_norm": 2.8886420726776123, |
|
"learning_rate": 5.0718621778765476e-05, |
|
"loss": 1.8643, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.14299706170421156, |
|
"grad_norm": 2.71232533454895, |
|
"learning_rate": 4.9641970958562366e-05, |
|
"loss": 1.5456, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.14365001632386548, |
|
"grad_norm": 3.262185573577881, |
|
"learning_rate": 4.8573084787687326e-05, |
|
"loss": 1.6296, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.14430297094351943, |
|
"grad_norm": 2.6300837993621826, |
|
"learning_rate": 4.751212807652806e-05, |
|
"loss": 1.6135, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.14495592556317335, |
|
"grad_norm": 3.008484363555908, |
|
"learning_rate": 4.6459264412837454e-05, |
|
"loss": 1.7263, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.1456088801828273, |
|
"grad_norm": 2.9057788848876953, |
|
"learning_rate": 4.5414656136510334e-05, |
|
"loss": 1.2964, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.14626183480248123, |
|
"grad_norm": 2.710711717605591, |
|
"learning_rate": 4.437846431455249e-05, |
|
"loss": 1.2502, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.14691478942213515, |
|
"grad_norm": 2.6256871223449707, |
|
"learning_rate": 4.335084871624572e-05, |
|
"loss": 0.9017, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.1475677440417891, |
|
"grad_norm": 3.905369997024536, |
|
"learning_rate": 4.2331967788513295e-05, |
|
"loss": 2.0128, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.14822069866144302, |
|
"grad_norm": 3.0856714248657227, |
|
"learning_rate": 4.132197863148925e-05, |
|
"loss": 1.7455, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.14887365328109697, |
|
"grad_norm": 3.04111385345459, |
|
"learning_rate": 4.0321036974295156e-05, |
|
"loss": 1.7097, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.14887365328109697, |
|
"eval_loss": NaN, |
|
"eval_runtime": 702.6357, |
|
"eval_samples_per_second": 3.672, |
|
"eval_steps_per_second": 0.918, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.1495266079007509, |
|
"grad_norm": 2.8979952335357666, |
|
"learning_rate": 3.932929715102863e-05, |
|
"loss": 1.4961, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.15017956252040482, |
|
"grad_norm": 2.8216731548309326, |
|
"learning_rate": 3.834691207696649e-05, |
|
"loss": 1.5215, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.15083251714005877, |
|
"grad_norm": 2.5655124187469482, |
|
"learning_rate": 3.7374033224987084e-05, |
|
"loss": 1.5517, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.1514854717597127, |
|
"grad_norm": 2.572943687438965, |
|
"learning_rate": 3.6410810602214684e-05, |
|
"loss": 1.2292, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.15213842637936664, |
|
"grad_norm": 2.6719319820404053, |
|
"learning_rate": 3.5457392726890236e-05, |
|
"loss": 1.1804, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.15279138099902057, |
|
"grad_norm": 2.3251240253448486, |
|
"learning_rate": 3.45139266054715e-05, |
|
"loss": 0.9235, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.15344433561867452, |
|
"grad_norm": 2.7189831733703613, |
|
"learning_rate": 3.3580557709966066e-05, |
|
"loss": 0.9769, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.15409729023832844, |
|
"grad_norm": 2.5098495483398438, |
|
"learning_rate": 3.2657429955501394e-05, |
|
"loss": 1.1983, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.15475024485798236, |
|
"grad_norm": 2.41629695892334, |
|
"learning_rate": 3.174468567813461e-05, |
|
"loss": 1.7956, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.1554031994776363, |
|
"grad_norm": 2.5007240772247314, |
|
"learning_rate": 3.0842465612905837e-05, |
|
"loss": 1.7452, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.15605615409729023, |
|
"grad_norm": 1.9491175413131714, |
|
"learning_rate": 2.9950908872138584e-05, |
|
"loss": 1.3075, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.15670910871694418, |
|
"grad_norm": 2.146371841430664, |
|
"learning_rate": 2.9070152923989946e-05, |
|
"loss": 1.7472, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.1573620633365981, |
|
"grad_norm": 2.1137588024139404, |
|
"learning_rate": 2.82003335712546e-05, |
|
"loss": 1.8301, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.15801501795625203, |
|
"grad_norm": 2.4170126914978027, |
|
"learning_rate": 2.7341584930425657e-05, |
|
"loss": 1.8519, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.15866797257590598, |
|
"grad_norm": 2.6382076740264893, |
|
"learning_rate": 2.6494039411015193e-05, |
|
"loss": 2.012, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.1593209271955599, |
|
"grad_norm": 2.1745855808258057, |
|
"learning_rate": 2.5657827695138372e-05, |
|
"loss": 1.7181, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.15997388181521385, |
|
"grad_norm": 2.9943010807037354, |
|
"learning_rate": 2.4833078717363544e-05, |
|
"loss": 1.5537, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.16062683643486778, |
|
"grad_norm": 3.608306407928467, |
|
"learning_rate": 2.4019919644832023e-05, |
|
"loss": 1.7291, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.1612797910545217, |
|
"grad_norm": 2.9223315715789795, |
|
"learning_rate": 2.3218475857650346e-05, |
|
"loss": 1.7161, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.16193274567417565, |
|
"grad_norm": 3.305222749710083, |
|
"learning_rate": 2.242887092955801e-05, |
|
"loss": 1.4356, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.16258570029382957, |
|
"grad_norm": 2.8326313495635986, |
|
"learning_rate": 2.1651226608873877e-05, |
|
"loss": 1.3055, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.16323865491348352, |
|
"grad_norm": 3.042375087738037, |
|
"learning_rate": 2.0885662799723804e-05, |
|
"loss": 0.8947, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.16389160953313744, |
|
"grad_norm": 2.6567325592041016, |
|
"learning_rate": 2.0132297543552757e-05, |
|
"loss": 1.776, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.16454456415279137, |
|
"grad_norm": 2.997938871383667, |
|
"learning_rate": 1.939124700092423e-05, |
|
"loss": 1.8796, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.16519751877244532, |
|
"grad_norm": 2.9053895473480225, |
|
"learning_rate": 1.866262543360958e-05, |
|
"loss": 1.662, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.16585047339209924, |
|
"grad_norm": 2.6591832637786865, |
|
"learning_rate": 1.7946545186970022e-05, |
|
"loss": 1.6189, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.1665034280117532, |
|
"grad_norm": 2.9311482906341553, |
|
"learning_rate": 1.7243116672634262e-05, |
|
"loss": 1.5663, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.1671563826314071, |
|
"grad_norm": 2.9493162631988525, |
|
"learning_rate": 1.6552448351474304e-05, |
|
"loss": 1.5125, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.16780933725106106, |
|
"grad_norm": 2.19980525970459, |
|
"learning_rate": 1.587464671688187e-05, |
|
"loss": 1.171, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.168462291870715, |
|
"grad_norm": 3.4264473915100098, |
|
"learning_rate": 1.520981627834851e-05, |
|
"loss": 1.0619, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.1691152464903689, |
|
"grad_norm": 2.755953550338745, |
|
"learning_rate": 1.4558059545351143e-05, |
|
"loss": 1.3709, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.16976820111002286, |
|
"grad_norm": 3.0671868324279785, |
|
"learning_rate": 1.3919477011546423e-05, |
|
"loss": 1.3373, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.17042115572967678, |
|
"grad_norm": 2.703378200531006, |
|
"learning_rate": 1.3294167139275593e-05, |
|
"loss": 1.4812, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.17107411034933073, |
|
"grad_norm": 2.7871146202087402, |
|
"learning_rate": 1.2682226344382796e-05, |
|
"loss": 1.5082, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.17172706496898466, |
|
"grad_norm": 2.1928627490997314, |
|
"learning_rate": 1.208374898134883e-05, |
|
"loss": 1.7859, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.17238001958863858, |
|
"grad_norm": 2.330217123031616, |
|
"learning_rate": 1.1498827328742623e-05, |
|
"loss": 1.9261, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.17303297420829253, |
|
"grad_norm": 2.532562732696533, |
|
"learning_rate": 1.0927551574992967e-05, |
|
"loss": 1.744, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.17368592882794645, |
|
"grad_norm": 2.453615188598633, |
|
"learning_rate": 1.0370009804482483e-05, |
|
"loss": 1.8084, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.1743388834476004, |
|
"grad_norm": 2.09717059135437, |
|
"learning_rate": 9.82628798396592e-06, |
|
"loss": 1.7786, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.17499183806725432, |
|
"grad_norm": 3.074509620666504, |
|
"learning_rate": 9.296469949315156e-06, |
|
"loss": 1.9725, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.17564479268690825, |
|
"grad_norm": 2.6540865898132324, |
|
"learning_rate": 8.780637392592495e-06, |
|
"loss": 1.6455, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.1762977473065622, |
|
"grad_norm": 3.0181193351745605, |
|
"learning_rate": 8.278869849454718e-06, |
|
"loss": 1.7383, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.17695070192621612, |
|
"grad_norm": 3.034897565841675, |
|
"learning_rate": 7.791244686889588e-06, |
|
"loss": 1.7912, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.17760365654587007, |
|
"grad_norm": 3.0905239582061768, |
|
"learning_rate": 7.317837091286706e-06, |
|
"loss": 1.5027, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.178256611165524, |
|
"grad_norm": 3.1407511234283447, |
|
"learning_rate": 6.858720056844614e-06, |
|
"loss": 1.2953, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.17890956578517794, |
|
"grad_norm": 2.7985000610351562, |
|
"learning_rate": 6.413964374315851e-06, |
|
"loss": 1.4958, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.17956252040483187, |
|
"grad_norm": 2.70698881149292, |
|
"learning_rate": 5.983638620091858e-06, |
|
"loss": 1.1345, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.1802154750244858, |
|
"grad_norm": 3.302675485610962, |
|
"learning_rate": 5.567809145629244e-06, |
|
"loss": 1.8954, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.18086842964413974, |
|
"grad_norm": 2.7587976455688477, |
|
"learning_rate": 5.1665400672190725e-06, |
|
"loss": 1.6744, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.18152138426379366, |
|
"grad_norm": 2.4605860710144043, |
|
"learning_rate": 4.7798932561009865e-06, |
|
"loss": 1.4846, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.1821743388834476, |
|
"grad_norm": 3.0369081497192383, |
|
"learning_rate": 4.407928328923194e-06, |
|
"loss": 1.5561, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.18282729350310153, |
|
"grad_norm": 2.6894164085388184, |
|
"learning_rate": 4.050702638550275e-06, |
|
"loss": 1.1031, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.18348024812275546, |
|
"grad_norm": 2.849666118621826, |
|
"learning_rate": 3.7082712652200867e-06, |
|
"loss": 1.366, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.1841332027424094, |
|
"grad_norm": 2.834463119506836, |
|
"learning_rate": 3.380687008050909e-06, |
|
"loss": 1.1118, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.18478615736206333, |
|
"grad_norm": 2.5449914932250977, |
|
"learning_rate": 3.068000376900515e-06, |
|
"loss": 1.0414, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.18543911198171728, |
|
"grad_norm": 2.6311261653900146, |
|
"learning_rate": 2.770259584577972e-06, |
|
"loss": 0.9975, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.1860920666013712, |
|
"grad_norm": 2.644763946533203, |
|
"learning_rate": 2.4875105394098654e-06, |
|
"loss": 1.1698, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.18674502122102513, |
|
"grad_norm": 2.4958488941192627, |
|
"learning_rate": 2.219796838161681e-06, |
|
"loss": 1.6067, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.18739797584067908, |
|
"grad_norm": 2.388773202896118, |
|
"learning_rate": 1.967159759315751e-06, |
|
"loss": 1.9297, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.188050930460333, |
|
"grad_norm": 2.2818784713745117, |
|
"learning_rate": 1.7296382567064672e-06, |
|
"loss": 1.7271, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.18870388507998695, |
|
"grad_norm": 2.581960439682007, |
|
"learning_rate": 1.5072689535141072e-06, |
|
"loss": 1.7268, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.18935683969964087, |
|
"grad_norm": 2.5180134773254395, |
|
"learning_rate": 1.3000861366179062e-06, |
|
"loss": 1.9191, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.19000979431929482, |
|
"grad_norm": 2.125568151473999, |
|
"learning_rate": 1.1081217513094212e-06, |
|
"loss": 1.7848, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.19066274893894875, |
|
"grad_norm": 2.7064099311828613, |
|
"learning_rate": 9.314053963669245e-07, |
|
"loss": 1.6782, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.19131570355860267, |
|
"grad_norm": 2.437058448791504, |
|
"learning_rate": 7.699643194915784e-07, |
|
"loss": 1.8157, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.19196865817825662, |
|
"grad_norm": 3.013352870941162, |
|
"learning_rate": 6.238234131061616e-07, |
|
"loss": 1.3813, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.19262161279791054, |
|
"grad_norm": 2.7739081382751465, |
|
"learning_rate": 4.93005210516928e-07, |
|
"loss": 1.7656, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.1932745674175645, |
|
"grad_norm": 3.043750286102295, |
|
"learning_rate": 3.775298824391982e-07, |
|
"loss": 1.5386, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.1939275220372184, |
|
"grad_norm": 2.856957197189331, |
|
"learning_rate": 2.774152338873126e-07, |
|
"loss": 1.4946, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.19458047665687234, |
|
"grad_norm": 2.9437901973724365, |
|
"learning_rate": 1.9267670142926187e-07, |
|
"loss": 1.3205, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.1952334312765263, |
|
"grad_norm": 2.6561126708984375, |
|
"learning_rate": 1.2332735080651248e-07, |
|
"loss": 1.1922, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.1958863858961802, |
|
"grad_norm": 2.5239169597625732, |
|
"learning_rate": 6.9377874919474e-08, |
|
"loss": 0.956, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.19653934051583416, |
|
"grad_norm": 2.8316755294799805, |
|
"learning_rate": 3.0836592178717926e-08, |
|
"loss": 1.8631, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.19719229513548808, |
|
"grad_norm": 2.5579283237457275, |
|
"learning_rate": 7.709445222403577e-09, |
|
"loss": 1.4522, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.197845249755142, |
|
"grad_norm": 2.6677441596984863, |
|
"learning_rate": 0.0, |
|
"loss": 1.7175, |
|
"step": 303 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 303, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 76, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 9.294759255217275e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|