|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 151.8324607329843, |
|
"eval_steps": 500, |
|
"global_step": 87000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.8726003490401396, |
|
"grad_norm": 0.931673526763916, |
|
"learning_rate": 0.0009956369982547994, |
|
"loss": 2.6507, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.7452006980802792, |
|
"grad_norm": 1.8975260257720947, |
|
"learning_rate": 0.0009912739965095986, |
|
"loss": 1.6969, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.6178010471204187, |
|
"grad_norm": 0.5923702120780945, |
|
"learning_rate": 0.000986910994764398, |
|
"loss": 1.287, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 3.4904013961605584, |
|
"grad_norm": 0.79058438539505, |
|
"learning_rate": 0.000982547993019197, |
|
"loss": 0.9398, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 4.363001745200698, |
|
"grad_norm": 0.7330807447433472, |
|
"learning_rate": 0.0009781849912739965, |
|
"loss": 0.6968, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 5.2356020942408374, |
|
"grad_norm": 0.6996338963508606, |
|
"learning_rate": 0.0009738219895287959, |
|
"loss": 0.4816, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 6.108202443280978, |
|
"grad_norm": 0.5383062362670898, |
|
"learning_rate": 0.0009694589877835951, |
|
"loss": 0.3189, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 6.980802792321117, |
|
"grad_norm": 0.6486771106719971, |
|
"learning_rate": 0.0009650959860383944, |
|
"loss": 0.2218, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 7.853403141361256, |
|
"grad_norm": 0.507366955280304, |
|
"learning_rate": 0.0009607329842931938, |
|
"loss": 0.1542, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 8.726003490401396, |
|
"grad_norm": 0.8452386260032654, |
|
"learning_rate": 0.000956369982547993, |
|
"loss": 0.1285, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 9.598603839441536, |
|
"grad_norm": 0.43142977356910706, |
|
"learning_rate": 0.0009520069808027923, |
|
"loss": 0.1175, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 10.471204188481675, |
|
"grad_norm": 0.2331060916185379, |
|
"learning_rate": 0.0009476439790575916, |
|
"loss": 0.1053, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 11.343804537521814, |
|
"grad_norm": 0.4272315800189972, |
|
"learning_rate": 0.000943280977312391, |
|
"loss": 0.092, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 12.216404886561955, |
|
"grad_norm": 0.2999955117702484, |
|
"learning_rate": 0.0009389179755671902, |
|
"loss": 0.0878, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 13.089005235602095, |
|
"grad_norm": 2.55757474899292, |
|
"learning_rate": 0.0009345549738219895, |
|
"loss": 0.0884, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 13.961605584642234, |
|
"grad_norm": 0.33913654088974, |
|
"learning_rate": 0.0009301919720767889, |
|
"loss": 0.0805, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 14.834205933682373, |
|
"grad_norm": 0.3642922341823578, |
|
"learning_rate": 0.0009258289703315882, |
|
"loss": 0.0721, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 15.706806282722512, |
|
"grad_norm": 0.7718423008918762, |
|
"learning_rate": 0.0009214659685863874, |
|
"loss": 0.0687, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 16.57940663176265, |
|
"grad_norm": 0.5820666551589966, |
|
"learning_rate": 0.0009171029668411868, |
|
"loss": 0.0644, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 17.452006980802793, |
|
"grad_norm": 0.2773011028766632, |
|
"learning_rate": 0.000912739965095986, |
|
"loss": 0.0598, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 18.324607329842934, |
|
"grad_norm": 0.25250518321990967, |
|
"learning_rate": 0.0009083769633507853, |
|
"loss": 0.0653, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 19.19720767888307, |
|
"grad_norm": 0.24120768904685974, |
|
"learning_rate": 0.0009040139616055847, |
|
"loss": 0.0632, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 20.069808027923212, |
|
"grad_norm": 0.2897244989871979, |
|
"learning_rate": 0.0008996509598603839, |
|
"loss": 0.0594, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 20.94240837696335, |
|
"grad_norm": 0.25145065784454346, |
|
"learning_rate": 0.0008952879581151833, |
|
"loss": 0.0556, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 21.81500872600349, |
|
"grad_norm": 0.27175939083099365, |
|
"learning_rate": 0.0008909249563699826, |
|
"loss": 0.0481, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 22.68760907504363, |
|
"grad_norm": 0.8626015782356262, |
|
"learning_rate": 0.0008865619546247818, |
|
"loss": 0.0466, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 23.56020942408377, |
|
"grad_norm": 0.18672889471054077, |
|
"learning_rate": 0.0008821989528795812, |
|
"loss": 0.0512, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 24.43280977312391, |
|
"grad_norm": 0.2387542873620987, |
|
"learning_rate": 0.0008778359511343804, |
|
"loss": 0.0512, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 25.305410122164048, |
|
"grad_norm": 0.2665075957775116, |
|
"learning_rate": 0.0008734729493891797, |
|
"loss": 0.0483, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 26.17801047120419, |
|
"grad_norm": 0.16715960204601288, |
|
"learning_rate": 0.0008691099476439791, |
|
"loss": 0.0441, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 27.050610820244327, |
|
"grad_norm": 0.1875993311405182, |
|
"learning_rate": 0.0008647469458987784, |
|
"loss": 0.0397, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 27.923211169284468, |
|
"grad_norm": 0.25451162457466125, |
|
"learning_rate": 0.0008603839441535776, |
|
"loss": 0.039, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 28.79581151832461, |
|
"grad_norm": 0.307054728269577, |
|
"learning_rate": 0.000856020942408377, |
|
"loss": 0.0443, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 29.668411867364746, |
|
"grad_norm": 0.2467741221189499, |
|
"learning_rate": 0.0008516579406631763, |
|
"loss": 0.038, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 30.541012216404887, |
|
"grad_norm": 0.23178012669086456, |
|
"learning_rate": 0.0008472949389179755, |
|
"loss": 0.0351, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 31.413612565445025, |
|
"grad_norm": 0.35061487555503845, |
|
"learning_rate": 0.0008429319371727748, |
|
"loss": 0.0323, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 32.28621291448517, |
|
"grad_norm": 0.3363337218761444, |
|
"learning_rate": 0.0008385689354275742, |
|
"loss": 0.038, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 33.1588132635253, |
|
"grad_norm": 0.24015972018241882, |
|
"learning_rate": 0.0008342059336823735, |
|
"loss": 0.0361, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 34.031413612565444, |
|
"grad_norm": 0.5147364735603333, |
|
"learning_rate": 0.0008298429319371727, |
|
"loss": 0.0318, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 34.904013961605585, |
|
"grad_norm": 0.17477251589298248, |
|
"learning_rate": 0.0008254799301919721, |
|
"loss": 0.0307, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 35.776614310645726, |
|
"grad_norm": 0.2511900067329407, |
|
"learning_rate": 0.0008211169284467714, |
|
"loss": 0.0348, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 36.64921465968587, |
|
"grad_norm": 0.2036200314760208, |
|
"learning_rate": 0.0008167539267015707, |
|
"loss": 0.0298, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 37.521815008726, |
|
"grad_norm": 0.16981545090675354, |
|
"learning_rate": 0.00081239092495637, |
|
"loss": 0.0319, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 38.39441535776614, |
|
"grad_norm": 0.22329097986221313, |
|
"learning_rate": 0.0008080279232111692, |
|
"loss": 0.0359, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 39.26701570680628, |
|
"grad_norm": 0.13258185982704163, |
|
"learning_rate": 0.0008036649214659686, |
|
"loss": 0.0266, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 40.139616055846425, |
|
"grad_norm": 0.12790647149085999, |
|
"learning_rate": 0.000799301919720768, |
|
"loss": 0.026, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 41.01221640488656, |
|
"grad_norm": 0.23635344207286835, |
|
"learning_rate": 0.0007949389179755671, |
|
"loss": 0.0279, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 41.8848167539267, |
|
"grad_norm": 0.11364254355430603, |
|
"learning_rate": 0.0007905759162303665, |
|
"loss": 0.0257, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 42.75741710296684, |
|
"grad_norm": 0.2781168520450592, |
|
"learning_rate": 0.0007862129144851659, |
|
"loss": 0.0295, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 43.63001745200698, |
|
"grad_norm": 0.106789730489254, |
|
"learning_rate": 0.0007818499127399651, |
|
"loss": 0.0308, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 44.50261780104712, |
|
"grad_norm": 0.16404911875724792, |
|
"learning_rate": 0.0007774869109947644, |
|
"loss": 0.0222, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 45.37521815008726, |
|
"grad_norm": 0.14249293506145477, |
|
"learning_rate": 0.0007731239092495637, |
|
"loss": 0.0225, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 46.2478184991274, |
|
"grad_norm": 0.1853444129228592, |
|
"learning_rate": 0.0007687609075043631, |
|
"loss": 0.0261, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 47.12041884816754, |
|
"grad_norm": 0.1456003040075302, |
|
"learning_rate": 0.0007643979057591623, |
|
"loss": 0.0252, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 47.99301919720768, |
|
"grad_norm": 0.16386698186397552, |
|
"learning_rate": 0.0007600349040139616, |
|
"loss": 0.0259, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 48.86561954624782, |
|
"grad_norm": 0.12221992015838623, |
|
"learning_rate": 0.000755671902268761, |
|
"loss": 0.0253, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 49.738219895287955, |
|
"grad_norm": 0.14093224704265594, |
|
"learning_rate": 0.0007513089005235602, |
|
"loss": 0.0203, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 50.610820244328096, |
|
"grad_norm": 0.1189383938908577, |
|
"learning_rate": 0.0007469458987783595, |
|
"loss": 0.0207, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 51.48342059336824, |
|
"grad_norm": 0.1471104919910431, |
|
"learning_rate": 0.0007425828970331589, |
|
"loss": 0.0209, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 52.35602094240838, |
|
"grad_norm": 0.08947575837373734, |
|
"learning_rate": 0.0007382198952879581, |
|
"loss": 0.0234, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 53.22862129144852, |
|
"grad_norm": 0.18746259808540344, |
|
"learning_rate": 0.0007338568935427574, |
|
"loss": 0.0245, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 54.10122164048865, |
|
"grad_norm": 0.1539311408996582, |
|
"learning_rate": 0.0007294938917975568, |
|
"loss": 0.0214, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 54.973821989528794, |
|
"grad_norm": 0.11201947182416916, |
|
"learning_rate": 0.000725130890052356, |
|
"loss": 0.0194, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 55.846422338568935, |
|
"grad_norm": 0.16618479788303375, |
|
"learning_rate": 0.0007207678883071554, |
|
"loss": 0.0185, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 56.719022687609076, |
|
"grad_norm": 0.1569599211215973, |
|
"learning_rate": 0.0007164048865619547, |
|
"loss": 0.0234, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 57.59162303664922, |
|
"grad_norm": 0.11062045395374298, |
|
"learning_rate": 0.0007120418848167539, |
|
"loss": 0.0187, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 58.46422338568935, |
|
"grad_norm": 0.1617700457572937, |
|
"learning_rate": 0.0007076788830715533, |
|
"loss": 0.0176, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 59.33682373472949, |
|
"grad_norm": 0.11750755459070206, |
|
"learning_rate": 0.0007033158813263525, |
|
"loss": 0.0188, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 60.20942408376963, |
|
"grad_norm": 0.24937282502651215, |
|
"learning_rate": 0.0006989528795811518, |
|
"loss": 0.0225, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 61.082024432809774, |
|
"grad_norm": 0.22410957515239716, |
|
"learning_rate": 0.0006945898778359512, |
|
"loss": 0.0192, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 61.954624781849915, |
|
"grad_norm": 0.18029357492923737, |
|
"learning_rate": 0.0006902268760907505, |
|
"loss": 0.0169, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 62.82722513089005, |
|
"grad_norm": 0.14077898859977722, |
|
"learning_rate": 0.0006858638743455497, |
|
"loss": 0.0167, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 63.69982547993019, |
|
"grad_norm": 0.12804169952869415, |
|
"learning_rate": 0.0006815008726003491, |
|
"loss": 0.0167, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 64.57242582897034, |
|
"grad_norm": 0.11356078833341599, |
|
"learning_rate": 0.0006771378708551484, |
|
"loss": 0.0183, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 65.44502617801047, |
|
"grad_norm": 0.07214757055044174, |
|
"learning_rate": 0.0006727748691099476, |
|
"loss": 0.0167, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 66.3176265270506, |
|
"grad_norm": 0.20497408509254456, |
|
"learning_rate": 0.0006684118673647469, |
|
"loss": 0.0186, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 67.19022687609075, |
|
"grad_norm": 0.09408937394618988, |
|
"learning_rate": 0.0006640488656195463, |
|
"loss": 0.0169, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 68.06282722513089, |
|
"grad_norm": 0.12423662841320038, |
|
"learning_rate": 0.0006596858638743456, |
|
"loss": 0.0179, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 68.93542757417103, |
|
"grad_norm": 0.4406953752040863, |
|
"learning_rate": 0.0006553228621291448, |
|
"loss": 0.019, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 69.80802792321117, |
|
"grad_norm": 0.11233725398778915, |
|
"learning_rate": 0.0006509598603839442, |
|
"loss": 0.0151, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 70.68062827225131, |
|
"grad_norm": 0.08892803639173508, |
|
"learning_rate": 0.0006465968586387435, |
|
"loss": 0.0137, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 71.55322862129145, |
|
"grad_norm": 0.11712398380041122, |
|
"learning_rate": 0.0006422338568935428, |
|
"loss": 0.0145, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 72.4258289703316, |
|
"grad_norm": 0.11560557782649994, |
|
"learning_rate": 0.0006378708551483421, |
|
"loss": 0.0181, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 73.29842931937172, |
|
"grad_norm": 0.10780952870845795, |
|
"learning_rate": 0.0006335078534031413, |
|
"loss": 0.0162, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 74.17102966841186, |
|
"grad_norm": 0.06540343165397644, |
|
"learning_rate": 0.0006291448516579407, |
|
"loss": 0.0157, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 75.043630017452, |
|
"grad_norm": 0.08087161183357239, |
|
"learning_rate": 0.00062478184991274, |
|
"loss": 0.0181, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 75.91623036649214, |
|
"grad_norm": 0.08909798413515091, |
|
"learning_rate": 0.0006204188481675392, |
|
"loss": 0.0136, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 76.78883071553228, |
|
"grad_norm": 0.09045905619859695, |
|
"learning_rate": 0.0006160558464223386, |
|
"loss": 0.0124, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 77.66143106457243, |
|
"grad_norm": 0.1375885307788849, |
|
"learning_rate": 0.000611692844677138, |
|
"loss": 0.013, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 78.53403141361257, |
|
"grad_norm": 0.06584478169679642, |
|
"learning_rate": 0.0006073298429319371, |
|
"loss": 0.0149, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 79.40663176265271, |
|
"grad_norm": 0.10484465211629868, |
|
"learning_rate": 0.0006029668411867365, |
|
"loss": 0.0141, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 80.27923211169285, |
|
"grad_norm": 0.11171326786279678, |
|
"learning_rate": 0.0005986038394415358, |
|
"loss": 0.0142, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 81.15183246073299, |
|
"grad_norm": 0.12636296451091766, |
|
"learning_rate": 0.000594240837696335, |
|
"loss": 0.0154, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 82.02443280977312, |
|
"grad_norm": 0.06967689841985703, |
|
"learning_rate": 0.0005898778359511344, |
|
"loss": 0.0144, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 82.89703315881326, |
|
"grad_norm": 0.09147424250841141, |
|
"learning_rate": 0.0005855148342059337, |
|
"loss": 0.0127, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 83.7696335078534, |
|
"grad_norm": 0.6871227025985718, |
|
"learning_rate": 0.000581151832460733, |
|
"loss": 0.0113, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 84.64223385689354, |
|
"grad_norm": 0.07160250097513199, |
|
"learning_rate": 0.0005767888307155323, |
|
"loss": 0.014, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 85.51483420593368, |
|
"grad_norm": 0.06839723885059357, |
|
"learning_rate": 0.0005724258289703316, |
|
"loss": 0.0121, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 86.38743455497382, |
|
"grad_norm": 0.05615156516432762, |
|
"learning_rate": 0.0005680628272251309, |
|
"loss": 0.0136, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 87.26003490401396, |
|
"grad_norm": 0.10079669952392578, |
|
"learning_rate": 0.0005636998254799302, |
|
"loss": 0.0116, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 88.1326352530541, |
|
"grad_norm": 0.09019599854946136, |
|
"learning_rate": 0.0005593368237347295, |
|
"loss": 0.0122, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 89.00523560209425, |
|
"grad_norm": 0.05459260195493698, |
|
"learning_rate": 0.0005549738219895288, |
|
"loss": 0.0113, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 89.87783595113439, |
|
"grad_norm": 0.07002388685941696, |
|
"learning_rate": 0.0005506108202443281, |
|
"loss": 0.0112, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 90.75043630017451, |
|
"grad_norm": 0.12079566717147827, |
|
"learning_rate": 0.0005462478184991274, |
|
"loss": 0.0115, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 91.62303664921465, |
|
"grad_norm": 0.08160880208015442, |
|
"learning_rate": 0.0005418848167539267, |
|
"loss": 0.0135, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 92.4956369982548, |
|
"grad_norm": 0.053758975118398666, |
|
"learning_rate": 0.000537521815008726, |
|
"loss": 0.0111, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 93.36823734729494, |
|
"grad_norm": 0.08977492153644562, |
|
"learning_rate": 0.0005331588132635254, |
|
"loss": 0.0118, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 94.24083769633508, |
|
"grad_norm": 0.14162831008434296, |
|
"learning_rate": 0.0005287958115183245, |
|
"loss": 0.0118, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 95.11343804537522, |
|
"grad_norm": 0.09927529096603394, |
|
"learning_rate": 0.0005244328097731239, |
|
"loss": 0.011, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 95.98603839441536, |
|
"grad_norm": 0.08714314550161362, |
|
"learning_rate": 0.0005200698080279233, |
|
"loss": 0.0101, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 96.8586387434555, |
|
"grad_norm": 0.09934234619140625, |
|
"learning_rate": 0.0005157068062827224, |
|
"loss": 0.0109, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 97.73123909249564, |
|
"grad_norm": 0.11947502195835114, |
|
"learning_rate": 0.0005113438045375218, |
|
"loss": 0.0114, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 98.60383944153578, |
|
"grad_norm": 0.0667385384440422, |
|
"learning_rate": 0.0005069808027923212, |
|
"loss": 0.0099, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 99.47643979057591, |
|
"grad_norm": 0.11594051122665405, |
|
"learning_rate": 0.0005026178010471204, |
|
"loss": 0.0117, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 100.34904013961605, |
|
"grad_norm": 0.06558683514595032, |
|
"learning_rate": 0.0004982547993019197, |
|
"loss": 0.0104, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 101.22164048865619, |
|
"grad_norm": 0.07644475996494293, |
|
"learning_rate": 0.000493891797556719, |
|
"loss": 0.0098, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 102.09424083769633, |
|
"grad_norm": 0.06626095622777939, |
|
"learning_rate": 0.0004895287958115183, |
|
"loss": 0.0095, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 102.96684118673647, |
|
"grad_norm": 0.10808754712343216, |
|
"learning_rate": 0.0004851657940663176, |
|
"loss": 0.0108, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 103.83944153577661, |
|
"grad_norm": 0.05127561092376709, |
|
"learning_rate": 0.000480802792321117, |
|
"loss": 0.01, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 104.71204188481676, |
|
"grad_norm": 0.12128196656703949, |
|
"learning_rate": 0.00047643979057591625, |
|
"loss": 0.0092, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 105.5846422338569, |
|
"grad_norm": 0.06882330775260925, |
|
"learning_rate": 0.0004720767888307155, |
|
"loss": 0.0111, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 106.45724258289704, |
|
"grad_norm": 0.05268734693527222, |
|
"learning_rate": 0.00046771378708551485, |
|
"loss": 0.0094, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 107.32984293193718, |
|
"grad_norm": 0.046268824487924576, |
|
"learning_rate": 0.00046335078534031417, |
|
"loss": 0.0082, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 108.2024432809773, |
|
"grad_norm": 0.04883289709687233, |
|
"learning_rate": 0.00045898778359511344, |
|
"loss": 0.0085, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 109.07504363001745, |
|
"grad_norm": 0.0723048597574234, |
|
"learning_rate": 0.00045462478184991276, |
|
"loss": 0.0093, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 109.94764397905759, |
|
"grad_norm": 0.06026541814208031, |
|
"learning_rate": 0.00045026178010471203, |
|
"loss": 0.0095, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 110.82024432809773, |
|
"grad_norm": 0.058908674865961075, |
|
"learning_rate": 0.00044589877835951136, |
|
"loss": 0.0085, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 111.69284467713787, |
|
"grad_norm": 0.05758107081055641, |
|
"learning_rate": 0.0004415357766143107, |
|
"loss": 0.01, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 112.56544502617801, |
|
"grad_norm": 0.06559444963932037, |
|
"learning_rate": 0.00043717277486910995, |
|
"loss": 0.0088, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 113.43804537521815, |
|
"grad_norm": 0.05080035701394081, |
|
"learning_rate": 0.0004328097731239092, |
|
"loss": 0.0083, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 114.3106457242583, |
|
"grad_norm": 0.0523524135351181, |
|
"learning_rate": 0.0004284467713787086, |
|
"loss": 0.0078, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 115.18324607329843, |
|
"grad_norm": 0.14169646799564362, |
|
"learning_rate": 0.00042408376963350787, |
|
"loss": 0.0085, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 116.05584642233858, |
|
"grad_norm": 0.05305915325880051, |
|
"learning_rate": 0.00041972076788830714, |
|
"loss": 0.0091, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 116.9284467713787, |
|
"grad_norm": 0.05915080010890961, |
|
"learning_rate": 0.00041535776614310646, |
|
"loss": 0.0085, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 117.80104712041884, |
|
"grad_norm": 0.07950141280889511, |
|
"learning_rate": 0.0004109947643979058, |
|
"loss": 0.0075, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 118.67364746945898, |
|
"grad_norm": 0.083484947681427, |
|
"learning_rate": 0.00040663176265270506, |
|
"loss": 0.0086, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 119.54624781849913, |
|
"grad_norm": 0.1149265244603157, |
|
"learning_rate": 0.0004022687609075044, |
|
"loss": 0.0087, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 120.41884816753927, |
|
"grad_norm": 0.10079418867826462, |
|
"learning_rate": 0.00039790575916230365, |
|
"loss": 0.0087, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 121.29144851657941, |
|
"grad_norm": 0.04444717988371849, |
|
"learning_rate": 0.000393542757417103, |
|
"loss": 0.0083, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 122.16404886561955, |
|
"grad_norm": 0.015783872455358505, |
|
"learning_rate": 0.0003891797556719023, |
|
"loss": 0.0074, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 123.03664921465969, |
|
"grad_norm": 0.07311473041772842, |
|
"learning_rate": 0.00038481675392670157, |
|
"loss": 0.007, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 123.90924956369983, |
|
"grad_norm": 0.03907659277319908, |
|
"learning_rate": 0.00038045375218150084, |
|
"loss": 0.0071, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 124.78184991273997, |
|
"grad_norm": 0.031013870611786842, |
|
"learning_rate": 0.0003760907504363002, |
|
"loss": 0.0071, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 125.6544502617801, |
|
"grad_norm": 0.05435263365507126, |
|
"learning_rate": 0.0003717277486910995, |
|
"loss": 0.0077, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 126.52705061082024, |
|
"grad_norm": 0.015180529095232487, |
|
"learning_rate": 0.00036736474694589876, |
|
"loss": 0.0072, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 127.39965095986038, |
|
"grad_norm": 0.02177223190665245, |
|
"learning_rate": 0.0003630017452006981, |
|
"loss": 0.0074, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 128.27225130890054, |
|
"grad_norm": 0.04897777736186981, |
|
"learning_rate": 0.0003586387434554974, |
|
"loss": 0.0076, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 129.14485165794068, |
|
"grad_norm": 0.02429993264377117, |
|
"learning_rate": 0.0003542757417102967, |
|
"loss": 0.0067, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 130.0174520069808, |
|
"grad_norm": 0.019401997327804565, |
|
"learning_rate": 0.000349912739965096, |
|
"loss": 0.0071, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 130.89005235602093, |
|
"grad_norm": 0.05895571410655975, |
|
"learning_rate": 0.00034554973821989527, |
|
"loss": 0.0067, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 131.76265270506107, |
|
"grad_norm": 0.08564560860395432, |
|
"learning_rate": 0.0003411867364746946, |
|
"loss": 0.0066, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 132.6352530541012, |
|
"grad_norm": 0.0574815534055233, |
|
"learning_rate": 0.0003368237347294939, |
|
"loss": 0.0063, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 133.50785340314135, |
|
"grad_norm": 0.08703868091106415, |
|
"learning_rate": 0.0003324607329842932, |
|
"loss": 0.0064, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 134.3804537521815, |
|
"grad_norm": 0.1667858362197876, |
|
"learning_rate": 0.00032809773123909246, |
|
"loss": 0.0074, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 135.25305410122164, |
|
"grad_norm": 0.07352814823389053, |
|
"learning_rate": 0.00032373472949389184, |
|
"loss": 0.0067, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 136.12565445026178, |
|
"grad_norm": 0.10321661829948425, |
|
"learning_rate": 0.0003193717277486911, |
|
"loss": 0.0065, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 136.99825479930192, |
|
"grad_norm": 0.046309106051921844, |
|
"learning_rate": 0.0003150087260034904, |
|
"loss": 0.0065, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 137.87085514834206, |
|
"grad_norm": 0.014806479215621948, |
|
"learning_rate": 0.0003106457242582897, |
|
"loss": 0.0061, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 138.7434554973822, |
|
"grad_norm": 0.041382819414138794, |
|
"learning_rate": 0.000306282722513089, |
|
"loss": 0.0059, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 139.61605584642234, |
|
"grad_norm": 0.07493896782398224, |
|
"learning_rate": 0.0003019197207678883, |
|
"loss": 0.0061, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 140.48865619546248, |
|
"grad_norm": 0.052245959639549255, |
|
"learning_rate": 0.0002975567190226876, |
|
"loss": 0.006, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 141.36125654450262, |
|
"grad_norm": 0.10472971946001053, |
|
"learning_rate": 0.0002931937172774869, |
|
"loss": 0.0061, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 142.23385689354276, |
|
"grad_norm": 0.033921029418706894, |
|
"learning_rate": 0.0002888307155322862, |
|
"loss": 0.0059, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 143.1064572425829, |
|
"grad_norm": 0.06276967376470566, |
|
"learning_rate": 0.00028446771378708553, |
|
"loss": 0.0059, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 143.97905759162305, |
|
"grad_norm": 0.022356705740094185, |
|
"learning_rate": 0.0002801047120418848, |
|
"loss": 0.0056, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 144.8516579406632, |
|
"grad_norm": 0.008057367987930775, |
|
"learning_rate": 0.0002757417102966841, |
|
"loss": 0.0054, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 145.72425828970333, |
|
"grad_norm": 0.00805575679987669, |
|
"learning_rate": 0.00027137870855148345, |
|
"loss": 0.0054, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 146.59685863874344, |
|
"grad_norm": 0.07681386172771454, |
|
"learning_rate": 0.0002670157068062827, |
|
"loss": 0.0053, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 147.46945898778358, |
|
"grad_norm": 0.041651804000139236, |
|
"learning_rate": 0.00026265270506108205, |
|
"loss": 0.0054, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 148.34205933682372, |
|
"grad_norm": 0.09435189515352249, |
|
"learning_rate": 0.0002582897033158813, |
|
"loss": 0.0057, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 149.21465968586386, |
|
"grad_norm": 0.02968364767730236, |
|
"learning_rate": 0.00025392670157068064, |
|
"loss": 0.0053, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 150.087260034904, |
|
"grad_norm": 0.04307426139712334, |
|
"learning_rate": 0.00024956369982547996, |
|
"loss": 0.0053, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 150.95986038394415, |
|
"grad_norm": 0.08466316014528275, |
|
"learning_rate": 0.00024520069808027923, |
|
"loss": 0.0051, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 151.8324607329843, |
|
"grad_norm": 0.14644889533519745, |
|
"learning_rate": 0.00024083769633507853, |
|
"loss": 0.005, |
|
"step": 87000 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 114600, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 200, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.5133036050382848e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|