ardaspear's picture
Training in progress, step 272, checkpoint
5dafa23 verified
raw
history blame
18.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.07733864088711971,
"eval_steps": 34,
"global_step": 272,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0002843332385555871,
"eval_loss": 1.5018354654312134,
"eval_runtime": 166.295,
"eval_samples_per_second": 35.623,
"eval_steps_per_second": 4.456,
"step": 1
},
{
"epoch": 0.0008529997156667614,
"grad_norm": 0.7210565209388733,
"learning_rate": 1.5e-05,
"loss": 1.5223,
"step": 3
},
{
"epoch": 0.0017059994313335229,
"grad_norm": 0.6832892298698425,
"learning_rate": 3e-05,
"loss": 1.5221,
"step": 6
},
{
"epoch": 0.0025589991470002845,
"grad_norm": 0.7228020429611206,
"learning_rate": 4.5e-05,
"loss": 1.408,
"step": 9
},
{
"epoch": 0.0034119988626670457,
"grad_norm": 0.6384875774383545,
"learning_rate": 4.999675562428437e-05,
"loss": 1.4404,
"step": 12
},
{
"epoch": 0.004264998578333807,
"grad_norm": 0.3351251482963562,
"learning_rate": 4.9979724954289244e-05,
"loss": 1.3736,
"step": 15
},
{
"epoch": 0.005117998294000569,
"grad_norm": 0.182565838098526,
"learning_rate": 4.994810682835951e-05,
"loss": 1.3079,
"step": 18
},
{
"epoch": 0.00597099800966733,
"grad_norm": 0.19113591313362122,
"learning_rate": 4.990191971059033e-05,
"loss": 1.316,
"step": 21
},
{
"epoch": 0.0068239977253340914,
"grad_norm": 0.1948336660861969,
"learning_rate": 4.984119057295783e-05,
"loss": 1.2426,
"step": 24
},
{
"epoch": 0.007676997441000853,
"grad_norm": 0.18684643507003784,
"learning_rate": 4.976595487956823e-05,
"loss": 1.2503,
"step": 27
},
{
"epoch": 0.008529997156667614,
"grad_norm": 0.18414735794067383,
"learning_rate": 4.967625656594782e-05,
"loss": 1.2323,
"step": 30
},
{
"epoch": 0.009382996872334376,
"grad_norm": 0.17294025421142578,
"learning_rate": 4.957214801338581e-05,
"loss": 1.1942,
"step": 33
},
{
"epoch": 0.009667330110889964,
"eval_loss": 1.1813108921051025,
"eval_runtime": 167.7438,
"eval_samples_per_second": 35.316,
"eval_steps_per_second": 4.417,
"step": 34
},
{
"epoch": 0.010235996588001138,
"grad_norm": 0.16702412068843842,
"learning_rate": 4.9453690018345144e-05,
"loss": 1.1981,
"step": 36
},
{
"epoch": 0.011088996303667898,
"grad_norm": 0.1968175172805786,
"learning_rate": 4.932095175695911e-05,
"loss": 1.1675,
"step": 39
},
{
"epoch": 0.01194199601933466,
"grad_norm": 0.18244469165802002,
"learning_rate": 4.917401074463441e-05,
"loss": 1.1584,
"step": 42
},
{
"epoch": 0.01279499573500142,
"grad_norm": 0.16749081015586853,
"learning_rate": 4.901295279078431e-05,
"loss": 1.1134,
"step": 45
},
{
"epoch": 0.013647995450668183,
"grad_norm": 0.17398597300052643,
"learning_rate": 4.883787194871841e-05,
"loss": 1.1139,
"step": 48
},
{
"epoch": 0.014500995166334945,
"grad_norm": 0.17164087295532227,
"learning_rate": 4.864887046071813e-05,
"loss": 1.079,
"step": 51
},
{
"epoch": 0.015353994882001705,
"grad_norm": 0.1644001007080078,
"learning_rate": 4.8446058698330115e-05,
"loss": 1.0646,
"step": 54
},
{
"epoch": 0.01620699459766847,
"grad_norm": 0.16490623354911804,
"learning_rate": 4.822955509791233e-05,
"loss": 1.0739,
"step": 57
},
{
"epoch": 0.017059994313335228,
"grad_norm": 0.17708458006381989,
"learning_rate": 4.799948609147061e-05,
"loss": 1.0897,
"step": 60
},
{
"epoch": 0.01791299402900199,
"grad_norm": 0.15597032010555267,
"learning_rate": 4.7755986032825864e-05,
"loss": 1.0566,
"step": 63
},
{
"epoch": 0.018765993744668752,
"grad_norm": 0.17728550732135773,
"learning_rate": 4.74991971191553e-05,
"loss": 1.0275,
"step": 66
},
{
"epoch": 0.019334660221779928,
"eval_loss": 1.0011852979660034,
"eval_runtime": 168.1174,
"eval_samples_per_second": 35.237,
"eval_steps_per_second": 4.408,
"step": 68
},
{
"epoch": 0.019618993460335514,
"grad_norm": 0.16994404792785645,
"learning_rate": 4.7229269307953235e-05,
"loss": 0.9841,
"step": 69
},
{
"epoch": 0.020471993176002276,
"grad_norm": 0.17181497812271118,
"learning_rate": 4.694636022946012e-05,
"loss": 0.9944,
"step": 72
},
{
"epoch": 0.021324992891669035,
"grad_norm": 0.21411621570587158,
"learning_rate": 4.665063509461097e-05,
"loss": 1.0013,
"step": 75
},
{
"epoch": 0.022177992607335797,
"grad_norm": 0.19594340026378632,
"learning_rate": 4.6342266598556814e-05,
"loss": 0.9969,
"step": 78
},
{
"epoch": 0.02303099232300256,
"grad_norm": 0.2057276964187622,
"learning_rate": 4.6021434819815555e-05,
"loss": 0.9808,
"step": 81
},
{
"epoch": 0.02388399203866932,
"grad_norm": 0.21778051555156708,
"learning_rate": 4.568832711511125e-05,
"loss": 0.9456,
"step": 84
},
{
"epoch": 0.024736991754336083,
"grad_norm": 0.21349306404590607,
"learning_rate": 4.534313800996299e-05,
"loss": 0.953,
"step": 87
},
{
"epoch": 0.02558999147000284,
"grad_norm": 0.21818219125270844,
"learning_rate": 4.498606908508754e-05,
"loss": 0.9133,
"step": 90
},
{
"epoch": 0.026442991185669604,
"grad_norm": 0.2510223388671875,
"learning_rate": 4.46173288586818e-05,
"loss": 0.9125,
"step": 93
},
{
"epoch": 0.027295990901336366,
"grad_norm": 0.24031595885753632,
"learning_rate": 4.4237132664654154e-05,
"loss": 0.8784,
"step": 96
},
{
"epoch": 0.028148990617003128,
"grad_norm": 0.2543729543685913,
"learning_rate": 4.384570252687542e-05,
"loss": 0.8984,
"step": 99
},
{
"epoch": 0.02900199033266989,
"grad_norm": 0.27020716667175293,
"learning_rate": 4.344326702952326e-05,
"loss": 0.8719,
"step": 102
},
{
"epoch": 0.02900199033266989,
"eval_loss": 0.8574855923652649,
"eval_runtime": 168.0651,
"eval_samples_per_second": 35.248,
"eval_steps_per_second": 4.409,
"step": 102
},
{
"epoch": 0.029854990048336652,
"grad_norm": 0.2670081853866577,
"learning_rate": 4.303006118359537e-05,
"loss": 0.8789,
"step": 105
},
{
"epoch": 0.03070798976400341,
"grad_norm": 0.3037591874599457,
"learning_rate": 4.260632628966974e-05,
"loss": 0.8526,
"step": 108
},
{
"epoch": 0.031560989479670176,
"grad_norm": 0.28440290689468384,
"learning_rate": 4.217230979699188e-05,
"loss": 0.8274,
"step": 111
},
{
"epoch": 0.03241398919533694,
"grad_norm": 0.49433842301368713,
"learning_rate": 4.172826515897146e-05,
"loss": 0.8104,
"step": 114
},
{
"epoch": 0.03326698891100369,
"grad_norm": 0.2911880314350128,
"learning_rate": 4.12744516851726e-05,
"loss": 0.8392,
"step": 117
},
{
"epoch": 0.034119988626670456,
"grad_norm": 0.3204285502433777,
"learning_rate": 4.0811134389884433e-05,
"loss": 0.788,
"step": 120
},
{
"epoch": 0.03497298834233722,
"grad_norm": 0.35650599002838135,
"learning_rate": 4.0338583837360225e-05,
"loss": 0.8004,
"step": 123
},
{
"epoch": 0.03582598805800398,
"grad_norm": 0.310162216424942,
"learning_rate": 3.985707598381544e-05,
"loss": 0.8085,
"step": 126
},
{
"epoch": 0.03667898777367074,
"grad_norm": 0.3370768129825592,
"learning_rate": 3.9366892016277096e-05,
"loss": 0.7317,
"step": 129
},
{
"epoch": 0.037531987489337504,
"grad_norm": 0.4600171446800232,
"learning_rate": 3.886831818837847e-05,
"loss": 0.7626,
"step": 132
},
{
"epoch": 0.038384987205004266,
"grad_norm": 0.43600377440452576,
"learning_rate": 3.8361645653195026e-05,
"loss": 0.7679,
"step": 135
},
{
"epoch": 0.038669320443559856,
"eval_loss": 0.7138456106185913,
"eval_runtime": 168.0686,
"eval_samples_per_second": 35.248,
"eval_steps_per_second": 4.409,
"step": 136
},
{
"epoch": 0.03923798692067103,
"grad_norm": 0.4538515508174896,
"learning_rate": 3.784717029321922e-05,
"loss": 0.7585,
"step": 138
},
{
"epoch": 0.04009098663633779,
"grad_norm": 0.5923985838890076,
"learning_rate": 3.732519254757344e-05,
"loss": 0.7211,
"step": 141
},
{
"epoch": 0.04094398635200455,
"grad_norm": 0.36592337489128113,
"learning_rate": 3.679601723656205e-05,
"loss": 0.6604,
"step": 144
},
{
"epoch": 0.04179698606767131,
"grad_norm": 0.44677990674972534,
"learning_rate": 3.625995338366492e-05,
"loss": 0.7846,
"step": 147
},
{
"epoch": 0.04264998578333807,
"grad_norm": 0.4921860098838806,
"learning_rate": 3.5717314035076355e-05,
"loss": 0.6709,
"step": 150
},
{
"epoch": 0.04350298549900483,
"grad_norm": 0.5457685589790344,
"learning_rate": 3.516841607689501e-05,
"loss": 0.6974,
"step": 153
},
{
"epoch": 0.044355985214671594,
"grad_norm": 0.46299344301223755,
"learning_rate": 3.461358005007128e-05,
"loss": 0.6643,
"step": 156
},
{
"epoch": 0.045208984930338356,
"grad_norm": 0.44877251982688904,
"learning_rate": 3.405312996322042e-05,
"loss": 0.6035,
"step": 159
},
{
"epoch": 0.04606198464600512,
"grad_norm": 0.3635808527469635,
"learning_rate": 3.348739310341068e-05,
"loss": 0.6781,
"step": 162
},
{
"epoch": 0.04691498436167188,
"grad_norm": 0.4172018766403198,
"learning_rate": 3.2916699845036816e-05,
"loss": 0.6195,
"step": 165
},
{
"epoch": 0.04776798407733864,
"grad_norm": 0.36372795701026917,
"learning_rate": 3.234138345689077e-05,
"loss": 0.6599,
"step": 168
},
{
"epoch": 0.048336650554449814,
"eval_loss": 0.606797456741333,
"eval_runtime": 168.0542,
"eval_samples_per_second": 35.251,
"eval_steps_per_second": 4.409,
"step": 170
},
{
"epoch": 0.048620983793005404,
"grad_norm": 0.4114479720592499,
"learning_rate": 3.17617799075421e-05,
"loss": 0.6287,
"step": 171
},
{
"epoch": 0.049473983508672166,
"grad_norm": 0.36733385920524597,
"learning_rate": 3.1178227669141744e-05,
"loss": 0.621,
"step": 174
},
{
"epoch": 0.05032698322433893,
"grad_norm": 0.3716995418071747,
"learning_rate": 3.0591067519763895e-05,
"loss": 0.5532,
"step": 177
},
{
"epoch": 0.05117998294000568,
"grad_norm": 0.4997945725917816,
"learning_rate": 3.0000642344401113e-05,
"loss": 0.6024,
"step": 180
},
{
"epoch": 0.052032982655672445,
"grad_norm": 0.43555817008018494,
"learning_rate": 2.9407296934729227e-05,
"loss": 0.5092,
"step": 183
},
{
"epoch": 0.05288598237133921,
"grad_norm": 0.49766504764556885,
"learning_rate": 2.8811377787758636e-05,
"loss": 0.6556,
"step": 186
},
{
"epoch": 0.05373898208700597,
"grad_norm": 0.6171467304229736,
"learning_rate": 2.8213232903489865e-05,
"loss": 0.5736,
"step": 189
},
{
"epoch": 0.05459198180267273,
"grad_norm": 0.8651450276374817,
"learning_rate": 2.761321158169134e-05,
"loss": 0.5568,
"step": 192
},
{
"epoch": 0.055444981518339494,
"grad_norm": 0.4486936926841736,
"learning_rate": 2.7011664217918154e-05,
"loss": 0.6145,
"step": 195
},
{
"epoch": 0.056297981234006256,
"grad_norm": 0.5348999500274658,
"learning_rate": 2.6408942098890936e-05,
"loss": 0.591,
"step": 198
},
{
"epoch": 0.05715098094967302,
"grad_norm": 0.4498997628688812,
"learning_rate": 2.580539719735433e-05,
"loss": 0.5572,
"step": 201
},
{
"epoch": 0.05800398066533978,
"grad_norm": 0.5493082404136658,
"learning_rate": 2.5201381966534748e-05,
"loss": 0.5316,
"step": 204
},
{
"epoch": 0.05800398066533978,
"eval_loss": 0.5341343879699707,
"eval_runtime": 167.9533,
"eval_samples_per_second": 35.272,
"eval_steps_per_second": 4.412,
"step": 204
},
{
"epoch": 0.05885698038100654,
"grad_norm": 0.47230827808380127,
"learning_rate": 2.459724913431772e-05,
"loss": 0.5152,
"step": 207
},
{
"epoch": 0.059709980096673304,
"grad_norm": 0.5665333867073059,
"learning_rate": 2.399335149726463e-05,
"loss": 0.5736,
"step": 210
},
{
"epoch": 0.06056297981234006,
"grad_norm": 0.47652244567871094,
"learning_rate": 2.3390041714589514e-05,
"loss": 0.5006,
"step": 213
},
{
"epoch": 0.06141597952800682,
"grad_norm": 0.7172293663024902,
"learning_rate": 2.2787672102216042e-05,
"loss": 0.4619,
"step": 216
},
{
"epoch": 0.06226897924367358,
"grad_norm": 0.5251888036727905,
"learning_rate": 2.2186594427034864e-05,
"loss": 0.4852,
"step": 219
},
{
"epoch": 0.06312197895934035,
"grad_norm": 0.45427563786506653,
"learning_rate": 2.1587159701481716e-05,
"loss": 0.4936,
"step": 222
},
{
"epoch": 0.06397497867500711,
"grad_norm": 0.49813351035118103,
"learning_rate": 2.098971797855599e-05,
"loss": 0.5062,
"step": 225
},
{
"epoch": 0.06482797839067388,
"grad_norm": 0.510427713394165,
"learning_rate": 2.0394618147399713e-05,
"loss": 0.4742,
"step": 228
},
{
"epoch": 0.06568097810634063,
"grad_norm": 0.5860615372657776,
"learning_rate": 1.980220772955602e-05,
"loss": 0.5472,
"step": 231
},
{
"epoch": 0.06653397782200739,
"grad_norm": 0.45956236124038696,
"learning_rate": 1.921283267602643e-05,
"loss": 0.5065,
"step": 234
},
{
"epoch": 0.06738697753767416,
"grad_norm": 0.46069616079330444,
"learning_rate": 1.8626837165245165e-05,
"loss": 0.4309,
"step": 237
},
{
"epoch": 0.06767131077622975,
"eval_loss": 0.4809924364089966,
"eval_runtime": 167.8669,
"eval_samples_per_second": 35.29,
"eval_steps_per_second": 4.414,
"step": 238
},
{
"epoch": 0.06823997725334091,
"grad_norm": 0.4818466603755951,
"learning_rate": 1.8044563402088684e-05,
"loss": 0.5623,
"step": 240
},
{
"epoch": 0.06909297696900768,
"grad_norm": 0.45440998673439026,
"learning_rate": 1.746635141803761e-05,
"loss": 0.5626,
"step": 243
},
{
"epoch": 0.06994597668467444,
"grad_norm": 0.5940297842025757,
"learning_rate": 1.6892538872607937e-05,
"loss": 0.5264,
"step": 246
},
{
"epoch": 0.0707989764003412,
"grad_norm": 0.41404712200164795,
"learning_rate": 1.6323460856167426e-05,
"loss": 0.4484,
"step": 249
},
{
"epoch": 0.07165197611600796,
"grad_norm": 0.44981393218040466,
"learning_rate": 1.5759449694252226e-05,
"loss": 0.473,
"step": 252
},
{
"epoch": 0.07250497583167473,
"grad_norm": 0.6551511883735657,
"learning_rate": 1.5200834753498128e-05,
"loss": 0.4956,
"step": 255
},
{
"epoch": 0.07335797554734148,
"grad_norm": 0.5032558441162109,
"learning_rate": 1.4647942249299707e-05,
"loss": 0.5262,
"step": 258
},
{
"epoch": 0.07421097526300825,
"grad_norm": 0.5430291891098022,
"learning_rate": 1.4101095055309746e-05,
"loss": 0.4899,
"step": 261
},
{
"epoch": 0.07506397497867501,
"grad_norm": 0.5020308494567871,
"learning_rate": 1.356061251489012e-05,
"loss": 0.494,
"step": 264
},
{
"epoch": 0.07591697469434176,
"grad_norm": 0.5282646417617798,
"learning_rate": 1.302681025462424e-05,
"loss": 0.5339,
"step": 267
},
{
"epoch": 0.07676997441000853,
"grad_norm": 0.48478028178215027,
"learning_rate": 1.2500000000000006e-05,
"loss": 0.5133,
"step": 270
},
{
"epoch": 0.07733864088711971,
"eval_loss": 0.4457505941390991,
"eval_runtime": 168.037,
"eval_samples_per_second": 35.254,
"eval_steps_per_second": 4.41,
"step": 272
}
],
"logging_steps": 3,
"max_steps": 400,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 34,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 8.710400478963302e+16,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}