|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9992217898832685, |
|
"eval_steps": 500, |
|
"global_step": 963, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0010376134889753567, |
|
"grad_norm": 23.969658904297845, |
|
"learning_rate": 1.0309278350515465e-07, |
|
"loss": 1.3725, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.005188067444876783, |
|
"grad_norm": 22.449071975220065, |
|
"learning_rate": 5.154639175257732e-07, |
|
"loss": 1.3719, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.010376134889753566, |
|
"grad_norm": 8.548339530816044, |
|
"learning_rate": 1.0309278350515464e-06, |
|
"loss": 1.2558, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01556420233463035, |
|
"grad_norm": 8.215121326664015, |
|
"learning_rate": 1.5463917525773197e-06, |
|
"loss": 1.081, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.020752269779507133, |
|
"grad_norm": 3.0841636846785283, |
|
"learning_rate": 2.061855670103093e-06, |
|
"loss": 0.9504, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02594033722438392, |
|
"grad_norm": 2.3540131831672575, |
|
"learning_rate": 2.577319587628866e-06, |
|
"loss": 0.9092, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0311284046692607, |
|
"grad_norm": 2.217217553056043, |
|
"learning_rate": 3.0927835051546395e-06, |
|
"loss": 0.8692, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03631647211413749, |
|
"grad_norm": 2.2563290398206615, |
|
"learning_rate": 3.6082474226804126e-06, |
|
"loss": 0.8445, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.041504539559014265, |
|
"grad_norm": 2.2874485501473907, |
|
"learning_rate": 4.123711340206186e-06, |
|
"loss": 0.8336, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04669260700389105, |
|
"grad_norm": 2.3330889165134967, |
|
"learning_rate": 4.639175257731959e-06, |
|
"loss": 0.8218, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.05188067444876784, |
|
"grad_norm": 2.16633956379982, |
|
"learning_rate": 5.154639175257732e-06, |
|
"loss": 0.8255, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.057068741893644616, |
|
"grad_norm": 2.355679025636223, |
|
"learning_rate": 5.670103092783505e-06, |
|
"loss": 0.7891, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.0622568093385214, |
|
"grad_norm": 2.4839432117685036, |
|
"learning_rate": 6.185567010309279e-06, |
|
"loss": 0.7814, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06744487678339818, |
|
"grad_norm": 2.480013950899919, |
|
"learning_rate": 6.701030927835052e-06, |
|
"loss": 0.7674, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.07263294422827497, |
|
"grad_norm": 2.3378408946103284, |
|
"learning_rate": 7.216494845360825e-06, |
|
"loss": 0.766, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07782101167315175, |
|
"grad_norm": 2.2751461973205482, |
|
"learning_rate": 7.731958762886599e-06, |
|
"loss": 0.7433, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08300907911802853, |
|
"grad_norm": 2.428511002623931, |
|
"learning_rate": 8.247422680412371e-06, |
|
"loss": 0.7414, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08819714656290532, |
|
"grad_norm": 2.4298836600324045, |
|
"learning_rate": 8.762886597938146e-06, |
|
"loss": 0.7358, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.0933852140077821, |
|
"grad_norm": 2.41313947506571, |
|
"learning_rate": 9.278350515463918e-06, |
|
"loss": 0.7319, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09857328145265888, |
|
"grad_norm": 2.4451429150679274, |
|
"learning_rate": 9.793814432989691e-06, |
|
"loss": 0.7323, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.10376134889753567, |
|
"grad_norm": 2.450193589248992, |
|
"learning_rate": 9.999703897419048e-06, |
|
"loss": 0.7231, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.10894941634241245, |
|
"grad_norm": 2.271786014084883, |
|
"learning_rate": 9.997894508649995e-06, |
|
"loss": 0.7149, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.11413748378728923, |
|
"grad_norm": 2.354564055245926, |
|
"learning_rate": 9.99444082710777e-06, |
|
"loss": 0.708, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11932555123216602, |
|
"grad_norm": 2.220428698962425, |
|
"learning_rate": 9.989343989043563e-06, |
|
"loss": 0.7216, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.1245136186770428, |
|
"grad_norm": 2.3141712328751396, |
|
"learning_rate": 9.982605671302293e-06, |
|
"loss": 0.7091, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1297016861219196, |
|
"grad_norm": 2.100396054783955, |
|
"learning_rate": 9.97422809077092e-06, |
|
"loss": 0.7066, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.13488975356679636, |
|
"grad_norm": 2.2484885982675413, |
|
"learning_rate": 9.9642140036491e-06, |
|
"loss": 0.7085, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.14007782101167315, |
|
"grad_norm": 2.1795476193729413, |
|
"learning_rate": 9.9525667045424e-06, |
|
"loss": 0.6889, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.14526588845654995, |
|
"grad_norm": 2.1757051871338593, |
|
"learning_rate": 9.93929002537839e-06, |
|
"loss": 0.6921, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1504539559014267, |
|
"grad_norm": 2.143005235580036, |
|
"learning_rate": 9.924388334145943e-06, |
|
"loss": 0.6907, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1556420233463035, |
|
"grad_norm": 2.1989760690420157, |
|
"learning_rate": 9.90786653345818e-06, |
|
"loss": 0.6912, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1608300907911803, |
|
"grad_norm": 2.004571277860471, |
|
"learning_rate": 9.889730058939529e-06, |
|
"loss": 0.6859, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.16601815823605706, |
|
"grad_norm": 2.05691987455993, |
|
"learning_rate": 9.869984877437413e-06, |
|
"loss": 0.6894, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.17120622568093385, |
|
"grad_norm": 2.230053895792029, |
|
"learning_rate": 9.848637485059183e-06, |
|
"loss": 0.6814, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.17639429312581065, |
|
"grad_norm": 1.9493958638517837, |
|
"learning_rate": 9.82569490503491e-06, |
|
"loss": 0.6731, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1815823605706874, |
|
"grad_norm": 2.133120594361784, |
|
"learning_rate": 9.80116468540677e-06, |
|
"loss": 0.6594, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1867704280155642, |
|
"grad_norm": 2.01624934264464, |
|
"learning_rate": 9.775054896545755e-06, |
|
"loss": 0.6751, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.191958495460441, |
|
"grad_norm": 2.1502691215852527, |
|
"learning_rate": 9.747374128496541e-06, |
|
"loss": 0.6789, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.19714656290531776, |
|
"grad_norm": 2.0484102083185194, |
|
"learning_rate": 9.718131488151399e-06, |
|
"loss": 0.6676, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.20233463035019456, |
|
"grad_norm": 2.0715841424222337, |
|
"learning_rate": 9.687336596254045e-06, |
|
"loss": 0.6616, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.20752269779507135, |
|
"grad_norm": 2.012157328183036, |
|
"learning_rate": 9.654999584234444e-06, |
|
"loss": 0.652, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2127107652399481, |
|
"grad_norm": 2.0669739212271923, |
|
"learning_rate": 9.621131090875603e-06, |
|
"loss": 0.6426, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.2178988326848249, |
|
"grad_norm": 2.0105636015375143, |
|
"learning_rate": 9.585742258813447e-06, |
|
"loss": 0.6445, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.2230869001297017, |
|
"grad_norm": 2.1108266544110688, |
|
"learning_rate": 9.548844730870903e-06, |
|
"loss": 0.6438, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.22827496757457846, |
|
"grad_norm": 2.072355913378756, |
|
"learning_rate": 9.51045064622747e-06, |
|
"loss": 0.6565, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.23346303501945526, |
|
"grad_norm": 2.166007360772802, |
|
"learning_rate": 9.470572636425451e-06, |
|
"loss": 0.647, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.23865110246433205, |
|
"grad_norm": 2.022875957881762, |
|
"learning_rate": 9.429223821214213e-06, |
|
"loss": 0.6325, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2438391699092088, |
|
"grad_norm": 2.006861087987301, |
|
"learning_rate": 9.386417804233836e-06, |
|
"loss": 0.6477, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.2490272373540856, |
|
"grad_norm": 2.0140489204477645, |
|
"learning_rate": 9.34216866853954e-06, |
|
"loss": 0.6391, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.25421530479896237, |
|
"grad_norm": 1.9489606047213677, |
|
"learning_rate": 9.296490971968416e-06, |
|
"loss": 0.6283, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.2594033722438392, |
|
"grad_norm": 2.072486707132733, |
|
"learning_rate": 9.249399742349928e-06, |
|
"loss": 0.6377, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.26459143968871596, |
|
"grad_norm": 1.9650189580925839, |
|
"learning_rate": 9.20091047256181e-06, |
|
"loss": 0.6261, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.2697795071335927, |
|
"grad_norm": 1.9241991797476943, |
|
"learning_rate": 9.151039115432946e-06, |
|
"loss": 0.6184, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.27496757457846954, |
|
"grad_norm": 1.9743470888532664, |
|
"learning_rate": 9.099802078494947e-06, |
|
"loss": 0.6142, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.2801556420233463, |
|
"grad_norm": 2.160988187935936, |
|
"learning_rate": 9.047216218584105e-06, |
|
"loss": 0.6094, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.2853437094682231, |
|
"grad_norm": 1.9697508480614465, |
|
"learning_rate": 8.993298836295556e-06, |
|
"loss": 0.6196, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.2905317769130999, |
|
"grad_norm": 1.8771524751425768, |
|
"learning_rate": 8.93806767029143e-06, |
|
"loss": 0.6163, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.29571984435797666, |
|
"grad_norm": 2.125863779805947, |
|
"learning_rate": 8.88154089146488e-06, |
|
"loss": 0.6167, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.3009079118028534, |
|
"grad_norm": 2.1188493077731514, |
|
"learning_rate": 8.823737096961916e-06, |
|
"loss": 0.5992, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.30609597924773024, |
|
"grad_norm": 2.1335267497592807, |
|
"learning_rate": 8.764675304062992e-06, |
|
"loss": 0.6071, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.311284046692607, |
|
"grad_norm": 2.036189297244598, |
|
"learning_rate": 8.704374943926386e-06, |
|
"loss": 0.609, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3164721141374838, |
|
"grad_norm": 1.915927299304865, |
|
"learning_rate": 8.642855855195394e-06, |
|
"loss": 0.5945, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.3216601815823606, |
|
"grad_norm": 2.005194485630929, |
|
"learning_rate": 8.580138277471476e-06, |
|
"loss": 0.5959, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.32684824902723736, |
|
"grad_norm": 2.1368034472887527, |
|
"learning_rate": 8.516242844655498e-06, |
|
"loss": 0.5941, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.3320363164721141, |
|
"grad_norm": 1.9360804934529585, |
|
"learning_rate": 8.45119057815922e-06, |
|
"loss": 0.5915, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.33722438391699094, |
|
"grad_norm": 1.9356101875463727, |
|
"learning_rate": 8.385002879989328e-06, |
|
"loss": 0.5838, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.3424124513618677, |
|
"grad_norm": 2.4311425501079023, |
|
"learning_rate": 8.317701525706226e-06, |
|
"loss": 0.5946, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.3476005188067445, |
|
"grad_norm": 2.356263841306792, |
|
"learning_rate": 8.249308657259943e-06, |
|
"loss": 0.567, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.3527885862516213, |
|
"grad_norm": 2.048334150791661, |
|
"learning_rate": 8.179846775705504e-06, |
|
"loss": 0.5795, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.35797665369649806, |
|
"grad_norm": 1.9977511587812506, |
|
"learning_rate": 8.109338733800132e-06, |
|
"loss": 0.5751, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.3631647211413748, |
|
"grad_norm": 1.8688618314869894, |
|
"learning_rate": 8.03780772848477e-06, |
|
"loss": 0.568, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.36835278858625164, |
|
"grad_norm": 1.93022130905715, |
|
"learning_rate": 7.965277293252354e-06, |
|
"loss": 0.5682, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.3735408560311284, |
|
"grad_norm": 2.0382225242835528, |
|
"learning_rate": 7.891771290405351e-06, |
|
"loss": 0.5617, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.3787289234760052, |
|
"grad_norm": 1.9924209327442368, |
|
"learning_rate": 7.817313903205148e-06, |
|
"loss": 0.5577, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.383916990920882, |
|
"grad_norm": 1.9678458173326334, |
|
"learning_rate": 7.741929627915814e-06, |
|
"loss": 0.56, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.38910505836575876, |
|
"grad_norm": 2.2405618654805215, |
|
"learning_rate": 7.66564326574491e-06, |
|
"loss": 0.5513, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.3942931258106355, |
|
"grad_norm": 1.9971872990885233, |
|
"learning_rate": 7.588479914683954e-06, |
|
"loss": 0.5445, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.39948119325551235, |
|
"grad_norm": 2.06807252227761, |
|
"learning_rate": 7.510464961251271e-06, |
|
"loss": 0.5674, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.4046692607003891, |
|
"grad_norm": 1.9627368535332135, |
|
"learning_rate": 7.431624072139884e-06, |
|
"loss": 0.5435, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.4098573281452659, |
|
"grad_norm": 1.9716804464407136, |
|
"learning_rate": 7.351983185773259e-06, |
|
"loss": 0.5552, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.4150453955901427, |
|
"grad_norm": 1.9693396583392846, |
|
"learning_rate": 7.271568503771632e-06, |
|
"loss": 0.5343, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.42023346303501946, |
|
"grad_norm": 1.9432949161104107, |
|
"learning_rate": 7.190406482331757e-06, |
|
"loss": 0.5475, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.4254215304798962, |
|
"grad_norm": 2.0194917717314045, |
|
"learning_rate": 7.108523823522891e-06, |
|
"loss": 0.5477, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.43060959792477305, |
|
"grad_norm": 2.206404974952941, |
|
"learning_rate": 7.0259474665018915e-06, |
|
"loss": 0.5425, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.4357976653696498, |
|
"grad_norm": 1.9526533277899327, |
|
"learning_rate": 6.942704578650312e-06, |
|
"loss": 0.5161, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.4409857328145266, |
|
"grad_norm": 2.0097466124913117, |
|
"learning_rate": 6.858822546636417e-06, |
|
"loss": 0.5331, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.4461738002594034, |
|
"grad_norm": 1.8348649689633039, |
|
"learning_rate": 6.774328967405035e-06, |
|
"loss": 0.523, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.45136186770428016, |
|
"grad_norm": 2.139084532722164, |
|
"learning_rate": 6.689251639098261e-06, |
|
"loss": 0.5251, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.4565499351491569, |
|
"grad_norm": 1.9708479629081865, |
|
"learning_rate": 6.603618551909935e-06, |
|
"loss": 0.5232, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.46173800259403375, |
|
"grad_norm": 1.9331722289768318, |
|
"learning_rate": 6.517457878876958e-06, |
|
"loss": 0.5305, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.4669260700389105, |
|
"grad_norm": 1.859009250403284, |
|
"learning_rate": 6.430797966610436e-06, |
|
"loss": 0.5159, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4721141374837873, |
|
"grad_norm": 1.986527066309499, |
|
"learning_rate": 6.343667325969736e-06, |
|
"loss": 0.5367, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.4773022049286641, |
|
"grad_norm": 1.9771277544299588, |
|
"learning_rate": 6.256094622682493e-06, |
|
"loss": 0.5123, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.48249027237354086, |
|
"grad_norm": 2.0022259730400904, |
|
"learning_rate": 6.168108667913666e-06, |
|
"loss": 0.5166, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.4876783398184176, |
|
"grad_norm": 1.9991961519932744, |
|
"learning_rate": 6.079738408786753e-06, |
|
"loss": 0.5161, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.49286640726329445, |
|
"grad_norm": 2.0805595238898307, |
|
"learning_rate": 5.9910129188602665e-06, |
|
"loss": 0.5179, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.4980544747081712, |
|
"grad_norm": 1.929253006230254, |
|
"learning_rate": 5.9019613885626235e-06, |
|
"loss": 0.5097, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.503242542153048, |
|
"grad_norm": 2.25129632838715, |
|
"learning_rate": 5.812613115588575e-06, |
|
"loss": 0.4971, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.5084306095979247, |
|
"grad_norm": 1.9119339241166262, |
|
"learning_rate": 5.722997495260348e-06, |
|
"loss": 0.4988, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5136186770428015, |
|
"grad_norm": 1.8300200112998326, |
|
"learning_rate": 5.6331440108566735e-06, |
|
"loss": 0.4941, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.5188067444876784, |
|
"grad_norm": 1.9591247994452368, |
|
"learning_rate": 5.543082223912875e-06, |
|
"loss": 0.492, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5239948119325551, |
|
"grad_norm": 1.99136453982626, |
|
"learning_rate": 5.452841764495203e-06, |
|
"loss": 0.5002, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.5291828793774319, |
|
"grad_norm": 1.9961024804052654, |
|
"learning_rate": 5.362452321452636e-06, |
|
"loss": 0.4772, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5343709468223087, |
|
"grad_norm": 1.9607124098040063, |
|
"learning_rate": 5.2719436326493255e-06, |
|
"loss": 0.4908, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.5395590142671854, |
|
"grad_norm": 1.9303906010446525, |
|
"learning_rate": 5.181345475180941e-06, |
|
"loss": 0.4866, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5447470817120622, |
|
"grad_norm": 2.0420688559734503, |
|
"learning_rate": 5.090687655578078e-06, |
|
"loss": 0.4769, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.5499351491569391, |
|
"grad_norm": 1.9908642175713687, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4742, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5551232166018158, |
|
"grad_norm": 1.9960779934532675, |
|
"learning_rate": 4.909312344421923e-06, |
|
"loss": 0.4666, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.5603112840466926, |
|
"grad_norm": 1.9274839933909422, |
|
"learning_rate": 4.8186545248190604e-06, |
|
"loss": 0.4866, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.5654993514915694, |
|
"grad_norm": 1.9162466337096817, |
|
"learning_rate": 4.7280563673506745e-06, |
|
"loss": 0.4692, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.5706874189364461, |
|
"grad_norm": 2.07386431606307, |
|
"learning_rate": 4.637547678547366e-06, |
|
"loss": 0.4859, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5758754863813229, |
|
"grad_norm": 2.0201984812958385, |
|
"learning_rate": 4.547158235504797e-06, |
|
"loss": 0.4718, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.5810635538261998, |
|
"grad_norm": 1.95015272613481, |
|
"learning_rate": 4.4569177760871255e-06, |
|
"loss": 0.475, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.5862516212710766, |
|
"grad_norm": 1.944586565605588, |
|
"learning_rate": 4.366855989143326e-06, |
|
"loss": 0.4551, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.5914396887159533, |
|
"grad_norm": 1.9208589567145171, |
|
"learning_rate": 4.277002504739653e-06, |
|
"loss": 0.4686, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.5966277561608301, |
|
"grad_norm": 1.8639671285460482, |
|
"learning_rate": 4.187386884411426e-06, |
|
"loss": 0.4557, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.6018158236057068, |
|
"grad_norm": 1.9975578797091653, |
|
"learning_rate": 4.098038611437377e-06, |
|
"loss": 0.4651, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.6070038910505836, |
|
"grad_norm": 1.961651938542185, |
|
"learning_rate": 4.008987081139734e-06, |
|
"loss": 0.4643, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.6121919584954605, |
|
"grad_norm": 1.9374158302120401, |
|
"learning_rate": 3.920261591213249e-06, |
|
"loss": 0.4556, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6173800259403373, |
|
"grad_norm": 1.9090835435895448, |
|
"learning_rate": 3.8318913320863355e-06, |
|
"loss": 0.4536, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.622568093385214, |
|
"grad_norm": 1.8975263865890188, |
|
"learning_rate": 3.7439053773175092e-06, |
|
"loss": 0.4615, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6277561608300908, |
|
"grad_norm": 1.9060390294655216, |
|
"learning_rate": 3.6563326740302664e-06, |
|
"loss": 0.4459, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.6329442282749675, |
|
"grad_norm": 1.9725006931962796, |
|
"learning_rate": 3.569202033389565e-06, |
|
"loss": 0.4451, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6381322957198443, |
|
"grad_norm": 1.9621067476956515, |
|
"learning_rate": 3.4825421211230437e-06, |
|
"loss": 0.4419, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.6433203631647212, |
|
"grad_norm": 2.098443239659209, |
|
"learning_rate": 3.3963814480900665e-06, |
|
"loss": 0.4415, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.648508430609598, |
|
"grad_norm": 1.8981208726840302, |
|
"learning_rate": 3.310748360901741e-06, |
|
"loss": 0.4456, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.6536964980544747, |
|
"grad_norm": 1.8947168989269416, |
|
"learning_rate": 3.225671032594966e-06, |
|
"loss": 0.4229, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.6588845654993515, |
|
"grad_norm": 2.0138652650509288, |
|
"learning_rate": 3.1411774533635854e-06, |
|
"loss": 0.437, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.6640726329442282, |
|
"grad_norm": 1.8903378440015823, |
|
"learning_rate": 3.0572954213496897e-06, |
|
"loss": 0.4454, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.669260700389105, |
|
"grad_norm": 1.8448484960177367, |
|
"learning_rate": 2.9740525334981105e-06, |
|
"loss": 0.4398, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.6744487678339819, |
|
"grad_norm": 1.9976530631786225, |
|
"learning_rate": 2.8914761764771093e-06, |
|
"loss": 0.429, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6796368352788587, |
|
"grad_norm": 1.9155018572353837, |
|
"learning_rate": 2.809593517668243e-06, |
|
"loss": 0.4309, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.6848249027237354, |
|
"grad_norm": 1.942714148946629, |
|
"learning_rate": 2.728431496228369e-06, |
|
"loss": 0.4248, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.6900129701686122, |
|
"grad_norm": 2.013023734418392, |
|
"learning_rate": 2.648016814226742e-06, |
|
"loss": 0.4304, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.695201037613489, |
|
"grad_norm": 1.9023117871214554, |
|
"learning_rate": 2.5683759278601174e-06, |
|
"loss": 0.4338, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.7003891050583657, |
|
"grad_norm": 1.8911448184302957, |
|
"learning_rate": 2.4895350387487304e-06, |
|
"loss": 0.4245, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.7055771725032426, |
|
"grad_norm": 2.0358392917626813, |
|
"learning_rate": 2.4115200853160475e-06, |
|
"loss": 0.4194, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.7107652399481194, |
|
"grad_norm": 1.9510576677492195, |
|
"learning_rate": 2.3343567342550933e-06, |
|
"loss": 0.4267, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.7159533073929961, |
|
"grad_norm": 1.8690267408594539, |
|
"learning_rate": 2.258070372084188e-06, |
|
"loss": 0.4312, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7211413748378729, |
|
"grad_norm": 1.8322122073891454, |
|
"learning_rate": 2.182686096794852e-06, |
|
"loss": 0.4207, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.7263294422827496, |
|
"grad_norm": 2.0311002524177253, |
|
"learning_rate": 2.108228709594649e-06, |
|
"loss": 0.4227, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7315175097276264, |
|
"grad_norm": 1.8678394687630775, |
|
"learning_rate": 2.0347227067476478e-06, |
|
"loss": 0.4149, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.7367055771725033, |
|
"grad_norm": 1.8521301731665931, |
|
"learning_rate": 1.962192271515232e-06, |
|
"loss": 0.4192, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.74189364461738, |
|
"grad_norm": 1.9291143236144128, |
|
"learning_rate": 1.8906612661998698e-06, |
|
"loss": 0.4128, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.7470817120622568, |
|
"grad_norm": 1.991134829662921, |
|
"learning_rate": 1.820153224294498e-06, |
|
"loss": 0.4102, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.7522697795071336, |
|
"grad_norm": 1.8597303553848081, |
|
"learning_rate": 1.750691342740058e-06, |
|
"loss": 0.4104, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.7574578469520103, |
|
"grad_norm": 1.8334844899907363, |
|
"learning_rate": 1.6822984742937764e-06, |
|
"loss": 0.4049, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.7626459143968871, |
|
"grad_norm": 1.839241095874111, |
|
"learning_rate": 1.6149971200106723e-06, |
|
"loss": 0.4009, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.767833981841764, |
|
"grad_norm": 1.8614288971061537, |
|
"learning_rate": 1.548809421840779e-06, |
|
"loss": 0.4029, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.7730220492866408, |
|
"grad_norm": 1.9712640153496117, |
|
"learning_rate": 1.483757155344503e-06, |
|
"loss": 0.4056, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.7782101167315175, |
|
"grad_norm": 1.7973975593361922, |
|
"learning_rate": 1.4198617225285244e-06, |
|
"loss": 0.409, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.7833981841763943, |
|
"grad_norm": 1.8870139707940816, |
|
"learning_rate": 1.3571441448046086e-06, |
|
"loss": 0.4117, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.788586251621271, |
|
"grad_norm": 1.9255920717839368, |
|
"learning_rate": 1.2956250560736143e-06, |
|
"loss": 0.4097, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.7937743190661478, |
|
"grad_norm": 1.8605570734597534, |
|
"learning_rate": 1.2353246959370086e-06, |
|
"loss": 0.3885, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.7989623865110247, |
|
"grad_norm": 1.8678208345700735, |
|
"learning_rate": 1.1762629030380867e-06, |
|
"loss": 0.4044, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.8041504539559015, |
|
"grad_norm": 1.8406724398818959, |
|
"learning_rate": 1.118459108535122e-06, |
|
"loss": 0.3991, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.8093385214007782, |
|
"grad_norm": 1.9447178497450672, |
|
"learning_rate": 1.061932329708572e-06, |
|
"loss": 0.3878, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.814526588845655, |
|
"grad_norm": 1.903347183666585, |
|
"learning_rate": 1.006701163704445e-06, |
|
"loss": 0.3994, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.8197146562905318, |
|
"grad_norm": 1.7817648380438804, |
|
"learning_rate": 9.527837814158963e-07, |
|
"loss": 0.3943, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.8249027237354085, |
|
"grad_norm": 1.89718875917406, |
|
"learning_rate": 9.001979215050544e-07, |
|
"loss": 0.3929, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.8300907911802854, |
|
"grad_norm": 1.8825895138353903, |
|
"learning_rate": 8.489608845670527e-07, |
|
"loss": 0.3924, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8352788586251622, |
|
"grad_norm": 1.7908515887362904, |
|
"learning_rate": 7.99089527438191e-07, |
|
"loss": 0.3919, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.8404669260700389, |
|
"grad_norm": 2.0314129121613034, |
|
"learning_rate": 7.506002576500732e-07, |
|
"loss": 0.3941, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.8456549935149157, |
|
"grad_norm": 1.838737045068825, |
|
"learning_rate": 7.035090280315854e-07, |
|
"loss": 0.398, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.8508430609597925, |
|
"grad_norm": 1.8701463175206698, |
|
"learning_rate": 6.578313314604612e-07, |
|
"loss": 0.395, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.8560311284046692, |
|
"grad_norm": 1.8734176088672492, |
|
"learning_rate": 6.135821957661658e-07, |
|
"loss": 0.3945, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.8612191958495461, |
|
"grad_norm": 1.8454123160341045, |
|
"learning_rate": 5.707761787857879e-07, |
|
"loss": 0.3855, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.8664072632944229, |
|
"grad_norm": 1.7755466173110739, |
|
"learning_rate": 5.294273635745517e-07, |
|
"loss": 0.3971, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.8715953307392996, |
|
"grad_norm": 1.8394864397787671, |
|
"learning_rate": 4.895493537725326e-07, |
|
"loss": 0.3966, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.8767833981841764, |
|
"grad_norm": 1.7915928948304078, |
|
"learning_rate": 4.511552691290988e-07, |
|
"loss": 0.3979, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.8819714656290532, |
|
"grad_norm": 1.7863787006400424, |
|
"learning_rate": 4.1425774118655505e-07, |
|
"loss": 0.3826, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.8871595330739299, |
|
"grad_norm": 1.8024141112662704, |
|
"learning_rate": 3.7886890912439633e-07, |
|
"loss": 0.3862, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.8923476005188068, |
|
"grad_norm": 1.8180972720099156, |
|
"learning_rate": 3.4500041576555733e-07, |
|
"loss": 0.3859, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.8975356679636836, |
|
"grad_norm": 1.7622051589037506, |
|
"learning_rate": 3.1266340374595693e-07, |
|
"loss": 0.3831, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.9027237354085603, |
|
"grad_norm": 1.7926800043760007, |
|
"learning_rate": 2.818685118486025e-07, |
|
"loss": 0.3927, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.9079118028534371, |
|
"grad_norm": 1.8515815235983688, |
|
"learning_rate": 2.526258715034602e-07, |
|
"loss": 0.3787, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.9130998702983139, |
|
"grad_norm": 1.761786866550431, |
|
"learning_rate": 2.2494510345424657e-07, |
|
"loss": 0.3881, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.9182879377431906, |
|
"grad_norm": 1.731506703869926, |
|
"learning_rate": 1.988353145932298e-07, |
|
"loss": 0.3762, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.9234760051880675, |
|
"grad_norm": 1.8427166106595052, |
|
"learning_rate": 1.7430509496508985e-07, |
|
"loss": 0.3975, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.9286640726329443, |
|
"grad_norm": 1.761769698023775, |
|
"learning_rate": 1.5136251494081822e-07, |
|
"loss": 0.3842, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.933852140077821, |
|
"grad_norm": 1.8297504100937483, |
|
"learning_rate": 1.3001512256258841e-07, |
|
"loss": 0.3916, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.9390402075226978, |
|
"grad_norm": 1.8143369848190358, |
|
"learning_rate": 1.1026994106047296e-07, |
|
"loss": 0.3911, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.9442282749675746, |
|
"grad_norm": 1.7462314691918333, |
|
"learning_rate": 9.213346654182054e-08, |
|
"loss": 0.3888, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.9494163424124513, |
|
"grad_norm": 1.842285372864709, |
|
"learning_rate": 7.561166585405789e-08, |
|
"loss": 0.3823, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.9546044098573282, |
|
"grad_norm": 1.798454935332072, |
|
"learning_rate": 6.070997462161055e-08, |
|
"loss": 0.4032, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.959792477302205, |
|
"grad_norm": 1.8579672164577692, |
|
"learning_rate": 4.743329545760122e-08, |
|
"loss": 0.3811, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.9649805447470817, |
|
"grad_norm": 1.764976690651984, |
|
"learning_rate": 3.578599635090163e-08, |
|
"loss": 0.3806, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.9701686121919585, |
|
"grad_norm": 1.7085373084916373, |
|
"learning_rate": 2.577190922908035e-08, |
|
"loss": 0.3888, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.9753566796368353, |
|
"grad_norm": 1.7431684765639506, |
|
"learning_rate": 1.7394328697707407e-08, |
|
"loss": 0.3901, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.980544747081712, |
|
"grad_norm": 1.8495600056895127, |
|
"learning_rate": 1.0656010956437979e-08, |
|
"loss": 0.3918, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.9857328145265889, |
|
"grad_norm": 1.8616847493274582, |
|
"learning_rate": 5.5591728922316235e-09, |
|
"loss": 0.3895, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.9909208819714657, |
|
"grad_norm": 1.8274058784400706, |
|
"learning_rate": 2.1054913500051512e-09, |
|
"loss": 0.3831, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.9961089494163424, |
|
"grad_norm": 1.7888916632814764, |
|
"learning_rate": 2.9610258095169596e-10, |
|
"loss": 0.3863, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.9992217898832685, |
|
"eval_loss": 0.35284245014190674, |
|
"eval_runtime": 0.9437, |
|
"eval_samples_per_second": 2.119, |
|
"eval_steps_per_second": 1.06, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.9992217898832685, |
|
"step": 963, |
|
"total_flos": 201580263505920.0, |
|
"train_loss": 0.5411187405403034, |
|
"train_runtime": 23935.6127, |
|
"train_samples_per_second": 1.288, |
|
"train_steps_per_second": 0.04 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 963, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 201580263505920.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|