|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.979591836734694, |
|
"eval_steps": 500, |
|
"global_step": 610, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 312.0, |
|
"learning_rate": 3.278688524590164e-06, |
|
"loss": 33.9196, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 354.0, |
|
"learning_rate": 1.6393442622950818e-05, |
|
"loss": 40.6749, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 125.0, |
|
"learning_rate": 3.2786885245901635e-05, |
|
"loss": 27.6429, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 29.875, |
|
"learning_rate": 4.918032786885246e-05, |
|
"loss": 19.7282, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 10.1875, |
|
"learning_rate": 6.557377049180327e-05, |
|
"loss": 16.6712, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 5.53125, |
|
"learning_rate": 8.19672131147541e-05, |
|
"loss": 14.7944, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 7.09375, |
|
"learning_rate": 9.836065573770493e-05, |
|
"loss": 14.0597, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 11.875, |
|
"learning_rate": 0.00011475409836065574, |
|
"loss": 12.5767, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 55.75, |
|
"learning_rate": 0.00013114754098360654, |
|
"loss": 9.9687, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 9.5, |
|
"learning_rate": 0.00014754098360655738, |
|
"loss": 2.8926, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 2.78125, |
|
"learning_rate": 0.0001639344262295082, |
|
"loss": 1.5807, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 2.609375, |
|
"learning_rate": 0.00018032786885245904, |
|
"loss": 1.3009, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 0.00019672131147540985, |
|
"loss": 1.1905, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 0.00019997380453683512, |
|
"loss": 1.1472, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 3.5, |
|
"learning_rate": 0.00019986740898848306, |
|
"loss": 1.0454, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 0.00019967926316561134, |
|
"loss": 0.9724, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 2.421875, |
|
"learning_rate": 0.00019940952108215526, |
|
"loss": 0.9521, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 6.9375, |
|
"learning_rate": 0.00019905840354577972, |
|
"loss": 0.9019, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 3.78125, |
|
"learning_rate": 0.00019862619797712846, |
|
"loss": 0.8665, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 0.00019811325817454517, |
|
"loss": 0.8629, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 0.00019752000402445825, |
|
"loss": 0.7888, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 0.0001968469211576668, |
|
"loss": 0.8141, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 3.65625, |
|
"learning_rate": 0.00019609456055180897, |
|
"loss": 0.8212, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 6.96875, |
|
"learning_rate": 0.00019526353808033825, |
|
"loss": 0.8073, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 0.00019435453400837684, |
|
"loss": 0.8192, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.2006253004074097, |
|
"eval_runtime": 1.1043, |
|
"eval_samples_per_second": 1.811, |
|
"eval_steps_per_second": 0.906, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.62890625, |
|
"learning_rate": 0.00019336829243585856, |
|
"loss": 0.7633, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 0.0001923056206884176, |
|
"loss": 0.7091, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 0.00019116738865652134, |
|
"loss": 0.6773, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 1.2578125, |
|
"learning_rate": 0.00018995452808338823, |
|
"loss": 0.7058, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 0.00018866803180227402, |
|
"loss": 0.7189, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 0.00018730895292375018, |
|
"loss": 0.6698, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 0.00018587840397364007, |
|
"loss": 0.6941, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.00018437755598231856, |
|
"loss": 0.6511, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 0.00018280763752612052, |
|
"loss": 0.6672, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.72265625, |
|
"learning_rate": 0.00018116993372164263, |
|
"loss": 0.698, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.7890625, |
|
"learning_rate": 0.0001794657851737625, |
|
"loss": 0.6736, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.00017769658687823525, |
|
"loss": 0.651, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001758637870797665, |
|
"loss": 0.6809, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 0.0001739688860864967, |
|
"loss": 0.6331, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.8359375, |
|
"learning_rate": 0.00017201343504186643, |
|
"loss": 0.6562, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 0.00016999903465486914, |
|
"loss": 0.6416, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 1.3359375, |
|
"learning_rate": 0.00016792733388972932, |
|
"loss": 0.6957, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.93359375, |
|
"learning_rate": 0.0001658000286160805, |
|
"loss": 0.6857, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 1.0703125, |
|
"learning_rate": 0.00016361886022074612, |
|
"loss": 0.6589, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 0.0001613856141822612, |
|
"loss": 0.6635, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 0.00015910211860930065, |
|
"loss": 0.6252, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.78515625, |
|
"learning_rate": 0.0001567702427442113, |
|
"loss": 0.6187, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 0.00015439189543287247, |
|
"loss": 0.6194, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 1.328125, |
|
"learning_rate": 0.00015196902356213714, |
|
"loss": 0.6396, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.87109375, |
|
"learning_rate": 0.0001495036104661337, |
|
"loss": 0.6377, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.1303967237472534, |
|
"eval_runtime": 1.0903, |
|
"eval_samples_per_second": 1.834, |
|
"eval_steps_per_second": 0.917, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 1.0234375, |
|
"learning_rate": 0.000146997674302732, |
|
"loss": 0.5446, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.76171875, |
|
"learning_rate": 0.0001444532664015037, |
|
"loss": 0.5915, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 0.0001418724695845277, |
|
"loss": 0.5521, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 0.0001392573964614172, |
|
"loss": 0.5279, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.00013661018769996227, |
|
"loss": 0.558, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.82421875, |
|
"learning_rate": 0.00013393301027380476, |
|
"loss": 0.5422, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.00013122805568857948, |
|
"loss": 0.5401, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.00012849753818797353, |
|
"loss": 0.5458, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.00012574369294117297, |
|
"loss": 0.5309, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001229687742131796, |
|
"loss": 0.5276, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.6796875, |
|
"learning_rate": 0.00012017505351949661, |
|
"loss": 0.5322, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00011736481776669306, |
|
"loss": 0.5242, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.00011454036738036899, |
|
"loss": 0.5104, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 1.0234375, |
|
"learning_rate": 0.00011170401442205474, |
|
"loss": 0.6306, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.00010885808069658451, |
|
"loss": 0.5246, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.86328125, |
|
"learning_rate": 0.00010600489585149484, |
|
"loss": 0.5307, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001031467954700025, |
|
"loss": 0.523, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 0.00010028611915912405, |
|
"loss": 0.5449, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 9.742520863450115e-05, |
|
"loss": 0.5226, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 9.456640580350018e-05, |
|
"loss": 0.5173, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.76953125, |
|
"learning_rate": 9.171205084815444e-05, |
|
"loss": 0.5281, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.6796875, |
|
"learning_rate": 8.886448030951912e-05, |
|
"loss": 0.543, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 8.602602517500651e-05, |
|
"loss": 0.5459, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.319900897026733e-05, |
|
"loss": 0.5334, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.1456480026245117, |
|
"eval_runtime": 1.0913, |
|
"eval_samples_per_second": 1.833, |
|
"eval_steps_per_second": 0.916, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.038574585718032e-05, |
|
"loss": 0.5025, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 7.758853873950676e-05, |
|
"loss": 0.4247, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.480967737776089e-05, |
|
"loss": 0.4551, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 7.205143651483906e-05, |
|
"loss": 0.4667, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 6.931607401394229e-05, |
|
"loss": 0.4581, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"grad_norm": 0.7421875, |
|
"learning_rate": 6.660582901031621e-05, |
|
"loss": 0.4441, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 6.392292007832168e-05, |
|
"loss": 0.4247, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"grad_norm": 0.66796875, |
|
"learning_rate": 6.126954341533599e-05, |
|
"loss": 0.4214, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 5.864787104397194e-05, |
|
"loss": 0.4413, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"grad_norm": 0.65234375, |
|
"learning_rate": 5.6060049034085815e-05, |
|
"loss": 0.4648, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 5.3508195746030296e-05, |
|
"loss": 0.5039, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.099440009658957e-05, |
|
"loss": 0.4493, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 4.852071984901696e-05, |
|
"loss": 0.4343, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 4.6089179928574487e-05, |
|
"loss": 0.4354, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 4.3701770764953124e-05, |
|
"loss": 0.4694, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 4.136044666293044e-05, |
|
"loss": 0.4446, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 3.9067124202600194e-05, |
|
"loss": 0.4398, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 3.6823680670482485e-05, |
|
"loss": 0.4449, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 3.463195252279939e-05, |
|
"loss": 0.4443, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 3.24937338821738e-05, |
|
"loss": 0.4346, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 3.0410775068981613e-05, |
|
"loss": 0.4423, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"grad_norm": 0.828125, |
|
"learning_rate": 2.8384781168560693e-05, |
|
"loss": 0.4304, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 2.6417410635448013e-05, |
|
"loss": 0.443, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 2.451027393578863e-05, |
|
"loss": 0.4392, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.5, |
|
"learning_rate": 2.2664932229027024e-05, |
|
"loss": 0.4454, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.1934518814086914, |
|
"eval_runtime": 1.0901, |
|
"eval_samples_per_second": 1.835, |
|
"eval_steps_per_second": 0.917, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 2.088289608996071e-05, |
|
"loss": 0.3925, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 1.9165624272201567e-05, |
|
"loss": 0.403, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"grad_norm": 0.5, |
|
"learning_rate": 1.7514522514057553e-05, |
|
"loss": 0.3824, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 1.593094238781191e-05, |
|
"loss": 0.3818, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 1.4416180193342144e-05, |
|
"loss": 0.4017, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 1.2971475896984475e-05, |
|
"loss": 0.4047, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 1.1598012116512092e-05, |
|
"loss": 0.3735, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 1.0296913153058163e-05, |
|
"loss": 0.3954, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.069244070776428e-06, |
|
"loss": 0.3955, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.916009824992254e-06, |
|
"loss": 0.3824, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 6.838154439558253e-06, |
|
"loss": 0.3846, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.836560234087418e-06, |
|
"loss": 0.4087, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 4.912047101696848e-06, |
|
"loss": 0.3785, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 4.065371837853016e-06, |
|
"loss": 0.4013, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 3.2972275208679625e-06, |
|
"loss": 0.4005, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 2.608242944553607e-06, |
|
"loss": 0.3842, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 1.9989821034986032e-06, |
|
"loss": 0.3791, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 1.4699437313891007e-06, |
|
"loss": 0.398, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 1.02156089275125e-06, |
|
"loss": 0.3836, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 6.542006284497304e-07, |
|
"loss": 0.3827, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 3.681636552324452e-07, |
|
"loss": 0.3846, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 1.6368411956741991e-07, |
|
"loss": 0.4347, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 4.092940597322237e-08, |
|
"loss": 0.3899, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0, |
|
"loss": 0.408, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"eval_loss": 1.2246639728546143, |
|
"eval_runtime": 1.1185, |
|
"eval_samples_per_second": 1.788, |
|
"eval_steps_per_second": 0.894, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"step": 610, |
|
"total_flos": 4.678055725535068e+17, |
|
"train_loss": 1.826815680011374, |
|
"train_runtime": 4235.156, |
|
"train_samples_per_second": 1.156, |
|
"train_steps_per_second": 0.144 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 610, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 100, |
|
"total_flos": 4.678055725535068e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|