|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.40160642570281124, |
|
"eval_steps": 9, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004016064257028112, |
|
"grad_norm": 0.7228589057922363, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.6631, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.004016064257028112, |
|
"eval_loss": 1.0697861909866333, |
|
"eval_runtime": 32.5901, |
|
"eval_samples_per_second": 6.444, |
|
"eval_steps_per_second": 0.828, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.008032128514056224, |
|
"grad_norm": 1.0123487710952759, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.6836, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.012048192771084338, |
|
"grad_norm": 1.005481481552124, |
|
"learning_rate": 6e-06, |
|
"loss": 1.9228, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01606425702811245, |
|
"grad_norm": 0.8270795941352844, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.45, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.020080321285140562, |
|
"grad_norm": 0.9999402165412903, |
|
"learning_rate": 1e-05, |
|
"loss": 2.668, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.024096385542168676, |
|
"grad_norm": 0.9072601199150085, |
|
"learning_rate": 1.2e-05, |
|
"loss": 1.8627, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.028112449799196786, |
|
"grad_norm": 1.1333314180374146, |
|
"learning_rate": 1.4e-05, |
|
"loss": 2.8953, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0321285140562249, |
|
"grad_norm": 0.9936800003051758, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 2.2984, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.03614457831325301, |
|
"grad_norm": 0.7884902954101562, |
|
"learning_rate": 1.8e-05, |
|
"loss": 1.4, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.03614457831325301, |
|
"eval_loss": 1.0618449449539185, |
|
"eval_runtime": 32.5955, |
|
"eval_samples_per_second": 6.443, |
|
"eval_steps_per_second": 0.828, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.040160642570281124, |
|
"grad_norm": 1.1049511432647705, |
|
"learning_rate": 2e-05, |
|
"loss": 2.3974, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04417670682730924, |
|
"grad_norm": 1.0087590217590332, |
|
"learning_rate": 1.999390827019096e-05, |
|
"loss": 2.1244, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.04819277108433735, |
|
"grad_norm": 1.021081805229187, |
|
"learning_rate": 1.9975640502598243e-05, |
|
"loss": 1.6924, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.05220883534136546, |
|
"grad_norm": 0.8872256278991699, |
|
"learning_rate": 1.9945218953682736e-05, |
|
"loss": 1.8039, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.05622489959839357, |
|
"grad_norm": 1.019092321395874, |
|
"learning_rate": 1.9902680687415704e-05, |
|
"loss": 1.8685, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.060240963855421686, |
|
"grad_norm": 1.4794679880142212, |
|
"learning_rate": 1.9848077530122083e-05, |
|
"loss": 2.1413, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0642570281124498, |
|
"grad_norm": 1.2401800155639648, |
|
"learning_rate": 1.9781476007338058e-05, |
|
"loss": 2.1287, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.06827309236947791, |
|
"grad_norm": 1.0547970533370972, |
|
"learning_rate": 1.9702957262759964e-05, |
|
"loss": 1.1896, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.07228915662650602, |
|
"grad_norm": 1.1523879766464233, |
|
"learning_rate": 1.961261695938319e-05, |
|
"loss": 1.5816, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.07228915662650602, |
|
"eval_loss": 0.985172688961029, |
|
"eval_runtime": 32.6746, |
|
"eval_samples_per_second": 6.427, |
|
"eval_steps_per_second": 0.826, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.07630522088353414, |
|
"grad_norm": 1.4140926599502563, |
|
"learning_rate": 1.9510565162951538e-05, |
|
"loss": 1.9035, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.08032128514056225, |
|
"grad_norm": 1.519662618637085, |
|
"learning_rate": 1.9396926207859085e-05, |
|
"loss": 1.9739, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.08433734939759036, |
|
"grad_norm": 0.977906346321106, |
|
"learning_rate": 1.9271838545667876e-05, |
|
"loss": 1.1063, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.08835341365461848, |
|
"grad_norm": 1.301216959953308, |
|
"learning_rate": 1.913545457642601e-05, |
|
"loss": 1.7746, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.09236947791164658, |
|
"grad_norm": 1.1488702297210693, |
|
"learning_rate": 1.8987940462991673e-05, |
|
"loss": 1.5895, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0963855421686747, |
|
"grad_norm": 1.351751685142517, |
|
"learning_rate": 1.8829475928589272e-05, |
|
"loss": 2.5842, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.10040160642570281, |
|
"grad_norm": 1.2297149896621704, |
|
"learning_rate": 1.866025403784439e-05, |
|
"loss": 1.935, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.10441767068273092, |
|
"grad_norm": 1.0193097591400146, |
|
"learning_rate": 1.848048096156426e-05, |
|
"loss": 1.6745, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.10843373493975904, |
|
"grad_norm": 1.5117055177688599, |
|
"learning_rate": 1.8290375725550417e-05, |
|
"loss": 1.2787, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.10843373493975904, |
|
"eval_loss": 0.8428009152412415, |
|
"eval_runtime": 32.6578, |
|
"eval_samples_per_second": 6.43, |
|
"eval_steps_per_second": 0.827, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.11244979919678715, |
|
"grad_norm": 0.8445839881896973, |
|
"learning_rate": 1.8090169943749477e-05, |
|
"loss": 1.6433, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.11646586345381527, |
|
"grad_norm": 1.1348373889923096, |
|
"learning_rate": 1.788010753606722e-05, |
|
"loss": 1.4937, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.12048192771084337, |
|
"grad_norm": 0.9660128355026245, |
|
"learning_rate": 1.766044443118978e-05, |
|
"loss": 1.2509, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.12449799196787148, |
|
"grad_norm": 1.046816110610962, |
|
"learning_rate": 1.7431448254773943e-05, |
|
"loss": 1.6319, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.1285140562248996, |
|
"grad_norm": 1.2746326923370361, |
|
"learning_rate": 1.7193398003386514e-05, |
|
"loss": 1.7526, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.13253012048192772, |
|
"grad_norm": 1.138624906539917, |
|
"learning_rate": 1.6946583704589973e-05, |
|
"loss": 1.5544, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.13654618473895583, |
|
"grad_norm": 1.0225257873535156, |
|
"learning_rate": 1.6691306063588583e-05, |
|
"loss": 1.3283, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.14056224899598393, |
|
"grad_norm": 1.3078521490097046, |
|
"learning_rate": 1.6427876096865394e-05, |
|
"loss": 1.6447, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.14457831325301204, |
|
"grad_norm": 1.1222729682922363, |
|
"learning_rate": 1.6156614753256583e-05, |
|
"loss": 1.482, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.14457831325301204, |
|
"eval_loss": 0.758996844291687, |
|
"eval_runtime": 32.6194, |
|
"eval_samples_per_second": 6.438, |
|
"eval_steps_per_second": 0.828, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.14859437751004015, |
|
"grad_norm": 1.4948545694351196, |
|
"learning_rate": 1.5877852522924733e-05, |
|
"loss": 1.4995, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.15261044176706828, |
|
"grad_norm": 1.0291143655776978, |
|
"learning_rate": 1.5591929034707468e-05, |
|
"loss": 1.2655, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.1566265060240964, |
|
"grad_norm": 1.1123580932617188, |
|
"learning_rate": 1.529919264233205e-05, |
|
"loss": 1.8731, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1606425702811245, |
|
"grad_norm": 0.7641251087188721, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 0.7626, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.1646586345381526, |
|
"grad_norm": 1.2452340126037598, |
|
"learning_rate": 1.469471562785891e-05, |
|
"loss": 1.3199, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.1686746987951807, |
|
"grad_norm": 1.4203380346298218, |
|
"learning_rate": 1.4383711467890776e-05, |
|
"loss": 1.3807, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.17269076305220885, |
|
"grad_norm": 1.2240241765975952, |
|
"learning_rate": 1.4067366430758004e-05, |
|
"loss": 1.3466, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.17670682730923695, |
|
"grad_norm": 1.2436611652374268, |
|
"learning_rate": 1.3746065934159123e-05, |
|
"loss": 1.4975, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.18072289156626506, |
|
"grad_norm": 1.1533771753311157, |
|
"learning_rate": 1.342020143325669e-05, |
|
"loss": 1.409, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.18072289156626506, |
|
"eval_loss": 0.7062954902648926, |
|
"eval_runtime": 32.6227, |
|
"eval_samples_per_second": 6.437, |
|
"eval_steps_per_second": 0.828, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.18473895582329317, |
|
"grad_norm": 1.3506983518600464, |
|
"learning_rate": 1.3090169943749475e-05, |
|
"loss": 1.1306, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.18875502008032127, |
|
"grad_norm": 1.8663361072540283, |
|
"learning_rate": 1.2756373558169992e-05, |
|
"loss": 1.3438, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.1927710843373494, |
|
"grad_norm": 1.1213346719741821, |
|
"learning_rate": 1.2419218955996677e-05, |
|
"loss": 1.3015, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.19678714859437751, |
|
"grad_norm": 1.3179396390914917, |
|
"learning_rate": 1.2079116908177592e-05, |
|
"loss": 1.4581, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.20080321285140562, |
|
"grad_norm": 1.3209939002990723, |
|
"learning_rate": 1.1736481776669307e-05, |
|
"loss": 1.3032, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.20481927710843373, |
|
"grad_norm": 1.0638103485107422, |
|
"learning_rate": 1.1391731009600655e-05, |
|
"loss": 1.3361, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.20883534136546184, |
|
"grad_norm": 1.422438144683838, |
|
"learning_rate": 1.1045284632676535e-05, |
|
"loss": 1.4487, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.21285140562248997, |
|
"grad_norm": 1.2671741247177124, |
|
"learning_rate": 1.0697564737441254e-05, |
|
"loss": 1.3626, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.21686746987951808, |
|
"grad_norm": 1.0545305013656616, |
|
"learning_rate": 1.0348994967025012e-05, |
|
"loss": 1.3192, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.21686746987951808, |
|
"eval_loss": 0.67100989818573, |
|
"eval_runtime": 32.6003, |
|
"eval_samples_per_second": 6.442, |
|
"eval_steps_per_second": 0.828, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.22088353413654618, |
|
"grad_norm": 1.1217244863510132, |
|
"learning_rate": 1e-05, |
|
"loss": 1.0894, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2248995983935743, |
|
"grad_norm": 1.5441569089889526, |
|
"learning_rate": 9.651005032974994e-06, |
|
"loss": 1.5076, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.2289156626506024, |
|
"grad_norm": 1.2194684743881226, |
|
"learning_rate": 9.302435262558748e-06, |
|
"loss": 1.3286, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.23293172690763053, |
|
"grad_norm": 1.2157171964645386, |
|
"learning_rate": 8.954715367323468e-06, |
|
"loss": 1.1355, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.23694779116465864, |
|
"grad_norm": 1.2383450269699097, |
|
"learning_rate": 8.60826899039935e-06, |
|
"loss": 1.3349, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.24096385542168675, |
|
"grad_norm": 1.0101817846298218, |
|
"learning_rate": 8.263518223330698e-06, |
|
"loss": 0.992, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.24497991967871485, |
|
"grad_norm": 1.3267630338668823, |
|
"learning_rate": 7.92088309182241e-06, |
|
"loss": 1.4941, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.24899598393574296, |
|
"grad_norm": 1.0341275930404663, |
|
"learning_rate": 7.580781044003324e-06, |
|
"loss": 0.8775, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.25301204819277107, |
|
"grad_norm": 1.5267117023468018, |
|
"learning_rate": 7.243626441830009e-06, |
|
"loss": 0.9736, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.25301204819277107, |
|
"eval_loss": 0.6493720412254333, |
|
"eval_runtime": 32.5813, |
|
"eval_samples_per_second": 6.445, |
|
"eval_steps_per_second": 0.829, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.2570281124497992, |
|
"grad_norm": 1.3753007650375366, |
|
"learning_rate": 6.909830056250527e-06, |
|
"loss": 1.4487, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.26104417670682734, |
|
"grad_norm": 1.2772159576416016, |
|
"learning_rate": 6.579798566743314e-06, |
|
"loss": 1.0686, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.26506024096385544, |
|
"grad_norm": 1.349326491355896, |
|
"learning_rate": 6.25393406584088e-06, |
|
"loss": 1.209, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.26907630522088355, |
|
"grad_norm": 1.2562471628189087, |
|
"learning_rate": 5.932633569242e-06, |
|
"loss": 1.3394, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.27309236947791166, |
|
"grad_norm": 1.4296122789382935, |
|
"learning_rate": 5.616288532109225e-06, |
|
"loss": 1.3279, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.27710843373493976, |
|
"grad_norm": 1.0856250524520874, |
|
"learning_rate": 5.305284372141095e-06, |
|
"loss": 0.6899, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.28112449799196787, |
|
"grad_norm": 1.3272840976715088, |
|
"learning_rate": 5.000000000000003e-06, |
|
"loss": 1.2571, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.285140562248996, |
|
"grad_norm": 1.5904431343078613, |
|
"learning_rate": 4.700807357667953e-06, |
|
"loss": 1.2008, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.2891566265060241, |
|
"grad_norm": 1.8825539350509644, |
|
"learning_rate": 4.408070965292534e-06, |
|
"loss": 1.9554, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.2891566265060241, |
|
"eval_loss": 0.6365012526512146, |
|
"eval_runtime": 32.5792, |
|
"eval_samples_per_second": 6.446, |
|
"eval_steps_per_second": 0.829, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.2931726907630522, |
|
"grad_norm": 1.4897550344467163, |
|
"learning_rate": 4.12214747707527e-06, |
|
"loss": 1.395, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.2971887550200803, |
|
"grad_norm": 1.3629871606826782, |
|
"learning_rate": 3.8433852467434175e-06, |
|
"loss": 1.3097, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.30120481927710846, |
|
"grad_norm": 1.578955888748169, |
|
"learning_rate": 3.5721239031346067e-06, |
|
"loss": 1.3274, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.30522088353413657, |
|
"grad_norm": 1.4249285459518433, |
|
"learning_rate": 3.308693936411421e-06, |
|
"loss": 1.2748, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.3092369477911647, |
|
"grad_norm": 1.1587796211242676, |
|
"learning_rate": 3.0534162954100264e-06, |
|
"loss": 1.0238, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.3132530120481928, |
|
"grad_norm": 1.302412748336792, |
|
"learning_rate": 2.8066019966134907e-06, |
|
"loss": 1.3392, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.3172690763052209, |
|
"grad_norm": 1.0670229196548462, |
|
"learning_rate": 2.5685517452260566e-06, |
|
"loss": 0.8408, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.321285140562249, |
|
"grad_norm": 1.086983561515808, |
|
"learning_rate": 2.339555568810221e-06, |
|
"loss": 0.6949, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.3253012048192771, |
|
"grad_norm": 1.277090311050415, |
|
"learning_rate": 2.119892463932781e-06, |
|
"loss": 0.8406, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.3253012048192771, |
|
"eval_loss": 0.6291564702987671, |
|
"eval_runtime": 32.5821, |
|
"eval_samples_per_second": 6.445, |
|
"eval_steps_per_second": 0.829, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.3293172690763052, |
|
"grad_norm": 1.555704951286316, |
|
"learning_rate": 1.9098300562505266e-06, |
|
"loss": 1.2604, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 1.5652434825897217, |
|
"learning_rate": 1.709624274449584e-06, |
|
"loss": 1.6627, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.3373493975903614, |
|
"grad_norm": 1.3343106508255005, |
|
"learning_rate": 1.5195190384357405e-06, |
|
"loss": 1.3886, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.3413654618473896, |
|
"grad_norm": 1.630918264389038, |
|
"learning_rate": 1.339745962155613e-06, |
|
"loss": 1.4081, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.3453815261044177, |
|
"grad_norm": 1.3480379581451416, |
|
"learning_rate": 1.1705240714107301e-06, |
|
"loss": 0.8545, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.3493975903614458, |
|
"grad_norm": 1.7904911041259766, |
|
"learning_rate": 1.012059537008332e-06, |
|
"loss": 1.2855, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.3534136546184739, |
|
"grad_norm": 1.5661286115646362, |
|
"learning_rate": 8.645454235739903e-07, |
|
"loss": 1.4537, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.357429718875502, |
|
"grad_norm": 1.5383715629577637, |
|
"learning_rate": 7.281614543321269e-07, |
|
"loss": 1.4315, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.3614457831325301, |
|
"grad_norm": 1.7544879913330078, |
|
"learning_rate": 6.030737921409169e-07, |
|
"loss": 1.5103, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3614457831325301, |
|
"eval_loss": 0.6263256072998047, |
|
"eval_runtime": 32.6143, |
|
"eval_samples_per_second": 6.439, |
|
"eval_steps_per_second": 0.828, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3654618473895582, |
|
"grad_norm": 1.6552611589431763, |
|
"learning_rate": 4.894348370484648e-07, |
|
"loss": 1.7058, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.36947791164658633, |
|
"grad_norm": 1.3507189750671387, |
|
"learning_rate": 3.8738304061681107e-07, |
|
"loss": 0.9932, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.37349397590361444, |
|
"grad_norm": 1.7394344806671143, |
|
"learning_rate": 2.970427372400353e-07, |
|
"loss": 1.4815, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.37751004016064255, |
|
"grad_norm": 1.3886154890060425, |
|
"learning_rate": 2.1852399266194312e-07, |
|
"loss": 0.9411, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.3815261044176707, |
|
"grad_norm": 1.3755466938018799, |
|
"learning_rate": 1.519224698779198e-07, |
|
"loss": 1.3225, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.3855421686746988, |
|
"grad_norm": 1.3697952032089233, |
|
"learning_rate": 9.731931258429638e-08, |
|
"loss": 1.307, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.3895582329317269, |
|
"grad_norm": 1.1580904722213745, |
|
"learning_rate": 5.4781046317267103e-08, |
|
"loss": 0.7828, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.39357429718875503, |
|
"grad_norm": 1.169704794883728, |
|
"learning_rate": 2.4359497401758026e-08, |
|
"loss": 1.1284, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.39759036144578314, |
|
"grad_norm": 1.4925367832183838, |
|
"learning_rate": 6.091729809042379e-09, |
|
"loss": 1.0595, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.39759036144578314, |
|
"eval_loss": 0.6255317330360413, |
|
"eval_runtime": 32.6, |
|
"eval_samples_per_second": 6.442, |
|
"eval_steps_per_second": 0.828, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.40160642570281124, |
|
"grad_norm": 1.5164817571640015, |
|
"learning_rate": 0.0, |
|
"loss": 1.4775, |
|
"step": 100 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.2695579394048e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|