|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.10707784559374665, |
|
"eval_steps": 500, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00010707784559374665, |
|
"grad_norm": 0.4909955859184265, |
|
"learning_rate": 0.00019999950652018584, |
|
"loss": 1.0891, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0002141556911874933, |
|
"grad_norm": 0.37118685245513916, |
|
"learning_rate": 0.0001999980260856137, |
|
"loss": 1.0543, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.00032123353678124, |
|
"grad_norm": 0.3935191035270691, |
|
"learning_rate": 0.000199995558710895, |
|
"loss": 1.0476, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0004283113823749866, |
|
"grad_norm": 0.47593897581100464, |
|
"learning_rate": 0.00019999210442038162, |
|
"loss": 0.9218, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0005353892279687332, |
|
"grad_norm": 0.4176044166088104, |
|
"learning_rate": 0.00019998766324816607, |
|
"loss": 0.8016, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.00064246707356248, |
|
"grad_norm": 0.45231443643569946, |
|
"learning_rate": 0.0001999822352380809, |
|
"loss": 0.7248, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0007495449191562266, |
|
"grad_norm": 0.4732208251953125, |
|
"learning_rate": 0.00019997582044369843, |
|
"loss": 0.6483, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0008566227647499732, |
|
"grad_norm": 0.4255562424659729, |
|
"learning_rate": 0.00019996841892833, |
|
"loss": 0.5931, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0009637006103437199, |
|
"grad_norm": 0.5560129284858704, |
|
"learning_rate": 0.00019996003076502565, |
|
"loss": 0.5045, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0010707784559374665, |
|
"grad_norm": 0.5253703594207764, |
|
"learning_rate": 0.00019995065603657316, |
|
"loss": 0.3965, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.001177856301531213, |
|
"grad_norm": 0.4471832513809204, |
|
"learning_rate": 0.0001999402948354973, |
|
"loss": 0.5783, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.00128493414712496, |
|
"grad_norm": 0.4470554292201996, |
|
"learning_rate": 0.00019992894726405893, |
|
"loss": 0.4207, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0013920119927187066, |
|
"grad_norm": 0.42513272166252136, |
|
"learning_rate": 0.000199916613434254, |
|
"loss": 0.4833, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0014990898383124532, |
|
"grad_norm": 0.34535884857177734, |
|
"learning_rate": 0.0001999032934678125, |
|
"loss": 0.3717, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0016061676839061998, |
|
"grad_norm": 0.33447375893592834, |
|
"learning_rate": 0.00019988898749619702, |
|
"loss": 0.3569, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0017132455294999465, |
|
"grad_norm": 0.50462406873703, |
|
"learning_rate": 0.00019987369566060176, |
|
"loss": 0.3444, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.001820323375093693, |
|
"grad_norm": 0.3608594834804535, |
|
"learning_rate": 0.00019985741811195097, |
|
"loss": 0.4535, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0019274012206874397, |
|
"grad_norm": 0.289742648601532, |
|
"learning_rate": 0.00019984015501089752, |
|
"loss": 0.3921, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0020344790662811863, |
|
"grad_norm": 0.3383671045303345, |
|
"learning_rate": 0.0001998219065278212, |
|
"loss": 0.4304, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.002141556911874933, |
|
"grad_norm": 0.3464438319206238, |
|
"learning_rate": 0.00019980267284282717, |
|
"loss": 0.3887, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0022486347574686796, |
|
"grad_norm": 0.3533989191055298, |
|
"learning_rate": 0.00019978245414574417, |
|
"loss": 0.3861, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.002355712603062426, |
|
"grad_norm": 0.3390829563140869, |
|
"learning_rate": 0.00019976125063612252, |
|
"loss": 0.3445, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.002462790448656173, |
|
"grad_norm": 0.23587378859519958, |
|
"learning_rate": 0.00019973906252323238, |
|
"loss": 0.2955, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.00256986829424992, |
|
"grad_norm": 0.42040207982063293, |
|
"learning_rate": 0.0001997158900260614, |
|
"loss": 0.4963, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.0026769461398436665, |
|
"grad_norm": 0.25572434067726135, |
|
"learning_rate": 0.0001996917333733128, |
|
"loss": 0.359, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.002784023985437413, |
|
"grad_norm": 0.44296184182167053, |
|
"learning_rate": 0.00019966659280340297, |
|
"loss": 0.366, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.0028911018310311598, |
|
"grad_norm": 0.2594900131225586, |
|
"learning_rate": 0.00019964046856445924, |
|
"loss": 0.2909, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0029981796766249064, |
|
"grad_norm": 0.2581496238708496, |
|
"learning_rate": 0.00019961336091431727, |
|
"loss": 0.2962, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.003105257522218653, |
|
"grad_norm": 0.29349207878112793, |
|
"learning_rate": 0.00019958527012051857, |
|
"loss": 0.2998, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.0032123353678123997, |
|
"grad_norm": 0.30551677942276, |
|
"learning_rate": 0.00019955619646030802, |
|
"loss": 0.4177, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0033194132134061463, |
|
"grad_norm": 0.22254447638988495, |
|
"learning_rate": 0.00019952614022063084, |
|
"loss": 0.3558, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.003426491058999893, |
|
"grad_norm": 0.2150162309408188, |
|
"learning_rate": 0.00019949510169813003, |
|
"loss": 0.376, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.0035335689045936395, |
|
"grad_norm": 0.2759721577167511, |
|
"learning_rate": 0.00019946308119914323, |
|
"loss": 0.3295, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.003640646750187386, |
|
"grad_norm": 0.2623266875743866, |
|
"learning_rate": 0.0001994300790396999, |
|
"loss": 0.3923, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.0037477245957811328, |
|
"grad_norm": 0.24647563695907593, |
|
"learning_rate": 0.000199396095545518, |
|
"loss": 0.3683, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0038548024413748794, |
|
"grad_norm": 0.20459140837192535, |
|
"learning_rate": 0.00019936113105200085, |
|
"loss": 0.2112, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.003961880286968626, |
|
"grad_norm": 0.23986490070819855, |
|
"learning_rate": 0.00019932518590423394, |
|
"loss": 0.3532, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.004068958132562373, |
|
"grad_norm": 0.21777169406414032, |
|
"learning_rate": 0.00019928826045698136, |
|
"loss": 0.3148, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.004176035978156119, |
|
"grad_norm": 0.2896265685558319, |
|
"learning_rate": 0.0001992503550746824, |
|
"loss": 0.3149, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.004283113823749866, |
|
"grad_norm": 0.24791480600833893, |
|
"learning_rate": 0.0001992114701314478, |
|
"loss": 0.3459, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0043901916693436125, |
|
"grad_norm": 0.25446805357933044, |
|
"learning_rate": 0.0001991716060110563, |
|
"loss": 0.3023, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.004497269514937359, |
|
"grad_norm": 0.22087767720222473, |
|
"learning_rate": 0.00019913076310695068, |
|
"loss": 0.3853, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.004604347360531106, |
|
"grad_norm": 0.2717120945453644, |
|
"learning_rate": 0.00019908894182223388, |
|
"loss": 0.3354, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.004711425206124852, |
|
"grad_norm": 0.20693597197532654, |
|
"learning_rate": 0.00019904614256966512, |
|
"loss": 0.3419, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.004818503051718599, |
|
"grad_norm": 0.21952880918979645, |
|
"learning_rate": 0.00019900236577165576, |
|
"loss": 0.2945, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.004925580897312346, |
|
"grad_norm": 0.2796410620212555, |
|
"learning_rate": 0.0001989576118602651, |
|
"loss": 0.3456, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.005032658742906092, |
|
"grad_norm": 0.21983903646469116, |
|
"learning_rate": 0.00019891188127719618, |
|
"loss": 0.3203, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.00513973658849984, |
|
"grad_norm": 0.1815391182899475, |
|
"learning_rate": 0.0001988651744737914, |
|
"loss": 0.2796, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.005246814434093586, |
|
"grad_norm": 0.2382379174232483, |
|
"learning_rate": 0.00019881749191102808, |
|
"loss": 0.3825, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.005353892279687333, |
|
"grad_norm": 0.19003333151340485, |
|
"learning_rate": 0.00019876883405951377, |
|
"loss": 0.3344, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.00546097012528108, |
|
"grad_norm": 0.2239263653755188, |
|
"learning_rate": 0.00019871920139948192, |
|
"loss": 0.4143, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.005568047970874826, |
|
"grad_norm": 0.1837572157382965, |
|
"learning_rate": 0.0001986685944207868, |
|
"loss": 0.3166, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.005675125816468573, |
|
"grad_norm": 0.22199811041355133, |
|
"learning_rate": 0.0001986170136228989, |
|
"loss": 0.3639, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.0057822036620623195, |
|
"grad_norm": 0.19471023976802826, |
|
"learning_rate": 0.00019856445951489982, |
|
"loss": 0.3314, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.005889281507656066, |
|
"grad_norm": 0.26663461327552795, |
|
"learning_rate": 0.0001985109326154774, |
|
"loss": 0.4204, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.005996359353249813, |
|
"grad_norm": 0.20248642563819885, |
|
"learning_rate": 0.00019845643345292054, |
|
"loss": 0.3668, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.006103437198843559, |
|
"grad_norm": 0.34333622455596924, |
|
"learning_rate": 0.00019840096256511398, |
|
"loss": 0.4426, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.006210515044437306, |
|
"grad_norm": 0.226885586977005, |
|
"learning_rate": 0.00019834452049953297, |
|
"loss": 0.3298, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.006317592890031053, |
|
"grad_norm": 0.2704862952232361, |
|
"learning_rate": 0.00019828710781323792, |
|
"loss": 0.4359, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.006424670735624799, |
|
"grad_norm": 0.24985526502132416, |
|
"learning_rate": 0.0001982287250728689, |
|
"loss": 0.4191, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.006531748581218546, |
|
"grad_norm": 0.1854284405708313, |
|
"learning_rate": 0.0001981693728546399, |
|
"loss": 0.3131, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.0066388264268122926, |
|
"grad_norm": 0.21722164750099182, |
|
"learning_rate": 0.0001981090517443334, |
|
"loss": 0.3055, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.006745904272406039, |
|
"grad_norm": 0.231235533952713, |
|
"learning_rate": 0.00019804776233729444, |
|
"loss": 0.3345, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.006852982117999786, |
|
"grad_norm": 0.22202789783477783, |
|
"learning_rate": 0.0001979855052384247, |
|
"loss": 0.3525, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.0069600599635935324, |
|
"grad_norm": 0.21439769864082336, |
|
"learning_rate": 0.00019792228106217658, |
|
"loss": 0.3413, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.007067137809187279, |
|
"grad_norm": 0.18203051388263702, |
|
"learning_rate": 0.00019785809043254722, |
|
"loss": 0.3312, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.007174215654781026, |
|
"grad_norm": 0.22350089251995087, |
|
"learning_rate": 0.0001977929339830722, |
|
"loss": 0.4871, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.007281293500374772, |
|
"grad_norm": 0.19005514681339264, |
|
"learning_rate": 0.00019772681235681936, |
|
"loss": 0.4379, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.007388371345968519, |
|
"grad_norm": 0.18949486315250397, |
|
"learning_rate": 0.00019765972620638248, |
|
"loss": 0.3185, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.0074954491915622656, |
|
"grad_norm": 0.1703757345676422, |
|
"learning_rate": 0.00019759167619387476, |
|
"loss": 0.3595, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.007602527037156012, |
|
"grad_norm": 0.27864259481430054, |
|
"learning_rate": 0.00019752266299092236, |
|
"loss": 0.3437, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.007709604882749759, |
|
"grad_norm": 0.17867398262023926, |
|
"learning_rate": 0.00019745268727865774, |
|
"loss": 0.3299, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.007816682728343505, |
|
"grad_norm": 0.1801465004682541, |
|
"learning_rate": 0.0001973817497477129, |
|
"loss": 0.2722, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.007923760573937252, |
|
"grad_norm": 0.16662195324897766, |
|
"learning_rate": 0.00019730985109821266, |
|
"loss": 0.3794, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.008030838419530999, |
|
"grad_norm": 0.1943478137254715, |
|
"learning_rate": 0.00019723699203976766, |
|
"loss": 0.3975, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.008137916265124745, |
|
"grad_norm": 0.1636917144060135, |
|
"learning_rate": 0.0001971631732914674, |
|
"loss": 0.3018, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.008244994110718492, |
|
"grad_norm": 0.1836482137441635, |
|
"learning_rate": 0.0001970883955818731, |
|
"loss": 0.2692, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.008352071956312239, |
|
"grad_norm": 0.17146596312522888, |
|
"learning_rate": 0.0001970126596490106, |
|
"loss": 0.271, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.008459149801905985, |
|
"grad_norm": 0.215361550450325, |
|
"learning_rate": 0.00019693596624036292, |
|
"loss": 0.4033, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.008566227647499732, |
|
"grad_norm": 0.16516317427158356, |
|
"learning_rate": 0.0001968583161128631, |
|
"loss": 0.3614, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.008673305493093478, |
|
"grad_norm": 0.24210558831691742, |
|
"learning_rate": 0.00019677971003288655, |
|
"loss": 0.5388, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.008780383338687225, |
|
"grad_norm": 0.16257977485656738, |
|
"learning_rate": 0.00019670014877624353, |
|
"loss": 0.3109, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.008887461184280972, |
|
"grad_norm": 0.26173606514930725, |
|
"learning_rate": 0.00019661963312817148, |
|
"loss": 0.389, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.008994539029874718, |
|
"grad_norm": 0.19191040098667145, |
|
"learning_rate": 0.0001965381638833274, |
|
"loss": 0.3695, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.009101616875468465, |
|
"grad_norm": 0.1762268841266632, |
|
"learning_rate": 0.00019645574184577982, |
|
"loss": 0.3374, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.009208694721062212, |
|
"grad_norm": 0.18683938682079315, |
|
"learning_rate": 0.000196372367829001, |
|
"loss": 0.3423, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.009315772566655958, |
|
"grad_norm": 0.17973066866397858, |
|
"learning_rate": 0.00019628804265585877, |
|
"loss": 0.398, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.009422850412249705, |
|
"grad_norm": 0.16043399274349213, |
|
"learning_rate": 0.0001962027671586086, |
|
"loss": 0.2939, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.009529928257843451, |
|
"grad_norm": 0.17222130298614502, |
|
"learning_rate": 0.0001961165421788852, |
|
"loss": 0.3419, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.009637006103437198, |
|
"grad_norm": 0.1862293928861618, |
|
"learning_rate": 0.0001960293685676943, |
|
"loss": 0.2944, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.009744083949030945, |
|
"grad_norm": 0.22960934042930603, |
|
"learning_rate": 0.0001959412471854043, |
|
"loss": 0.4391, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.009851161794624691, |
|
"grad_norm": 0.16783665120601654, |
|
"learning_rate": 0.0001958521789017376, |
|
"loss": 0.3055, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.009958239640218438, |
|
"grad_norm": 0.1998833566904068, |
|
"learning_rate": 0.00019576216459576222, |
|
"loss": 0.351, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.010065317485812185, |
|
"grad_norm": 0.19366595149040222, |
|
"learning_rate": 0.00019567120515588308, |
|
"loss": 0.4888, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.010172395331405933, |
|
"grad_norm": 0.20969517529010773, |
|
"learning_rate": 0.00019557930147983302, |
|
"loss": 0.429, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.01027947317699968, |
|
"grad_norm": 0.17586903274059296, |
|
"learning_rate": 0.00019548645447466431, |
|
"loss": 0.2981, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.010386551022593426, |
|
"grad_norm": 0.2091716229915619, |
|
"learning_rate": 0.00019539266505673938, |
|
"loss": 0.3118, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.010493628868187173, |
|
"grad_norm": 0.2315663993358612, |
|
"learning_rate": 0.00019529793415172192, |
|
"loss": 0.4724, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.01060070671378092, |
|
"grad_norm": 0.19753670692443848, |
|
"learning_rate": 0.00019520226269456768, |
|
"loss": 0.3185, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.010707784559374666, |
|
"grad_norm": 0.1516932249069214, |
|
"learning_rate": 0.00019510565162951537, |
|
"loss": 0.2947, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.010814862404968413, |
|
"grad_norm": 0.1840781718492508, |
|
"learning_rate": 0.00019500810191007718, |
|
"loss": 0.3685, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.01092194025056216, |
|
"grad_norm": 0.1962490826845169, |
|
"learning_rate": 0.00019490961449902946, |
|
"loss": 0.3657, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.011029018096155906, |
|
"grad_norm": 0.22264431416988373, |
|
"learning_rate": 0.0001948101903684032, |
|
"loss": 0.4791, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.011136095941749653, |
|
"grad_norm": 0.17540693283081055, |
|
"learning_rate": 0.00019470983049947444, |
|
"loss": 0.3603, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.0112431737873434, |
|
"grad_norm": 0.1619989275932312, |
|
"learning_rate": 0.00019460853588275454, |
|
"loss": 0.3965, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.011350251632937146, |
|
"grad_norm": 0.17258454859256744, |
|
"learning_rate": 0.00019450630751798048, |
|
"loss": 0.2544, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.011457329478530892, |
|
"grad_norm": 0.23875324428081512, |
|
"learning_rate": 0.000194403146414105, |
|
"loss": 0.3889, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.011564407324124639, |
|
"grad_norm": 0.17445383965969086, |
|
"learning_rate": 0.00019429905358928646, |
|
"loss": 0.3413, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.011671485169718386, |
|
"grad_norm": 0.21563485264778137, |
|
"learning_rate": 0.00019419403007087907, |
|
"loss": 0.3474, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.011778563015312132, |
|
"grad_norm": 0.15843790769577026, |
|
"learning_rate": 0.00019408807689542257, |
|
"loss": 0.3574, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.011885640860905879, |
|
"grad_norm": 0.2041575163602829, |
|
"learning_rate": 0.00019398119510863197, |
|
"loss": 0.4127, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.011992718706499626, |
|
"grad_norm": 0.16513045132160187, |
|
"learning_rate": 0.00019387338576538744, |
|
"loss": 0.2633, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.012099796552093372, |
|
"grad_norm": 0.18099965155124664, |
|
"learning_rate": 0.00019376464992972356, |
|
"loss": 0.3851, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.012206874397687119, |
|
"grad_norm": 0.22513428330421448, |
|
"learning_rate": 0.00019365498867481923, |
|
"loss": 0.4314, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.012313952243280865, |
|
"grad_norm": 0.15893103182315826, |
|
"learning_rate": 0.00019354440308298675, |
|
"loss": 0.2835, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.012421030088874612, |
|
"grad_norm": 0.17109234631061554, |
|
"learning_rate": 0.00019343289424566122, |
|
"loss": 0.4174, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.012528107934468359, |
|
"grad_norm": 0.13781225681304932, |
|
"learning_rate": 0.00019332046326338986, |
|
"loss": 0.2647, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.012635185780062105, |
|
"grad_norm": 0.16822676360607147, |
|
"learning_rate": 0.0001932071112458211, |
|
"loss": 0.3348, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.012742263625655852, |
|
"grad_norm": 0.17825409770011902, |
|
"learning_rate": 0.00019309283931169356, |
|
"loss": 0.3831, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.012849341471249599, |
|
"grad_norm": 0.2027072161436081, |
|
"learning_rate": 0.00019297764858882514, |
|
"loss": 0.4212, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.012956419316843345, |
|
"grad_norm": 0.17872022092342377, |
|
"learning_rate": 0.00019286154021410173, |
|
"loss": 0.32, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.013063497162437092, |
|
"grad_norm": 0.19515572488307953, |
|
"learning_rate": 0.00019274451533346615, |
|
"loss": 0.4162, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.013170575008030838, |
|
"grad_norm": 0.16127873957157135, |
|
"learning_rate": 0.00019262657510190666, |
|
"loss": 0.2789, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.013277652853624585, |
|
"grad_norm": 0.16778552532196045, |
|
"learning_rate": 0.0001925077206834458, |
|
"loss": 0.3651, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.013384730699218332, |
|
"grad_norm": 0.1893121302127838, |
|
"learning_rate": 0.0001923879532511287, |
|
"loss": 0.381, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.013491808544812078, |
|
"grad_norm": 0.17123256623744965, |
|
"learning_rate": 0.0001922672739870115, |
|
"loss": 0.2907, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.013598886390405825, |
|
"grad_norm": 0.19237865507602692, |
|
"learning_rate": 0.00019214568408214985, |
|
"loss": 0.3571, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.013705964235999572, |
|
"grad_norm": 0.20815183222293854, |
|
"learning_rate": 0.00019202318473658705, |
|
"loss": 0.4195, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.013813042081593318, |
|
"grad_norm": 0.19400858879089355, |
|
"learning_rate": 0.00019189977715934213, |
|
"loss": 0.443, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.013920119927187065, |
|
"grad_norm": 0.20479348301887512, |
|
"learning_rate": 0.00019177546256839812, |
|
"loss": 0.3082, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.014027197772780811, |
|
"grad_norm": 0.20510035753250122, |
|
"learning_rate": 0.0001916502421906898, |
|
"loss": 0.3948, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.014134275618374558, |
|
"grad_norm": 0.18292556703090668, |
|
"learning_rate": 0.00019152411726209176, |
|
"loss": 0.2958, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.014241353463968305, |
|
"grad_norm": 0.17402611672878265, |
|
"learning_rate": 0.00019139708902740613, |
|
"loss": 0.3426, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.014348431309562051, |
|
"grad_norm": 0.1712314933538437, |
|
"learning_rate": 0.0001912691587403503, |
|
"loss": 0.3618, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.014455509155155798, |
|
"grad_norm": 0.15368402004241943, |
|
"learning_rate": 0.00019114032766354453, |
|
"loss": 0.2978, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.014562587000749545, |
|
"grad_norm": 0.2001761794090271, |
|
"learning_rate": 0.00019101059706849957, |
|
"loss": 0.3388, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.014669664846343291, |
|
"grad_norm": 0.17164276540279388, |
|
"learning_rate": 0.00019087996823560402, |
|
"loss": 0.3974, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.014776742691937038, |
|
"grad_norm": 0.15010999143123627, |
|
"learning_rate": 0.0001907484424541117, |
|
"loss": 0.2875, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.014883820537530785, |
|
"grad_norm": 0.19477848708629608, |
|
"learning_rate": 0.00019061602102212898, |
|
"loss": 0.4388, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.014990898383124531, |
|
"grad_norm": 0.18081682920455933, |
|
"learning_rate": 0.00019048270524660196, |
|
"loss": 0.3336, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.015097976228718278, |
|
"grad_norm": 0.22473369538784027, |
|
"learning_rate": 0.0001903484964433035, |
|
"loss": 0.294, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.015205054074312024, |
|
"grad_norm": 0.16706225275993347, |
|
"learning_rate": 0.00019021339593682028, |
|
"loss": 0.302, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.015312131919905771, |
|
"grad_norm": 0.15441536903381348, |
|
"learning_rate": 0.00019007740506053983, |
|
"loss": 0.309, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.015419209765499518, |
|
"grad_norm": 0.1709998995065689, |
|
"learning_rate": 0.0001899405251566371, |
|
"loss": 0.3257, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.015526287611093264, |
|
"grad_norm": 0.18251051008701324, |
|
"learning_rate": 0.00018980275757606157, |
|
"loss": 0.3423, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.01563336545668701, |
|
"grad_norm": 0.155283123254776, |
|
"learning_rate": 0.00018966410367852362, |
|
"loss": 0.2204, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.01574044330228076, |
|
"grad_norm": 0.19766977429389954, |
|
"learning_rate": 0.00018952456483248119, |
|
"loss": 0.4286, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.015847521147874504, |
|
"grad_norm": 0.17703601717948914, |
|
"learning_rate": 0.0001893841424151264, |
|
"loss": 0.276, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.015954598993468252, |
|
"grad_norm": 0.178870290517807, |
|
"learning_rate": 0.0001892428378123718, |
|
"loss": 0.3468, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.016061676839061997, |
|
"grad_norm": 0.17410863935947418, |
|
"learning_rate": 0.0001891006524188368, |
|
"loss": 0.3231, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.016168754684655746, |
|
"grad_norm": 0.1725064218044281, |
|
"learning_rate": 0.00018895758763783383, |
|
"loss": 0.3679, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.01627583253024949, |
|
"grad_norm": 0.194014772772789, |
|
"learning_rate": 0.00018881364488135448, |
|
"loss": 0.3218, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.01638291037584324, |
|
"grad_norm": 0.16831275820732117, |
|
"learning_rate": 0.00018866882557005567, |
|
"loss": 0.338, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.016489988221436984, |
|
"grad_norm": 0.14952607452869415, |
|
"learning_rate": 0.00018852313113324552, |
|
"loss": 0.2548, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.016597066067030732, |
|
"grad_norm": 0.1442236751317978, |
|
"learning_rate": 0.00018837656300886937, |
|
"loss": 0.2202, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.016704143912624477, |
|
"grad_norm": 0.1802881807088852, |
|
"learning_rate": 0.00018822912264349534, |
|
"loss": 0.4736, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.016811221758218226, |
|
"grad_norm": 0.2469218522310257, |
|
"learning_rate": 0.00018808081149230036, |
|
"loss": 0.4753, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.01691829960381197, |
|
"grad_norm": 0.16127042472362518, |
|
"learning_rate": 0.00018793163101905563, |
|
"loss": 0.3649, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.01702537744940572, |
|
"grad_norm": 0.17643609642982483, |
|
"learning_rate": 0.00018778158269611218, |
|
"loss": 0.311, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.017132455294999464, |
|
"grad_norm": 0.14460398256778717, |
|
"learning_rate": 0.00018763066800438636, |
|
"loss": 0.2213, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.017239533140593212, |
|
"grad_norm": 0.15388457477092743, |
|
"learning_rate": 0.0001874788884333453, |
|
"loss": 0.3528, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.017346610986186957, |
|
"grad_norm": 0.15638676285743713, |
|
"learning_rate": 0.00018732624548099204, |
|
"loss": 0.3714, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.017453688831780705, |
|
"grad_norm": 0.1833023875951767, |
|
"learning_rate": 0.0001871727406538509, |
|
"loss": 0.3702, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.01756076667737445, |
|
"grad_norm": 0.24797490239143372, |
|
"learning_rate": 0.0001870183754669526, |
|
"loss": 0.3922, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.0176678445229682, |
|
"grad_norm": 0.1474548578262329, |
|
"learning_rate": 0.00018686315144381913, |
|
"loss": 0.2486, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.017774922368561943, |
|
"grad_norm": 0.19909460842609406, |
|
"learning_rate": 0.000186707070116449, |
|
"loss": 0.3681, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.017882000214155692, |
|
"grad_norm": 0.15004006028175354, |
|
"learning_rate": 0.0001865501330253019, |
|
"loss": 0.3179, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.017989078059749437, |
|
"grad_norm": 0.2146318256855011, |
|
"learning_rate": 0.00018639234171928353, |
|
"loss": 0.3937, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.018096155905343185, |
|
"grad_norm": 0.17869114875793457, |
|
"learning_rate": 0.0001862336977557304, |
|
"loss": 0.4543, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.01820323375093693, |
|
"grad_norm": 0.17866748571395874, |
|
"learning_rate": 0.0001860742027003944, |
|
"loss": 0.3069, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.01831031159653068, |
|
"grad_norm": 0.17818759381771088, |
|
"learning_rate": 0.00018591385812742725, |
|
"loss": 0.2896, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.018417389442124423, |
|
"grad_norm": 0.18158233165740967, |
|
"learning_rate": 0.00018575266561936523, |
|
"loss": 0.3399, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.01852446728771817, |
|
"grad_norm": 0.17146669328212738, |
|
"learning_rate": 0.00018559062676711332, |
|
"loss": 0.2842, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.018631545133311916, |
|
"grad_norm": 0.2015829235315323, |
|
"learning_rate": 0.0001854277431699295, |
|
"loss": 0.3274, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.018738622978905665, |
|
"grad_norm": 0.18138733506202698, |
|
"learning_rate": 0.00018526401643540922, |
|
"loss": 0.3509, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.01884570082449941, |
|
"grad_norm": 0.18845680356025696, |
|
"learning_rate": 0.00018509944817946922, |
|
"loss": 0.4202, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.018952778670093158, |
|
"grad_norm": 0.19233529269695282, |
|
"learning_rate": 0.00018493404002633166, |
|
"loss": 0.4569, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.019059856515686903, |
|
"grad_norm": 0.1837352216243744, |
|
"learning_rate": 0.00018476779360850832, |
|
"loss": 0.4374, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.01916693436128065, |
|
"grad_norm": 0.19979973137378693, |
|
"learning_rate": 0.00018460071056678422, |
|
"loss": 0.4437, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.019274012206874396, |
|
"grad_norm": 0.3244735598564148, |
|
"learning_rate": 0.00018443279255020152, |
|
"loss": 0.3831, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.019381090052468145, |
|
"grad_norm": 0.21024329960346222, |
|
"learning_rate": 0.00018426404121604323, |
|
"loss": 0.4677, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.01948816789806189, |
|
"grad_norm": 0.22042858600616455, |
|
"learning_rate": 0.00018409445822981693, |
|
"loss": 0.3745, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.019595245743655638, |
|
"grad_norm": 0.17360983788967133, |
|
"learning_rate": 0.00018392404526523817, |
|
"loss": 0.3558, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.019702323589249383, |
|
"grad_norm": 0.1609887331724167, |
|
"learning_rate": 0.0001837528040042142, |
|
"loss": 0.2913, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.01980940143484313, |
|
"grad_norm": 0.17246180772781372, |
|
"learning_rate": 0.00018358073613682706, |
|
"loss": 0.3632, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.019916479280436876, |
|
"grad_norm": 0.24439960718154907, |
|
"learning_rate": 0.00018340784336131713, |
|
"loss": 0.423, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.020023557126030624, |
|
"grad_norm": 0.17695844173431396, |
|
"learning_rate": 0.00018323412738406635, |
|
"loss": 0.3018, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.02013063497162437, |
|
"grad_norm": 0.18329234421253204, |
|
"learning_rate": 0.00018305958991958127, |
|
"loss": 0.4078, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.020237712817218118, |
|
"grad_norm": 0.1552525758743286, |
|
"learning_rate": 0.0001828842326904762, |
|
"loss": 0.2596, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.020344790662811866, |
|
"grad_norm": 0.15624456107616425, |
|
"learning_rate": 0.00018270805742745617, |
|
"loss": 0.2375, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.02045186850840561, |
|
"grad_norm": 0.19527368247509003, |
|
"learning_rate": 0.00018253106586929997, |
|
"loss": 0.4228, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.02055894635399936, |
|
"grad_norm": 0.1665705144405365, |
|
"learning_rate": 0.00018235325976284275, |
|
"loss": 0.2786, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.020666024199593104, |
|
"grad_norm": 0.15713033080101013, |
|
"learning_rate": 0.00018217464086295904, |
|
"loss": 0.2833, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.020773102045186852, |
|
"grad_norm": 0.15957246720790863, |
|
"learning_rate": 0.00018199521093254523, |
|
"loss": 0.2854, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.020880179890780597, |
|
"grad_norm": 0.20240135490894318, |
|
"learning_rate": 0.00018181497174250236, |
|
"loss": 0.342, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.020987257736374346, |
|
"grad_norm": 0.21210376918315887, |
|
"learning_rate": 0.00018163392507171842, |
|
"loss": 0.3008, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.02109433558196809, |
|
"grad_norm": 0.1445528119802475, |
|
"learning_rate": 0.00018145207270705096, |
|
"loss": 0.2464, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.02120141342756184, |
|
"grad_norm": 0.1597689390182495, |
|
"learning_rate": 0.0001812694164433094, |
|
"loss": 0.3196, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.021308491273155584, |
|
"grad_norm": 0.1642187237739563, |
|
"learning_rate": 0.00018108595808323736, |
|
"loss": 0.2816, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.021415569118749332, |
|
"grad_norm": 0.17551757395267487, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 0.3887, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.021522646964343077, |
|
"grad_norm": 0.18966630101203918, |
|
"learning_rate": 0.00018071664232464002, |
|
"loss": 0.4184, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.021629724809936825, |
|
"grad_norm": 0.20549455285072327, |
|
"learning_rate": 0.0001805307885711122, |
|
"loss": 0.3353, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.02173680265553057, |
|
"grad_norm": 0.16554607450962067, |
|
"learning_rate": 0.00018034414001121278, |
|
"loss": 0.4082, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.02184388050112432, |
|
"grad_norm": 0.18681012094020844, |
|
"learning_rate": 0.00018015669848708767, |
|
"loss": 0.2714, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.021950958346718064, |
|
"grad_norm": 0.18294884264469147, |
|
"learning_rate": 0.00017996846584870908, |
|
"loss": 0.3567, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.022058036192311812, |
|
"grad_norm": 0.1881534308195114, |
|
"learning_rate": 0.0001797794439538571, |
|
"loss": 0.3717, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.022165114037905557, |
|
"grad_norm": 0.16362693905830383, |
|
"learning_rate": 0.0001795896346681016, |
|
"loss": 0.3797, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.022272191883499305, |
|
"grad_norm": 0.16024716198444366, |
|
"learning_rate": 0.00017939903986478355, |
|
"loss": 0.2268, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.02237926972909305, |
|
"grad_norm": 0.1983472853899002, |
|
"learning_rate": 0.00017920766142499672, |
|
"loss": 0.318, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.0224863475746868, |
|
"grad_norm": 0.1842176765203476, |
|
"learning_rate": 0.00017901550123756906, |
|
"loss": 0.3663, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.022593425420280543, |
|
"grad_norm": 0.21368902921676636, |
|
"learning_rate": 0.00017882256119904403, |
|
"loss": 0.4759, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.02270050326587429, |
|
"grad_norm": 0.1486794352531433, |
|
"learning_rate": 0.00017862884321366188, |
|
"loss": 0.3391, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.022807581111468037, |
|
"grad_norm": 0.19507035613059998, |
|
"learning_rate": 0.000178434349193341, |
|
"loss": 0.3386, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.022914658957061785, |
|
"grad_norm": 0.18279972672462463, |
|
"learning_rate": 0.0001782390810576588, |
|
"loss": 0.3325, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.02302173680265553, |
|
"grad_norm": 0.18135708570480347, |
|
"learning_rate": 0.000178043040733833, |
|
"loss": 0.3102, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.023128814648249278, |
|
"grad_norm": 0.2114657461643219, |
|
"learning_rate": 0.00017784623015670238, |
|
"loss": 0.4783, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.023235892493843023, |
|
"grad_norm": 0.1833896040916443, |
|
"learning_rate": 0.00017764865126870786, |
|
"loss": 0.3528, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.02334297033943677, |
|
"grad_norm": 0.15574641525745392, |
|
"learning_rate": 0.00017745030601987337, |
|
"loss": 0.3342, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.023450048185030516, |
|
"grad_norm": 0.14896464347839355, |
|
"learning_rate": 0.00017725119636778644, |
|
"loss": 0.1749, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.023557126030624265, |
|
"grad_norm": 0.13782522082328796, |
|
"learning_rate": 0.00017705132427757895, |
|
"loss": 0.2891, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.02366420387621801, |
|
"grad_norm": 0.18487310409545898, |
|
"learning_rate": 0.00017685069172190766, |
|
"loss": 0.3989, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.023771281721811758, |
|
"grad_norm": 0.16647568345069885, |
|
"learning_rate": 0.00017664930068093498, |
|
"loss": 0.3998, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.023878359567405503, |
|
"grad_norm": 0.1517525464296341, |
|
"learning_rate": 0.00017644715314230918, |
|
"loss": 0.3215, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.02398543741299925, |
|
"grad_norm": 0.16107575595378876, |
|
"learning_rate": 0.0001762442511011448, |
|
"loss": 0.3801, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.024092515258592996, |
|
"grad_norm": 0.18223942816257477, |
|
"learning_rate": 0.0001760405965600031, |
|
"loss": 0.4009, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.024199593104186744, |
|
"grad_norm": 0.12823620438575745, |
|
"learning_rate": 0.0001758361915288722, |
|
"loss": 0.2689, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.02430667094978049, |
|
"grad_norm": 0.1777333915233612, |
|
"learning_rate": 0.0001756310380251472, |
|
"loss": 0.3016, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.024413748795374238, |
|
"grad_norm": 0.20274615287780762, |
|
"learning_rate": 0.00017542513807361037, |
|
"loss": 0.3624, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.024520826640967983, |
|
"grad_norm": 0.4989624619483948, |
|
"learning_rate": 0.00017521849370641114, |
|
"loss": 0.237, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.02462790448656173, |
|
"grad_norm": 0.1811639964580536, |
|
"learning_rate": 0.00017501110696304596, |
|
"loss": 0.3738, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.024734982332155476, |
|
"grad_norm": 0.2014167159795761, |
|
"learning_rate": 0.00017480297989033825, |
|
"loss": 0.3726, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.024842060177749224, |
|
"grad_norm": 0.16710756719112396, |
|
"learning_rate": 0.00017459411454241822, |
|
"loss": 0.3655, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.02494913802334297, |
|
"grad_norm": 0.1497872918844223, |
|
"learning_rate": 0.00017438451298070252, |
|
"loss": 0.1916, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.025056215868936717, |
|
"grad_norm": 0.17602579295635223, |
|
"learning_rate": 0.00017417417727387394, |
|
"loss": 0.3154, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.025163293714530462, |
|
"grad_norm": 0.17509448528289795, |
|
"learning_rate": 0.000173963109497861, |
|
"loss": 0.2654, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.02527037156012421, |
|
"grad_norm": 0.22780412435531616, |
|
"learning_rate": 0.0001737513117358174, |
|
"loss": 0.4443, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.025377449405717956, |
|
"grad_norm": 0.15670035779476166, |
|
"learning_rate": 0.0001735387860781016, |
|
"loss": 0.2997, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.025484527251311704, |
|
"grad_norm": 0.20101696252822876, |
|
"learning_rate": 0.00017332553462225602, |
|
"loss": 0.3639, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.02559160509690545, |
|
"grad_norm": 0.18168163299560547, |
|
"learning_rate": 0.00017311155947298643, |
|
"loss": 0.4077, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.025698682942499197, |
|
"grad_norm": 0.1704721599817276, |
|
"learning_rate": 0.00017289686274214118, |
|
"loss": 0.2374, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.025805760788092942, |
|
"grad_norm": 0.18674886226654053, |
|
"learning_rate": 0.0001726814465486903, |
|
"loss": 0.368, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.02591283863368669, |
|
"grad_norm": 0.19927389919757843, |
|
"learning_rate": 0.0001724653130187047, |
|
"loss": 0.4098, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.026019916479280435, |
|
"grad_norm": 0.159666508436203, |
|
"learning_rate": 0.00017224846428533499, |
|
"loss": 0.3666, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.026126994324874184, |
|
"grad_norm": 0.15318766236305237, |
|
"learning_rate": 0.0001720309024887907, |
|
"loss": 0.2852, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.02623407217046793, |
|
"grad_norm": 0.1432930827140808, |
|
"learning_rate": 0.00017181262977631888, |
|
"loss": 0.2757, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.026341150016061677, |
|
"grad_norm": 0.18784955143928528, |
|
"learning_rate": 0.00017159364830218312, |
|
"loss": 0.2719, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.026448227861655422, |
|
"grad_norm": 0.15523239970207214, |
|
"learning_rate": 0.00017137396022764214, |
|
"loss": 0.2753, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.02655530570724917, |
|
"grad_norm": 0.17909985780715942, |
|
"learning_rate": 0.00017115356772092857, |
|
"loss": 0.3062, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.026662383552842915, |
|
"grad_norm": 0.17707036435604095, |
|
"learning_rate": 0.0001709324729572274, |
|
"loss": 0.3064, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.026769461398436663, |
|
"grad_norm": 0.15941859781742096, |
|
"learning_rate": 0.00017071067811865476, |
|
"loss": 0.3092, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.026876539244030412, |
|
"grad_norm": 0.18870612978935242, |
|
"learning_rate": 0.00017048818539423615, |
|
"loss": 0.4432, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.026983617089624157, |
|
"grad_norm": 0.1703352928161621, |
|
"learning_rate": 0.00017026499697988493, |
|
"loss": 0.3708, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.027090694935217905, |
|
"grad_norm": 0.2195807695388794, |
|
"learning_rate": 0.00017004111507838064, |
|
"loss": 0.4352, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.02719777278081165, |
|
"grad_norm": 0.2179376184940338, |
|
"learning_rate": 0.00016981654189934727, |
|
"loss": 0.3906, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.0273048506264054, |
|
"grad_norm": 0.16640064120292664, |
|
"learning_rate": 0.00016959127965923142, |
|
"loss": 0.2906, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.027411928471999143, |
|
"grad_norm": 0.1458374410867691, |
|
"learning_rate": 0.0001693653305812805, |
|
"loss": 0.2691, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.02751900631759289, |
|
"grad_norm": 0.15912285447120667, |
|
"learning_rate": 0.00016913869689552064, |
|
"loss": 0.3578, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.027626084163186636, |
|
"grad_norm": 0.17868658900260925, |
|
"learning_rate": 0.00016891138083873487, |
|
"loss": 0.4068, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.027733162008780385, |
|
"grad_norm": 0.1656470149755478, |
|
"learning_rate": 0.00016868338465444085, |
|
"loss": 0.3617, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.02784023985437413, |
|
"grad_norm": 0.1929635852575302, |
|
"learning_rate": 0.00016845471059286887, |
|
"loss": 0.3202, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.027947317699967878, |
|
"grad_norm": 0.1525058150291443, |
|
"learning_rate": 0.00016822536091093965, |
|
"loss": 0.3161, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.028054395545561623, |
|
"grad_norm": 0.1779276728630066, |
|
"learning_rate": 0.00016799533787224192, |
|
"loss": 0.3852, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.02816147339115537, |
|
"grad_norm": 0.16779017448425293, |
|
"learning_rate": 0.00016776464374701025, |
|
"loss": 0.3606, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.028268551236749116, |
|
"grad_norm": 0.14932067692279816, |
|
"learning_rate": 0.00016753328081210245, |
|
"loss": 0.2798, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.028375629082342865, |
|
"grad_norm": 0.12646813690662384, |
|
"learning_rate": 0.00016730125135097735, |
|
"loss": 0.2752, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.02848270692793661, |
|
"grad_norm": 0.1634105145931244, |
|
"learning_rate": 0.000167068557653672, |
|
"loss": 0.3424, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.028589784773530358, |
|
"grad_norm": 0.12478397786617279, |
|
"learning_rate": 0.0001668352020167793, |
|
"loss": 0.2156, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.028696862619124103, |
|
"grad_norm": 0.13762520253658295, |
|
"learning_rate": 0.00016660118674342517, |
|
"loss": 0.2768, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.02880394046471785, |
|
"grad_norm": 0.15719301998615265, |
|
"learning_rate": 0.00016636651414324587, |
|
"loss": 0.3163, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.028911018310311596, |
|
"grad_norm": 0.19167502224445343, |
|
"learning_rate": 0.00016613118653236518, |
|
"loss": 0.3347, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.029018096155905344, |
|
"grad_norm": 0.218104287981987, |
|
"learning_rate": 0.0001658952062333717, |
|
"loss": 0.3021, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.02912517400149909, |
|
"grad_norm": 0.16969047486782074, |
|
"learning_rate": 0.00016565857557529566, |
|
"loss": 0.3387, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.029232251847092838, |
|
"grad_norm": 0.19966812431812286, |
|
"learning_rate": 0.00016542129689358612, |
|
"loss": 0.3035, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.029339329692686583, |
|
"grad_norm": 0.12486845999956131, |
|
"learning_rate": 0.0001651833725300879, |
|
"loss": 0.2951, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.02944640753828033, |
|
"grad_norm": 0.1498948633670807, |
|
"learning_rate": 0.00016494480483301836, |
|
"loss": 0.3087, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.029553485383874076, |
|
"grad_norm": 0.16795475780963898, |
|
"learning_rate": 0.00016470559615694446, |
|
"loss": 0.3827, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.029660563229467824, |
|
"grad_norm": 0.1276923269033432, |
|
"learning_rate": 0.00016446574886275913, |
|
"loss": 0.2476, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.02976764107506157, |
|
"grad_norm": 0.14436961710453033, |
|
"learning_rate": 0.00016422526531765846, |
|
"loss": 0.3199, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.029874718920655317, |
|
"grad_norm": 0.2020837962627411, |
|
"learning_rate": 0.00016398414789511786, |
|
"loss": 0.3996, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.029981796766249062, |
|
"grad_norm": 0.15087634325027466, |
|
"learning_rate": 0.000163742398974869, |
|
"loss": 0.2648, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.03008887461184281, |
|
"grad_norm": 0.17405913770198822, |
|
"learning_rate": 0.00016350002094287609, |
|
"loss": 0.3969, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.030195952457436556, |
|
"grad_norm": 0.15879938006401062, |
|
"learning_rate": 0.00016325701619131246, |
|
"loss": 0.3349, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.030303030303030304, |
|
"grad_norm": 0.16552022099494934, |
|
"learning_rate": 0.00016301338711853693, |
|
"loss": 0.3489, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.03041010814862405, |
|
"grad_norm": 0.201536163687706, |
|
"learning_rate": 0.00016276913612907007, |
|
"loss": 0.4232, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.030517185994217797, |
|
"grad_norm": 0.22116144001483917, |
|
"learning_rate": 0.00016252426563357055, |
|
"loss": 0.4583, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.030624263839811542, |
|
"grad_norm": 0.21266256272792816, |
|
"learning_rate": 0.00016227877804881127, |
|
"loss": 0.4128, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.03073134168540529, |
|
"grad_norm": 0.16596044600009918, |
|
"learning_rate": 0.00016203267579765563, |
|
"loss": 0.2805, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.030838419530999035, |
|
"grad_norm": 0.233629509806633, |
|
"learning_rate": 0.00016178596130903344, |
|
"loss": 0.3752, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.030945497376592784, |
|
"grad_norm": 0.182915598154068, |
|
"learning_rate": 0.00016153863701791717, |
|
"loss": 0.3252, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.03105257522218653, |
|
"grad_norm": 0.16230888664722443, |
|
"learning_rate": 0.00016129070536529766, |
|
"loss": 0.2728, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.031159653067780277, |
|
"grad_norm": 0.21524645388126373, |
|
"learning_rate": 0.00016104216879816026, |
|
"loss": 0.4, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.03126673091337402, |
|
"grad_norm": 0.18035411834716797, |
|
"learning_rate": 0.00016079302976946055, |
|
"loss": 0.3949, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.03137380875896777, |
|
"grad_norm": 0.15152864158153534, |
|
"learning_rate": 0.00016054329073810015, |
|
"loss": 0.2806, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.03148088660456152, |
|
"grad_norm": 0.16412824392318726, |
|
"learning_rate": 0.00016029295416890248, |
|
"loss": 0.2953, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.03158796445015526, |
|
"grad_norm": 0.155482217669487, |
|
"learning_rate": 0.00016004202253258842, |
|
"loss": 0.2799, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.03169504229574901, |
|
"grad_norm": 0.13629105687141418, |
|
"learning_rate": 0.0001597904983057519, |
|
"loss": 0.2761, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.03180212014134275, |
|
"grad_norm": 0.18656809628009796, |
|
"learning_rate": 0.00015953838397083552, |
|
"loss": 0.2649, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.031909197986936505, |
|
"grad_norm": 0.1716616302728653, |
|
"learning_rate": 0.00015928568201610595, |
|
"loss": 0.2967, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.03201627583253025, |
|
"grad_norm": 0.19273731112480164, |
|
"learning_rate": 0.00015903239493562948, |
|
"loss": 0.4114, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.032123353678123995, |
|
"grad_norm": 0.16990764439105988, |
|
"learning_rate": 0.00015877852522924732, |
|
"loss": 0.3486, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03223043152371774, |
|
"grad_norm": 0.17067308723926544, |
|
"learning_rate": 0.00015852407540255104, |
|
"loss": 0.3064, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.03233750936931149, |
|
"grad_norm": 0.15841452777385712, |
|
"learning_rate": 0.00015826904796685762, |
|
"loss": 0.3338, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.032444587214905236, |
|
"grad_norm": 0.1738758087158203, |
|
"learning_rate": 0.00015801344543918495, |
|
"loss": 0.3908, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.03255166506049898, |
|
"grad_norm": 0.16537289321422577, |
|
"learning_rate": 0.00015775727034222675, |
|
"loss": 0.4218, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.032658742906092726, |
|
"grad_norm": 0.1708252876996994, |
|
"learning_rate": 0.00015750052520432787, |
|
"loss": 0.3284, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.03276582075168648, |
|
"grad_norm": 0.14976796507835388, |
|
"learning_rate": 0.0001572432125594591, |
|
"loss": 0.2875, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.03287289859728022, |
|
"grad_norm": 0.16668827831745148, |
|
"learning_rate": 0.00015698533494719238, |
|
"loss": 0.3899, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.03297997644287397, |
|
"grad_norm": 0.12908723950386047, |
|
"learning_rate": 0.00015672689491267567, |
|
"loss": 0.2418, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.03308705428846771, |
|
"grad_norm": 0.15876644849777222, |
|
"learning_rate": 0.00015646789500660773, |
|
"loss": 0.3975, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.033194132134061465, |
|
"grad_norm": 0.17148083448410034, |
|
"learning_rate": 0.00015620833778521307, |
|
"loss": 0.4312, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.03330120997965521, |
|
"grad_norm": 0.18296833336353302, |
|
"learning_rate": 0.0001559482258102167, |
|
"loss": 0.3408, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.033408287825248954, |
|
"grad_norm": 0.16649243235588074, |
|
"learning_rate": 0.00015568756164881882, |
|
"loss": 0.3898, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.0335153656708427, |
|
"grad_norm": 0.12311869859695435, |
|
"learning_rate": 0.00015542634787366942, |
|
"loss": 0.1999, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.03362244351643645, |
|
"grad_norm": 0.1413881629705429, |
|
"learning_rate": 0.00015516458706284303, |
|
"loss": 0.2685, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.033729521362030196, |
|
"grad_norm": 0.1433306783437729, |
|
"learning_rate": 0.0001549022817998132, |
|
"loss": 0.3465, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.03383659920762394, |
|
"grad_norm": 0.17652204632759094, |
|
"learning_rate": 0.00015463943467342693, |
|
"loss": 0.4089, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.03394367705321769, |
|
"grad_norm": 0.12554718554019928, |
|
"learning_rate": 0.00015437604827787927, |
|
"loss": 0.2809, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.03405075489881144, |
|
"grad_norm": 0.1779223084449768, |
|
"learning_rate": 0.00015411212521268758, |
|
"loss": 0.2576, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.03415783274440518, |
|
"grad_norm": 0.2108379304409027, |
|
"learning_rate": 0.00015384766808266602, |
|
"loss": 0.3919, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.03426491058999893, |
|
"grad_norm": 0.15554186701774597, |
|
"learning_rate": 0.00015358267949789966, |
|
"loss": 0.3479, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.03437198843559268, |
|
"grad_norm": 0.1714179813861847, |
|
"learning_rate": 0.00015331716207371888, |
|
"loss": 0.3595, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.034479066281186424, |
|
"grad_norm": 0.2408221811056137, |
|
"learning_rate": 0.0001530511184306734, |
|
"loss": 0.3556, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.03458614412678017, |
|
"grad_norm": 0.17631877958774567, |
|
"learning_rate": 0.00015278455119450664, |
|
"loss": 0.3639, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.034693221972373914, |
|
"grad_norm": 0.16605743765830994, |
|
"learning_rate": 0.0001525174629961296, |
|
"loss": 0.3989, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.034800299817967666, |
|
"grad_norm": 0.14433328807353973, |
|
"learning_rate": 0.0001522498564715949, |
|
"loss": 0.3316, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.03490737766356141, |
|
"grad_norm": 0.16017615795135498, |
|
"learning_rate": 0.00015198173426207094, |
|
"loss": 0.3285, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.035014455509155155, |
|
"grad_norm": 0.1323675662279129, |
|
"learning_rate": 0.00015171309901381572, |
|
"loss": 0.2674, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.0351215333547489, |
|
"grad_norm": 0.17207059264183044, |
|
"learning_rate": 0.00015144395337815064, |
|
"loss": 0.2994, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.03522861120034265, |
|
"grad_norm": 0.18844406306743622, |
|
"learning_rate": 0.00015117430001143452, |
|
"loss": 0.3926, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.0353356890459364, |
|
"grad_norm": 0.15473999083042145, |
|
"learning_rate": 0.00015090414157503714, |
|
"loss": 0.279, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.03544276689153014, |
|
"grad_norm": 0.1484527736902237, |
|
"learning_rate": 0.00015063348073531324, |
|
"loss": 0.2345, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.03554984473712389, |
|
"grad_norm": 0.14658762514591217, |
|
"learning_rate": 0.0001503623201635761, |
|
"loss": 0.2724, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.03565692258271764, |
|
"grad_norm": 0.14875711500644684, |
|
"learning_rate": 0.000150090662536071, |
|
"loss": 0.2749, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.035764000428311384, |
|
"grad_norm": 0.1902220994234085, |
|
"learning_rate": 0.0001498185105339491, |
|
"loss": 0.3198, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.03587107827390513, |
|
"grad_norm": 0.15855100750923157, |
|
"learning_rate": 0.00014954586684324078, |
|
"loss": 0.3031, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.03597815611949887, |
|
"grad_norm": 0.16661474108695984, |
|
"learning_rate": 0.00014927273415482915, |
|
"loss": 0.2691, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.036085233965092625, |
|
"grad_norm": 0.16410493850708008, |
|
"learning_rate": 0.00014899911516442365, |
|
"loss": 0.2248, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.03619231181068637, |
|
"grad_norm": 0.14755503833293915, |
|
"learning_rate": 0.00014872501257253323, |
|
"loss": 0.275, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.036299389656280115, |
|
"grad_norm": 0.1402604728937149, |
|
"learning_rate": 0.0001484504290844398, |
|
"loss": 0.2979, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.03640646750187386, |
|
"grad_norm": 0.20232371985912323, |
|
"learning_rate": 0.00014817536741017152, |
|
"loss": 0.3584, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.03651354534746761, |
|
"grad_norm": 0.13137860596179962, |
|
"learning_rate": 0.00014789983026447612, |
|
"loss": 0.2216, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.03662062319306136, |
|
"grad_norm": 0.18769167363643646, |
|
"learning_rate": 0.0001476238203667939, |
|
"loss": 0.2955, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.0367277010386551, |
|
"grad_norm": 0.17783814668655396, |
|
"learning_rate": 0.0001473473404412312, |
|
"loss": 0.2859, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.036834778884248846, |
|
"grad_norm": 0.17168748378753662, |
|
"learning_rate": 0.0001470703932165333, |
|
"loss": 0.3324, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.0369418567298426, |
|
"grad_norm": 0.2099820077419281, |
|
"learning_rate": 0.00014679298142605734, |
|
"loss": 0.4388, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.03704893457543634, |
|
"grad_norm": 0.17469075322151184, |
|
"learning_rate": 0.00014651510780774583, |
|
"loss": 0.3557, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.03715601242103009, |
|
"grad_norm": 0.1499973088502884, |
|
"learning_rate": 0.00014623677510409918, |
|
"loss": 0.2724, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.03726309026662383, |
|
"grad_norm": 0.1490471214056015, |
|
"learning_rate": 0.00014595798606214882, |
|
"loss": 0.3109, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.037370168112217585, |
|
"grad_norm": 0.18719898164272308, |
|
"learning_rate": 0.00014567874343342997, |
|
"loss": 0.3609, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.03747724595781133, |
|
"grad_norm": 0.17404448986053467, |
|
"learning_rate": 0.00014539904997395468, |
|
"loss": 0.3542, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.037584323803405074, |
|
"grad_norm": 0.17870911955833435, |
|
"learning_rate": 0.00014511890844418453, |
|
"loss": 0.253, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.03769140164899882, |
|
"grad_norm": 0.1721985787153244, |
|
"learning_rate": 0.00014483832160900326, |
|
"loss": 0.4102, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.03779847949459257, |
|
"grad_norm": 0.1740548461675644, |
|
"learning_rate": 0.00014455729223768966, |
|
"loss": 0.3727, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.037905557340186316, |
|
"grad_norm": 0.18063552677631378, |
|
"learning_rate": 0.0001442758231038902, |
|
"loss": 0.4705, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.03801263518578006, |
|
"grad_norm": 0.16232942044734955, |
|
"learning_rate": 0.00014399391698559152, |
|
"loss": 0.3015, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.038119713031373806, |
|
"grad_norm": 0.2140200287103653, |
|
"learning_rate": 0.0001437115766650933, |
|
"loss": 0.3251, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.03822679087696756, |
|
"grad_norm": 0.19321036338806152, |
|
"learning_rate": 0.00014342880492898048, |
|
"loss": 0.2617, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.0383338687225613, |
|
"grad_norm": 0.14323490858078003, |
|
"learning_rate": 0.0001431456045680959, |
|
"loss": 0.3512, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.03844094656815505, |
|
"grad_norm": 0.14133870601654053, |
|
"learning_rate": 0.00014286197837751286, |
|
"loss": 0.2648, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.03854802441374879, |
|
"grad_norm": 0.1461154669523239, |
|
"learning_rate": 0.00014257792915650728, |
|
"loss": 0.2978, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.038655102259342544, |
|
"grad_norm": 0.13456670939922333, |
|
"learning_rate": 0.00014229345970853032, |
|
"loss": 0.2399, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.03876218010493629, |
|
"grad_norm": 0.19285790622234344, |
|
"learning_rate": 0.00014200857284118066, |
|
"loss": 0.3538, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.038869257950530034, |
|
"grad_norm": 0.17173908650875092, |
|
"learning_rate": 0.00014172327136617656, |
|
"loss": 0.325, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.03897633579612378, |
|
"grad_norm": 0.16437563300132751, |
|
"learning_rate": 0.00014143755809932845, |
|
"loss": 0.2856, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.03908341364171753, |
|
"grad_norm": 0.16356609761714935, |
|
"learning_rate": 0.00014115143586051088, |
|
"loss": 0.3665, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.039190491487311276, |
|
"grad_norm": 0.17940573394298553, |
|
"learning_rate": 0.00014086490747363493, |
|
"loss": 0.3673, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.03929756933290502, |
|
"grad_norm": 0.17950116097927094, |
|
"learning_rate": 0.00014057797576662, |
|
"loss": 0.3944, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.039404647178498765, |
|
"grad_norm": 0.1494416892528534, |
|
"learning_rate": 0.00014029064357136628, |
|
"loss": 0.2745, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.03951172502409252, |
|
"grad_norm": 0.16802102327346802, |
|
"learning_rate": 0.00014000291372372647, |
|
"loss": 0.3329, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.03961880286968626, |
|
"grad_norm": 0.16510355472564697, |
|
"learning_rate": 0.00013971478906347806, |
|
"loss": 0.415, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.03972588071528001, |
|
"grad_norm": 0.14013710618019104, |
|
"learning_rate": 0.00013942627243429512, |
|
"loss": 0.2186, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.03983295856087375, |
|
"grad_norm": 0.20402979850769043, |
|
"learning_rate": 0.00013913736668372026, |
|
"loss": 0.4516, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.039940036406467504, |
|
"grad_norm": 0.1825643926858902, |
|
"learning_rate": 0.00013884807466313663, |
|
"loss": 0.314, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.04004711425206125, |
|
"grad_norm": 0.13864319026470184, |
|
"learning_rate": 0.00013855839922773968, |
|
"loss": 0.2619, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.040154192097654993, |
|
"grad_norm": 0.16547183692455292, |
|
"learning_rate": 0.000138268343236509, |
|
"loss": 0.3125, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.04026126994324874, |
|
"grad_norm": 0.17819920182228088, |
|
"learning_rate": 0.00013797790955218014, |
|
"loss": 0.4062, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.04036834778884249, |
|
"grad_norm": 0.154698446393013, |
|
"learning_rate": 0.00013768710104121627, |
|
"loss": 0.2868, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.040475425634436235, |
|
"grad_norm": 0.14120882749557495, |
|
"learning_rate": 0.00013739592057378003, |
|
"loss": 0.3406, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.04058250348002998, |
|
"grad_norm": 0.17801566421985626, |
|
"learning_rate": 0.0001371043710237051, |
|
"loss": 0.3826, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.04068958132562373, |
|
"grad_norm": 0.1434149593114853, |
|
"learning_rate": 0.00013681245526846783, |
|
"loss": 0.2858, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.04079665917121748, |
|
"grad_norm": 0.14281722903251648, |
|
"learning_rate": 0.0001365201761891588, |
|
"loss": 0.2466, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.04090373701681122, |
|
"grad_norm": 0.16065680980682373, |
|
"learning_rate": 0.00013622753667045457, |
|
"loss": 0.3634, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.041010814862404966, |
|
"grad_norm": 0.19151967763900757, |
|
"learning_rate": 0.00013593453960058908, |
|
"loss": 0.3986, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.04111789270799872, |
|
"grad_norm": 0.13544686138629913, |
|
"learning_rate": 0.00013564118787132506, |
|
"loss": 0.3097, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.04122497055359246, |
|
"grad_norm": 0.13807305693626404, |
|
"learning_rate": 0.00013534748437792573, |
|
"loss": 0.2874, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.04133204839918621, |
|
"grad_norm": 0.1964596062898636, |
|
"learning_rate": 0.0001350534320191259, |
|
"loss": 0.3109, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.04143912624477995, |
|
"grad_norm": 0.14541715383529663, |
|
"learning_rate": 0.0001347590336971037, |
|
"loss": 0.3624, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.041546204090373705, |
|
"grad_norm": 0.20359724760055542, |
|
"learning_rate": 0.0001344642923174517, |
|
"loss": 0.3734, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.04165328193596745, |
|
"grad_norm": 0.12174426764249802, |
|
"learning_rate": 0.00013416921078914835, |
|
"loss": 0.2217, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.041760359781561195, |
|
"grad_norm": 0.13207176327705383, |
|
"learning_rate": 0.00013387379202452917, |
|
"loss": 0.2343, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.04186743762715494, |
|
"grad_norm": 0.1388169378042221, |
|
"learning_rate": 0.00013357803893925807, |
|
"loss": 0.2587, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.04197451547274869, |
|
"grad_norm": 0.1489415019750595, |
|
"learning_rate": 0.00013328195445229868, |
|
"loss": 0.2599, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.042081593318342436, |
|
"grad_norm": 0.20052435994148254, |
|
"learning_rate": 0.00013298554148588528, |
|
"loss": 0.3817, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.04218867116393618, |
|
"grad_norm": 0.19129063189029694, |
|
"learning_rate": 0.00013268880296549425, |
|
"loss": 0.3565, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.042295749009529926, |
|
"grad_norm": 0.18302471935749054, |
|
"learning_rate": 0.00013239174181981495, |
|
"loss": 0.3647, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.04240282685512368, |
|
"grad_norm": 0.23317281901836395, |
|
"learning_rate": 0.00013209436098072095, |
|
"loss": 0.3992, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.04250990470071742, |
|
"grad_norm": 0.20024320483207703, |
|
"learning_rate": 0.00013179666338324108, |
|
"loss": 0.3974, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.04261698254631117, |
|
"grad_norm": 0.19934377074241638, |
|
"learning_rate": 0.0001314986519655305, |
|
"loss": 0.449, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.04272406039190491, |
|
"grad_norm": 0.1918846070766449, |
|
"learning_rate": 0.0001312003296688415, |
|
"loss": 0.5255, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.042831138237498664, |
|
"grad_norm": 0.14945387840270996, |
|
"learning_rate": 0.00013090169943749476, |
|
"loss": 0.2609, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.04293821608309241, |
|
"grad_norm": 0.16165776550769806, |
|
"learning_rate": 0.0001306027642188501, |
|
"loss": 0.3412, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.043045293928686154, |
|
"grad_norm": 0.15505026280879974, |
|
"learning_rate": 0.00013030352696327742, |
|
"loss": 0.3109, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.0431523717742799, |
|
"grad_norm": 0.17844939231872559, |
|
"learning_rate": 0.00013000399062412763, |
|
"loss": 0.3441, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.04325944961987365, |
|
"grad_norm": 0.19538049399852753, |
|
"learning_rate": 0.0001297041581577035, |
|
"loss": 0.3154, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.043366527465467396, |
|
"grad_norm": 0.12841279804706573, |
|
"learning_rate": 0.0001294040325232304, |
|
"loss": 0.2992, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.04347360531106114, |
|
"grad_norm": 0.15678054094314575, |
|
"learning_rate": 0.00012910361668282719, |
|
"loss": 0.3667, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.043580683156654886, |
|
"grad_norm": 0.19353367388248444, |
|
"learning_rate": 0.00012880291360147693, |
|
"loss": 0.3586, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.04368776100224864, |
|
"grad_norm": 0.18881572782993317, |
|
"learning_rate": 0.0001285019262469976, |
|
"loss": 0.3125, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.04379483884784238, |
|
"grad_norm": 0.14315952360630035, |
|
"learning_rate": 0.00012820065759001293, |
|
"loss": 0.2743, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.04390191669343613, |
|
"grad_norm": 0.1901596039533615, |
|
"learning_rate": 0.00012789911060392294, |
|
"loss": 0.405, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.04400899453902987, |
|
"grad_norm": 0.17396406829357147, |
|
"learning_rate": 0.0001275972882648746, |
|
"loss": 0.3503, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.044116072384623624, |
|
"grad_norm": 0.19500721991062164, |
|
"learning_rate": 0.00012729519355173254, |
|
"loss": 0.4379, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.04422315023021737, |
|
"grad_norm": 0.15421026945114136, |
|
"learning_rate": 0.00012699282944604967, |
|
"loss": 0.3766, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.044330228075811114, |
|
"grad_norm": 0.16204634308815002, |
|
"learning_rate": 0.00012669019893203759, |
|
"loss": 0.3692, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.04443730592140486, |
|
"grad_norm": 0.16477806866168976, |
|
"learning_rate": 0.0001263873049965373, |
|
"loss": 0.2539, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.04454438376699861, |
|
"grad_norm": 0.15045298635959625, |
|
"learning_rate": 0.00012608415062898972, |
|
"loss": 0.3063, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.044651461612592355, |
|
"grad_norm": 0.1610919088125229, |
|
"learning_rate": 0.000125780738821406, |
|
"loss": 0.3427, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.0447585394581861, |
|
"grad_norm": 0.19420892000198364, |
|
"learning_rate": 0.00012547707256833823, |
|
"loss": 0.3954, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.044865617303779845, |
|
"grad_norm": 0.18012918531894684, |
|
"learning_rate": 0.00012517315486684972, |
|
"loss": 0.3788, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.0449726951493736, |
|
"grad_norm": 0.162590891122818, |
|
"learning_rate": 0.0001248689887164855, |
|
"loss": 0.357, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.04507977299496734, |
|
"grad_norm": 0.15058323740959167, |
|
"learning_rate": 0.00012456457711924266, |
|
"loss": 0.245, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.04518685084056109, |
|
"grad_norm": 0.12929019331932068, |
|
"learning_rate": 0.00012425992307954075, |
|
"loss": 0.2408, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.04529392868615483, |
|
"grad_norm": 0.1378922462463379, |
|
"learning_rate": 0.0001239550296041922, |
|
"loss": 0.2867, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.04540100653174858, |
|
"grad_norm": 0.14614133536815643, |
|
"learning_rate": 0.00012364989970237248, |
|
"loss": 0.2963, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.04550808437734233, |
|
"grad_norm": 0.15353389084339142, |
|
"learning_rate": 0.00012334453638559057, |
|
"loss": 0.3199, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.04561516222293607, |
|
"grad_norm": 0.1354091763496399, |
|
"learning_rate": 0.00012303894266765908, |
|
"loss": 0.2396, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.04572224006852982, |
|
"grad_norm": 0.1608564555644989, |
|
"learning_rate": 0.00012273312156466464, |
|
"loss": 0.2776, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.04582931791412357, |
|
"grad_norm": 0.17579632997512817, |
|
"learning_rate": 0.00012242707609493814, |
|
"loss": 0.3, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.045936395759717315, |
|
"grad_norm": 0.1844303011894226, |
|
"learning_rate": 0.00012212080927902474, |
|
"loss": 0.3654, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.04604347360531106, |
|
"grad_norm": 0.15440614521503448, |
|
"learning_rate": 0.00012181432413965428, |
|
"loss": 0.2888, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.046150551450904805, |
|
"grad_norm": 0.1817633956670761, |
|
"learning_rate": 0.00012150762370171136, |
|
"loss": 0.4253, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.046257629296498556, |
|
"grad_norm": 0.17288000881671906, |
|
"learning_rate": 0.00012120071099220549, |
|
"loss": 0.3776, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.0463647071420923, |
|
"grad_norm": 0.18507465720176697, |
|
"learning_rate": 0.00012089358904024117, |
|
"loss": 0.3965, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.046471784987686046, |
|
"grad_norm": 0.1360858529806137, |
|
"learning_rate": 0.00012058626087698814, |
|
"loss": 0.2018, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.04657886283327979, |
|
"grad_norm": 0.18026049435138702, |
|
"learning_rate": 0.00012027872953565125, |
|
"loss": 0.3193, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.04668594067887354, |
|
"grad_norm": 0.2125885784626007, |
|
"learning_rate": 0.00011997099805144069, |
|
"loss": 0.4414, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.04679301852446729, |
|
"grad_norm": 0.18527942895889282, |
|
"learning_rate": 0.000119663069461542, |
|
"loss": 0.2822, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.04690009637006103, |
|
"grad_norm": 0.15211868286132812, |
|
"learning_rate": 0.00011935494680508606, |
|
"loss": 0.3093, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.047007174215654784, |
|
"grad_norm": 0.20865000784397125, |
|
"learning_rate": 0.00011904663312311901, |
|
"loss": 0.3999, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.04711425206124853, |
|
"grad_norm": 0.21300359070301056, |
|
"learning_rate": 0.00011873813145857249, |
|
"loss": 0.3397, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.047221329906842274, |
|
"grad_norm": 0.18172712624073029, |
|
"learning_rate": 0.00011842944485623335, |
|
"loss": 0.3376, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.04732840775243602, |
|
"grad_norm": 0.16084377467632294, |
|
"learning_rate": 0.00011812057636271374, |
|
"loss": 0.2233, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.04743548559802977, |
|
"grad_norm": 0.1644655168056488, |
|
"learning_rate": 0.000117811529026421, |
|
"loss": 0.3709, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.047542563443623516, |
|
"grad_norm": 0.1645745038986206, |
|
"learning_rate": 0.00011750230589752762, |
|
"loss": 0.3592, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.04764964128921726, |
|
"grad_norm": 0.19533959031105042, |
|
"learning_rate": 0.00011719291002794096, |
|
"loss": 0.3205, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.047756719134811006, |
|
"grad_norm": 0.14579270780086517, |
|
"learning_rate": 0.00011688334447127338, |
|
"loss": 0.3762, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.04786379698040476, |
|
"grad_norm": 0.175794318318367, |
|
"learning_rate": 0.00011657361228281199, |
|
"loss": 0.361, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.0479708748259985, |
|
"grad_norm": 0.1505967080593109, |
|
"learning_rate": 0.00011626371651948838, |
|
"loss": 0.2582, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.04807795267159225, |
|
"grad_norm": 0.10504525899887085, |
|
"learning_rate": 0.00011595366023984864, |
|
"loss": 0.1823, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.04818503051718599, |
|
"grad_norm": 0.19151091575622559, |
|
"learning_rate": 0.0001156434465040231, |
|
"loss": 0.3385, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.048292108362779744, |
|
"grad_norm": 0.19403024017810822, |
|
"learning_rate": 0.00011533307837369607, |
|
"loss": 0.3538, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.04839918620837349, |
|
"grad_norm": 0.185320183634758, |
|
"learning_rate": 0.00011502255891207572, |
|
"loss": 0.3393, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.048506264053967234, |
|
"grad_norm": 0.20061573386192322, |
|
"learning_rate": 0.00011471189118386375, |
|
"loss": 0.3567, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.04861334189956098, |
|
"grad_norm": 0.15041832625865936, |
|
"learning_rate": 0.00011440107825522521, |
|
"loss": 0.2794, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.04872041974515473, |
|
"grad_norm": 0.16546788811683655, |
|
"learning_rate": 0.00011409012319375827, |
|
"loss": 0.4382, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.048827497590748475, |
|
"grad_norm": 0.17634829878807068, |
|
"learning_rate": 0.0001137790290684638, |
|
"loss": 0.4121, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.04893457543634222, |
|
"grad_norm": 0.2208283543586731, |
|
"learning_rate": 0.00011346779894971527, |
|
"loss": 0.454, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.049041653281935965, |
|
"grad_norm": 0.16910183429718018, |
|
"learning_rate": 0.00011315643590922827, |
|
"loss": 0.3679, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.04914873112752972, |
|
"grad_norm": 0.18956156075000763, |
|
"learning_rate": 0.0001128449430200303, |
|
"loss": 0.4173, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.04925580897312346, |
|
"grad_norm": 0.1655164510011673, |
|
"learning_rate": 0.00011253332335643043, |
|
"loss": 0.3229, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.04936288681871721, |
|
"grad_norm": 0.1741032898426056, |
|
"learning_rate": 0.00011222157999398895, |
|
"loss": 0.2208, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.04946996466431095, |
|
"grad_norm": 0.14394614100456238, |
|
"learning_rate": 0.00011190971600948699, |
|
"loss": 0.3147, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.049577042509904704, |
|
"grad_norm": 0.19279611110687256, |
|
"learning_rate": 0.00011159773448089614, |
|
"loss": 0.4006, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.04968412035549845, |
|
"grad_norm": 0.17145444452762604, |
|
"learning_rate": 0.00011128563848734816, |
|
"loss": 0.3711, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.04979119820109219, |
|
"grad_norm": 0.15201549232006073, |
|
"learning_rate": 0.00011097343110910452, |
|
"loss": 0.3944, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.04989827604668594, |
|
"grad_norm": 0.19721528887748718, |
|
"learning_rate": 0.000110661115427526, |
|
"loss": 0.3839, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.05000535389227969, |
|
"grad_norm": 0.16234098374843597, |
|
"learning_rate": 0.00011034869452504226, |
|
"loss": 0.3303, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.050112431737873435, |
|
"grad_norm": 0.14313015341758728, |
|
"learning_rate": 0.00011003617148512149, |
|
"loss": 0.2862, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.05021950958346718, |
|
"grad_norm": 0.15435555577278137, |
|
"learning_rate": 0.00010972354939223996, |
|
"loss": 0.2875, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.050326587429060925, |
|
"grad_norm": 0.1893438845872879, |
|
"learning_rate": 0.00010941083133185146, |
|
"loss": 0.303, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.050433665274654677, |
|
"grad_norm": 0.17701826989650726, |
|
"learning_rate": 0.00010909802039035701, |
|
"loss": 0.3937, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.05054074312024842, |
|
"grad_norm": 0.15118297934532166, |
|
"learning_rate": 0.00010878511965507434, |
|
"loss": 0.2687, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.050647820965842166, |
|
"grad_norm": 0.16342921555042267, |
|
"learning_rate": 0.00010847213221420736, |
|
"loss": 0.2945, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.05075489881143591, |
|
"grad_norm": 0.16562896966934204, |
|
"learning_rate": 0.00010815906115681578, |
|
"loss": 0.3528, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.05086197665702966, |
|
"grad_norm": 0.17907415330410004, |
|
"learning_rate": 0.0001078459095727845, |
|
"loss": 0.2738, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.05096905450262341, |
|
"grad_norm": 0.20665884017944336, |
|
"learning_rate": 0.00010753268055279329, |
|
"loss": 0.3369, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.05107613234821715, |
|
"grad_norm": 0.14618425071239471, |
|
"learning_rate": 0.0001072193771882861, |
|
"loss": 0.3057, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.0511832101938109, |
|
"grad_norm": 0.15848585963249207, |
|
"learning_rate": 0.00010690600257144061, |
|
"loss": 0.3142, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.05129028803940465, |
|
"grad_norm": 0.1518888771533966, |
|
"learning_rate": 0.0001065925597951378, |
|
"loss": 0.2689, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.051397365884998394, |
|
"grad_norm": 0.1687459647655487, |
|
"learning_rate": 0.00010627905195293135, |
|
"loss": 0.2871, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.05150444373059214, |
|
"grad_norm": 0.16327998042106628, |
|
"learning_rate": 0.00010596548213901708, |
|
"loss": 0.2737, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.051611521576185884, |
|
"grad_norm": 0.17769210040569305, |
|
"learning_rate": 0.00010565185344820247, |
|
"loss": 0.3618, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.051718599421779636, |
|
"grad_norm": 0.17134082317352295, |
|
"learning_rate": 0.00010533816897587606, |
|
"loss": 0.3736, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.05182567726737338, |
|
"grad_norm": 0.17243537306785583, |
|
"learning_rate": 0.00010502443181797697, |
|
"loss": 0.3158, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.051932755112967126, |
|
"grad_norm": 0.172983318567276, |
|
"learning_rate": 0.00010471064507096426, |
|
"loss": 0.3272, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.05203983295856087, |
|
"grad_norm": 0.16794148087501526, |
|
"learning_rate": 0.0001043968118317865, |
|
"loss": 0.4177, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.05214691080415462, |
|
"grad_norm": 0.14923620223999023, |
|
"learning_rate": 0.00010408293519785101, |
|
"loss": 0.2119, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.05225398864974837, |
|
"grad_norm": 0.148018017411232, |
|
"learning_rate": 0.00010376901826699348, |
|
"loss": 0.3396, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.05236106649534211, |
|
"grad_norm": 0.21457895636558533, |
|
"learning_rate": 0.00010345506413744726, |
|
"loss": 0.2789, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.05246814434093586, |
|
"grad_norm": 0.16679152846336365, |
|
"learning_rate": 0.00010314107590781284, |
|
"loss": 0.3228, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.05257522218652961, |
|
"grad_norm": 0.17177151143550873, |
|
"learning_rate": 0.00010282705667702734, |
|
"loss": 0.2813, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.052682300032123354, |
|
"grad_norm": 0.1440981924533844, |
|
"learning_rate": 0.00010251300954433376, |
|
"loss": 0.3063, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.0527893778777171, |
|
"grad_norm": 0.1973811239004135, |
|
"learning_rate": 0.00010219893760925052, |
|
"loss": 0.4275, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.052896455723310844, |
|
"grad_norm": 0.15629740059375763, |
|
"learning_rate": 0.00010188484397154084, |
|
"loss": 0.3702, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.053003533568904596, |
|
"grad_norm": 0.15883494913578033, |
|
"learning_rate": 0.00010157073173118208, |
|
"loss": 0.3856, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.05311061141449834, |
|
"grad_norm": 0.21182221174240112, |
|
"learning_rate": 0.00010125660398833528, |
|
"loss": 0.4464, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.053217689260092085, |
|
"grad_norm": 0.179497629404068, |
|
"learning_rate": 0.00010094246384331442, |
|
"loss": 0.3077, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.05332476710568583, |
|
"grad_norm": 0.2272624969482422, |
|
"learning_rate": 0.00010062831439655591, |
|
"loss": 0.4539, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.05343184495127958, |
|
"grad_norm": 0.19556404650211334, |
|
"learning_rate": 0.00010031415874858797, |
|
"loss": 0.5309, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.05353892279687333, |
|
"grad_norm": 0.14988619089126587, |
|
"learning_rate": 0.0001, |
|
"loss": 0.2452, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05364600064246707, |
|
"grad_norm": 0.15577977895736694, |
|
"learning_rate": 9.968584125141204e-05, |
|
"loss": 0.3528, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.053753078488060824, |
|
"grad_norm": 0.2019302248954773, |
|
"learning_rate": 9.937168560344412e-05, |
|
"loss": 0.4053, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.05386015633365457, |
|
"grad_norm": 0.17697198688983917, |
|
"learning_rate": 9.90575361566856e-05, |
|
"loss": 0.3189, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.05396723417924831, |
|
"grad_norm": 0.15755611658096313, |
|
"learning_rate": 9.874339601166473e-05, |
|
"loss": 0.2804, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.05407431202484206, |
|
"grad_norm": 0.1574283093214035, |
|
"learning_rate": 9.842926826881796e-05, |
|
"loss": 0.3097, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.05418138987043581, |
|
"grad_norm": 0.11551764607429504, |
|
"learning_rate": 9.81151560284592e-05, |
|
"loss": 0.2281, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.054288467716029555, |
|
"grad_norm": 0.17114503681659698, |
|
"learning_rate": 9.78010623907495e-05, |
|
"loss": 0.306, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.0543955455616233, |
|
"grad_norm": 0.15964044630527496, |
|
"learning_rate": 9.748699045566626e-05, |
|
"loss": 0.3484, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.054502623407217045, |
|
"grad_norm": 0.1551811248064041, |
|
"learning_rate": 9.717294332297268e-05, |
|
"loss": 0.2986, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.0546097012528108, |
|
"grad_norm": 0.15225476026535034, |
|
"learning_rate": 9.685892409218717e-05, |
|
"loss": 0.3387, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.05471677909840454, |
|
"grad_norm": 0.18411734700202942, |
|
"learning_rate": 9.654493586255278e-05, |
|
"loss": 0.384, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.054823856943998286, |
|
"grad_norm": 0.1901894211769104, |
|
"learning_rate": 9.623098173300654e-05, |
|
"loss": 0.4127, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.05493093478959203, |
|
"grad_norm": 0.18249692022800446, |
|
"learning_rate": 9.591706480214901e-05, |
|
"loss": 0.2972, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.05503801263518578, |
|
"grad_norm": 0.16920414566993713, |
|
"learning_rate": 9.560318816821353e-05, |
|
"loss": 0.3322, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.05514509048077953, |
|
"grad_norm": 0.1470497101545334, |
|
"learning_rate": 9.528935492903575e-05, |
|
"loss": 0.2718, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.05525216832637327, |
|
"grad_norm": 0.1421034187078476, |
|
"learning_rate": 9.497556818202306e-05, |
|
"loss": 0.3081, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.05535924617196702, |
|
"grad_norm": 0.17241626977920532, |
|
"learning_rate": 9.466183102412395e-05, |
|
"loss": 0.3708, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.05546632401756077, |
|
"grad_norm": 0.16179907321929932, |
|
"learning_rate": 9.434814655179755e-05, |
|
"loss": 0.2864, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.055573401863154515, |
|
"grad_norm": 0.16402095556259155, |
|
"learning_rate": 9.403451786098294e-05, |
|
"loss": 0.3596, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.05568047970874826, |
|
"grad_norm": 0.20050428807735443, |
|
"learning_rate": 9.372094804706867e-05, |
|
"loss": 0.5162, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.055787557554342004, |
|
"grad_norm": 0.15083353221416473, |
|
"learning_rate": 9.340744020486222e-05, |
|
"loss": 0.3693, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.055894635399935756, |
|
"grad_norm": 0.15941321849822998, |
|
"learning_rate": 9.309399742855942e-05, |
|
"loss": 0.3116, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.0560017132455295, |
|
"grad_norm": 0.13679879903793335, |
|
"learning_rate": 9.278062281171393e-05, |
|
"loss": 0.2461, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.056108791091123246, |
|
"grad_norm": 0.1818103790283203, |
|
"learning_rate": 9.246731944720675e-05, |
|
"loss": 0.3061, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.05621586893671699, |
|
"grad_norm": 0.15920798480510712, |
|
"learning_rate": 9.215409042721552e-05, |
|
"loss": 0.3344, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.05632294678231074, |
|
"grad_norm": 0.2193126231431961, |
|
"learning_rate": 9.184093884318425e-05, |
|
"loss": 0.5115, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.05643002462790449, |
|
"grad_norm": 0.16002865135669708, |
|
"learning_rate": 9.152786778579267e-05, |
|
"loss": 0.3645, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.05653710247349823, |
|
"grad_norm": 0.14729884266853333, |
|
"learning_rate": 9.121488034492569e-05, |
|
"loss": 0.3054, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.05664418031909198, |
|
"grad_norm": 0.15512928366661072, |
|
"learning_rate": 9.090197960964301e-05, |
|
"loss": 0.3074, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.05675125816468573, |
|
"grad_norm": 0.16242021322250366, |
|
"learning_rate": 9.058916866814858e-05, |
|
"loss": 0.3717, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.056858336010279474, |
|
"grad_norm": 0.17270362377166748, |
|
"learning_rate": 9.027645060776006e-05, |
|
"loss": 0.4058, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.05696541385587322, |
|
"grad_norm": 0.1453240066766739, |
|
"learning_rate": 8.99638285148785e-05, |
|
"loss": 0.2409, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.057072491701466964, |
|
"grad_norm": 0.1585230678319931, |
|
"learning_rate": 8.965130547495776e-05, |
|
"loss": 0.3502, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.057179569547060716, |
|
"grad_norm": 0.16663610935211182, |
|
"learning_rate": 8.933888457247402e-05, |
|
"loss": 0.3985, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.05728664739265446, |
|
"grad_norm": 0.16193991899490356, |
|
"learning_rate": 8.902656889089548e-05, |
|
"loss": 0.308, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.057393725238248205, |
|
"grad_norm": 0.1661953181028366, |
|
"learning_rate": 8.871436151265184e-05, |
|
"loss": 0.3457, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.05750080308384195, |
|
"grad_norm": 0.15939539670944214, |
|
"learning_rate": 8.840226551910387e-05, |
|
"loss": 0.3158, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.0576078809294357, |
|
"grad_norm": 0.16002564132213593, |
|
"learning_rate": 8.809028399051302e-05, |
|
"loss": 0.3278, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.05771495877502945, |
|
"grad_norm": 0.1772623062133789, |
|
"learning_rate": 8.777842000601105e-05, |
|
"loss": 0.2166, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.05782203662062319, |
|
"grad_norm": 0.22258137166500092, |
|
"learning_rate": 8.746667664356956e-05, |
|
"loss": 0.3896, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.05792911446621694, |
|
"grad_norm": 0.1760905683040619, |
|
"learning_rate": 8.715505697996971e-05, |
|
"loss": 0.2794, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.05803619231181069, |
|
"grad_norm": 0.18258985877037048, |
|
"learning_rate": 8.684356409077176e-05, |
|
"loss": 0.3412, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.058143270157404434, |
|
"grad_norm": 0.20860691368579865, |
|
"learning_rate": 8.653220105028474e-05, |
|
"loss": 0.4274, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.05825034800299818, |
|
"grad_norm": 0.14682573080062866, |
|
"learning_rate": 8.62209709315362e-05, |
|
"loss": 0.3266, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.05835742584859192, |
|
"grad_norm": 0.15006284415721893, |
|
"learning_rate": 8.590987680624174e-05, |
|
"loss": 0.2551, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.058464503694185675, |
|
"grad_norm": 0.17763254046440125, |
|
"learning_rate": 8.559892174477479e-05, |
|
"loss": 0.4648, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.05857158153977942, |
|
"grad_norm": 0.21015936136245728, |
|
"learning_rate": 8.528810881613626e-05, |
|
"loss": 0.3025, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.058678659385373165, |
|
"grad_norm": 0.1679753214120865, |
|
"learning_rate": 8.497744108792429e-05, |
|
"loss": 0.2861, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.05878573723096691, |
|
"grad_norm": 0.18313933908939362, |
|
"learning_rate": 8.466692162630392e-05, |
|
"loss": 0.3483, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.05889281507656066, |
|
"grad_norm": 0.16596145927906036, |
|
"learning_rate": 8.435655349597689e-05, |
|
"loss": 0.3251, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.05899989292215441, |
|
"grad_norm": 0.18846194446086884, |
|
"learning_rate": 8.404633976015134e-05, |
|
"loss": 0.2898, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.05910697076774815, |
|
"grad_norm": 0.15795886516571045, |
|
"learning_rate": 8.373628348051165e-05, |
|
"loss": 0.2182, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.059214048613341896, |
|
"grad_norm": 0.14111833274364471, |
|
"learning_rate": 8.342638771718802e-05, |
|
"loss": 0.2136, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.05932112645893565, |
|
"grad_norm": 0.2854287624359131, |
|
"learning_rate": 8.311665552872662e-05, |
|
"loss": 0.4156, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.05942820430452939, |
|
"grad_norm": 0.16893833875656128, |
|
"learning_rate": 8.280708997205904e-05, |
|
"loss": 0.3394, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.05953528215012314, |
|
"grad_norm": 0.14959296584129333, |
|
"learning_rate": 8.249769410247239e-05, |
|
"loss": 0.2749, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.05964235999571688, |
|
"grad_norm": 0.16701342165470123, |
|
"learning_rate": 8.218847097357898e-05, |
|
"loss": 0.257, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.059749437841310635, |
|
"grad_norm": 0.16094553470611572, |
|
"learning_rate": 8.187942363728625e-05, |
|
"loss": 0.2504, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.05985651568690438, |
|
"grad_norm": 0.1865304857492447, |
|
"learning_rate": 8.157055514376666e-05, |
|
"loss": 0.3331, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.059963593532498125, |
|
"grad_norm": 0.1482362002134323, |
|
"learning_rate": 8.126186854142752e-05, |
|
"loss": 0.2317, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.06007067137809187, |
|
"grad_norm": 0.14238731563091278, |
|
"learning_rate": 8.095336687688102e-05, |
|
"loss": 0.2891, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.06017774922368562, |
|
"grad_norm": 0.19169366359710693, |
|
"learning_rate": 8.064505319491398e-05, |
|
"loss": 0.455, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.060284827069279366, |
|
"grad_norm": 0.1366347372531891, |
|
"learning_rate": 8.033693053845801e-05, |
|
"loss": 0.2489, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.06039190491487311, |
|
"grad_norm": 0.168013796210289, |
|
"learning_rate": 8.002900194855932e-05, |
|
"loss": 0.3256, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.06049898276046686, |
|
"grad_norm": 0.18004128336906433, |
|
"learning_rate": 7.972127046434878e-05, |
|
"loss": 0.3891, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.06060606060606061, |
|
"grad_norm": 0.18658877909183502, |
|
"learning_rate": 7.941373912301189e-05, |
|
"loss": 0.4804, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.06071313845165435, |
|
"grad_norm": 0.20196396112442017, |
|
"learning_rate": 7.910641095975886e-05, |
|
"loss": 0.4308, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.0608202162972481, |
|
"grad_norm": 0.18546605110168457, |
|
"learning_rate": 7.879928900779456e-05, |
|
"loss": 0.3777, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.06092729414284185, |
|
"grad_norm": 0.16258494555950165, |
|
"learning_rate": 7.849237629828869e-05, |
|
"loss": 0.3056, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.061034371988435594, |
|
"grad_norm": 0.207724466919899, |
|
"learning_rate": 7.818567586034577e-05, |
|
"loss": 0.3235, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.06114144983402934, |
|
"grad_norm": 0.15525054931640625, |
|
"learning_rate": 7.787919072097531e-05, |
|
"loss": 0.3045, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.061248527679623084, |
|
"grad_norm": 0.1632733941078186, |
|
"learning_rate": 7.75729239050619e-05, |
|
"loss": 0.3197, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.061355605525216836, |
|
"grad_norm": 0.19123557209968567, |
|
"learning_rate": 7.726687843533538e-05, |
|
"loss": 0.4378, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.06146268337081058, |
|
"grad_norm": 0.1749892383813858, |
|
"learning_rate": 7.696105733234098e-05, |
|
"loss": 0.332, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.061569761216404326, |
|
"grad_norm": 0.16800446808338165, |
|
"learning_rate": 7.66554636144095e-05, |
|
"loss": 0.4398, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.06167683906199807, |
|
"grad_norm": 0.164847269654274, |
|
"learning_rate": 7.635010029762756e-05, |
|
"loss": 0.2423, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.06178391690759182, |
|
"grad_norm": 0.13671499490737915, |
|
"learning_rate": 7.604497039580785e-05, |
|
"loss": 0.2792, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.06189099475318557, |
|
"grad_norm": 0.17302730679512024, |
|
"learning_rate": 7.574007692045928e-05, |
|
"loss": 0.3691, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.06199807259877931, |
|
"grad_norm": 0.16944162547588348, |
|
"learning_rate": 7.543542288075739e-05, |
|
"loss": 0.2405, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.06210515044437306, |
|
"grad_norm": 0.20139697194099426, |
|
"learning_rate": 7.513101128351454e-05, |
|
"loss": 0.3166, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.06221222828996681, |
|
"grad_norm": 0.16109436750411987, |
|
"learning_rate": 7.48268451331503e-05, |
|
"loss": 0.4416, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.062319306135560554, |
|
"grad_norm": 0.1434623748064041, |
|
"learning_rate": 7.45229274316618e-05, |
|
"loss": 0.266, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.0624263839811543, |
|
"grad_norm": 0.22343096137046814, |
|
"learning_rate": 7.421926117859403e-05, |
|
"loss": 0.4556, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.06253346182674804, |
|
"grad_norm": 0.18569520115852356, |
|
"learning_rate": 7.391584937101033e-05, |
|
"loss": 0.3634, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.06264053967234179, |
|
"grad_norm": 0.1524548977613449, |
|
"learning_rate": 7.361269500346274e-05, |
|
"loss": 0.3065, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.06274761751793553, |
|
"grad_norm": 0.2060038298368454, |
|
"learning_rate": 7.330980106796246e-05, |
|
"loss": 0.3112, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.06285469536352929, |
|
"grad_norm": 0.1781323105096817, |
|
"learning_rate": 7.300717055395039e-05, |
|
"loss": 0.3519, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.06296177320912304, |
|
"grad_norm": 0.16407997906208038, |
|
"learning_rate": 7.270480644826749e-05, |
|
"loss": 0.261, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.06306885105471678, |
|
"grad_norm": 0.15888680517673492, |
|
"learning_rate": 7.240271173512546e-05, |
|
"loss": 0.3337, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.06317592890031053, |
|
"grad_norm": 0.14741234481334686, |
|
"learning_rate": 7.210088939607708e-05, |
|
"loss": 0.3529, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.06328300674590427, |
|
"grad_norm": 0.17969931662082672, |
|
"learning_rate": 7.179934240998706e-05, |
|
"loss": 0.3359, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.06339008459149802, |
|
"grad_norm": 0.15772207081317902, |
|
"learning_rate": 7.149807375300239e-05, |
|
"loss": 0.2282, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.06349716243709176, |
|
"grad_norm": 0.23129960894584656, |
|
"learning_rate": 7.119708639852312e-05, |
|
"loss": 0.4251, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.0636042402826855, |
|
"grad_norm": 0.19827698171138763, |
|
"learning_rate": 7.089638331717284e-05, |
|
"loss": 0.4167, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.06371131812827927, |
|
"grad_norm": 0.18540829420089722, |
|
"learning_rate": 7.059596747676962e-05, |
|
"loss": 0.3376, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.06381839597387301, |
|
"grad_norm": 0.13156816363334656, |
|
"learning_rate": 7.029584184229653e-05, |
|
"loss": 0.2571, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.06392547381946675, |
|
"grad_norm": 0.1740744709968567, |
|
"learning_rate": 6.999600937587239e-05, |
|
"loss": 0.3639, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.0640325516650605, |
|
"grad_norm": 0.170998677611351, |
|
"learning_rate": 6.969647303672262e-05, |
|
"loss": 0.2915, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.06413962951065424, |
|
"grad_norm": 0.15191666781902313, |
|
"learning_rate": 6.939723578114993e-05, |
|
"loss": 0.2864, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.06424670735624799, |
|
"grad_norm": 0.150756374001503, |
|
"learning_rate": 6.909830056250527e-05, |
|
"loss": 0.3503, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.06435378520184173, |
|
"grad_norm": 0.1681331992149353, |
|
"learning_rate": 6.879967033115853e-05, |
|
"loss": 0.336, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.06446086304743548, |
|
"grad_norm": 0.1576251983642578, |
|
"learning_rate": 6.850134803446954e-05, |
|
"loss": 0.4084, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.06456794089302924, |
|
"grad_norm": 0.17241467535495758, |
|
"learning_rate": 6.820333661675893e-05, |
|
"loss": 0.3258, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.06467501873862298, |
|
"grad_norm": 0.15638768672943115, |
|
"learning_rate": 6.790563901927907e-05, |
|
"loss": 0.319, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.06478209658421673, |
|
"grad_norm": 0.15127749741077423, |
|
"learning_rate": 6.760825818018508e-05, |
|
"loss": 0.385, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.06488917442981047, |
|
"grad_norm": 0.18429574370384216, |
|
"learning_rate": 6.731119703450577e-05, |
|
"loss": 0.3338, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.06499625227540422, |
|
"grad_norm": 0.1837109476327896, |
|
"learning_rate": 6.701445851411472e-05, |
|
"loss": 0.3073, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.06510333012099796, |
|
"grad_norm": 0.1658046543598175, |
|
"learning_rate": 6.671804554770135e-05, |
|
"loss": 0.2748, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.06521040796659171, |
|
"grad_norm": 0.15374763309955597, |
|
"learning_rate": 6.642196106074194e-05, |
|
"loss": 0.3071, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.06531748581218545, |
|
"grad_norm": 0.15231020748615265, |
|
"learning_rate": 6.612620797547087e-05, |
|
"loss": 0.3016, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.06542456365777921, |
|
"grad_norm": 0.15196466445922852, |
|
"learning_rate": 6.583078921085167e-05, |
|
"loss": 0.2787, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.06553164150337296, |
|
"grad_norm": 0.1837112009525299, |
|
"learning_rate": 6.55357076825483e-05, |
|
"loss": 0.3969, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.0656387193489667, |
|
"grad_norm": 0.14765381813049316, |
|
"learning_rate": 6.52409663028963e-05, |
|
"loss": 0.3398, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.06574579719456045, |
|
"grad_norm": 0.21580912172794342, |
|
"learning_rate": 6.494656798087412e-05, |
|
"loss": 0.3179, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.06585287504015419, |
|
"grad_norm": 0.14525169134140015, |
|
"learning_rate": 6.465251562207431e-05, |
|
"loss": 0.2758, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.06595995288574794, |
|
"grad_norm": 0.1643163412809372, |
|
"learning_rate": 6.435881212867493e-05, |
|
"loss": 0.3803, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.06606703073134168, |
|
"grad_norm": 0.18450172245502472, |
|
"learning_rate": 6.406546039941094e-05, |
|
"loss": 0.3461, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.06617410857693543, |
|
"grad_norm": 0.17202648520469666, |
|
"learning_rate": 6.377246332954544e-05, |
|
"loss": 0.3287, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.06628118642252918, |
|
"grad_norm": 0.16204772889614105, |
|
"learning_rate": 6.347982381084123e-05, |
|
"loss": 0.2965, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.06638826426812293, |
|
"grad_norm": 0.14521515369415283, |
|
"learning_rate": 6.318754473153221e-05, |
|
"loss": 0.2263, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.06649534211371667, |
|
"grad_norm": 0.13584968447685242, |
|
"learning_rate": 6.289562897629492e-05, |
|
"loss": 0.2796, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.06660241995931042, |
|
"grad_norm": 0.15651963651180267, |
|
"learning_rate": 6.260407942621998e-05, |
|
"loss": 0.4405, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.06670949780490416, |
|
"grad_norm": 0.1673763245344162, |
|
"learning_rate": 6.231289895878375e-05, |
|
"loss": 0.2816, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.06681657565049791, |
|
"grad_norm": 0.16071754693984985, |
|
"learning_rate": 6.20220904478199e-05, |
|
"loss": 0.3041, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.06692365349609165, |
|
"grad_norm": 0.15596891939640045, |
|
"learning_rate": 6.173165676349103e-05, |
|
"loss": 0.3273, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.0670307313416854, |
|
"grad_norm": 0.14081811904907227, |
|
"learning_rate": 6.144160077226036e-05, |
|
"loss": 0.2684, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.06713780918727916, |
|
"grad_norm": 0.15451277792453766, |
|
"learning_rate": 6.11519253368634e-05, |
|
"loss": 0.2725, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.0672448870328729, |
|
"grad_norm": 0.1893017739057541, |
|
"learning_rate": 6.086263331627976e-05, |
|
"loss": 0.5062, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.06735196487846665, |
|
"grad_norm": 0.17904160916805267, |
|
"learning_rate": 6.05737275657049e-05, |
|
"loss": 0.4317, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.06745904272406039, |
|
"grad_norm": 0.20292861759662628, |
|
"learning_rate": 6.0285210936521955e-05, |
|
"loss": 0.3869, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.06756612056965414, |
|
"grad_norm": 0.1582748293876648, |
|
"learning_rate": 5.999708627627354e-05, |
|
"loss": 0.2296, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.06767319841524788, |
|
"grad_norm": 0.16759826242923737, |
|
"learning_rate": 5.9709356428633746e-05, |
|
"loss": 0.2913, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.06778027626084163, |
|
"grad_norm": 0.18039844930171967, |
|
"learning_rate": 5.9422024233380013e-05, |
|
"loss": 0.3225, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.06788735410643539, |
|
"grad_norm": 0.18086355924606323, |
|
"learning_rate": 5.913509252636511e-05, |
|
"loss": 0.2374, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.06799443195202913, |
|
"grad_norm": 0.1444912850856781, |
|
"learning_rate": 5.884856413948913e-05, |
|
"loss": 0.3337, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.06810150979762288, |
|
"grad_norm": 0.1468546986579895, |
|
"learning_rate": 5.856244190067159e-05, |
|
"loss": 0.2902, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.06820858764321662, |
|
"grad_norm": 0.17324474453926086, |
|
"learning_rate": 5.82767286338235e-05, |
|
"loss": 0.3888, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.06831566548881036, |
|
"grad_norm": 0.17539095878601074, |
|
"learning_rate": 5.799142715881938e-05, |
|
"loss": 0.439, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.06842274333440411, |
|
"grad_norm": 0.1325037181377411, |
|
"learning_rate": 5.770654029146969e-05, |
|
"loss": 0.2649, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.06852982117999785, |
|
"grad_norm": 0.1581648737192154, |
|
"learning_rate": 5.7422070843492734e-05, |
|
"loss": 0.2837, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.0686368990255916, |
|
"grad_norm": 0.17844368517398834, |
|
"learning_rate": 5.713802162248718e-05, |
|
"loss": 0.3192, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.06874397687118536, |
|
"grad_norm": 0.13828836381435394, |
|
"learning_rate": 5.6854395431904094e-05, |
|
"loss": 0.3098, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.0688510547167791, |
|
"grad_norm": 0.16136354207992554, |
|
"learning_rate": 5.657119507101954e-05, |
|
"loss": 0.4036, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.06895813256237285, |
|
"grad_norm": 0.17593613266944885, |
|
"learning_rate": 5.6288423334906735e-05, |
|
"loss": 0.3596, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.06906521040796659, |
|
"grad_norm": 0.1810455173254013, |
|
"learning_rate": 5.6006083014408484e-05, |
|
"loss": 0.3002, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.06917228825356034, |
|
"grad_norm": 0.18986423313617706, |
|
"learning_rate": 5.572417689610987e-05, |
|
"loss": 0.3121, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.06927936609915408, |
|
"grad_norm": 0.1990794688463211, |
|
"learning_rate": 5.544270776231038e-05, |
|
"loss": 0.3284, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.06938644394474783, |
|
"grad_norm": 0.15112511813640594, |
|
"learning_rate": 5.5161678390996796e-05, |
|
"loss": 0.3088, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.06949352179034157, |
|
"grad_norm": 0.23028168082237244, |
|
"learning_rate": 5.488109155581549e-05, |
|
"loss": 0.4338, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.06960059963593533, |
|
"grad_norm": 0.1991870403289795, |
|
"learning_rate": 5.4600950026045326e-05, |
|
"loss": 0.3372, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.06970767748152908, |
|
"grad_norm": 0.20563647150993347, |
|
"learning_rate": 5.4321256566570036e-05, |
|
"loss": 0.4628, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.06981475532712282, |
|
"grad_norm": 0.15499360859394073, |
|
"learning_rate": 5.404201393785122e-05, |
|
"loss": 0.2367, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.06992183317271657, |
|
"grad_norm": 0.20121176540851593, |
|
"learning_rate": 5.3763224895900846e-05, |
|
"loss": 0.3584, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.07002891101831031, |
|
"grad_norm": 0.16816240549087524, |
|
"learning_rate": 5.348489219225416e-05, |
|
"loss": 0.3287, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.07013598886390406, |
|
"grad_norm": 0.19340524077415466, |
|
"learning_rate": 5.320701857394268e-05, |
|
"loss": 0.3215, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.0702430667094978, |
|
"grad_norm": 0.16680116951465607, |
|
"learning_rate": 5.292960678346675e-05, |
|
"loss": 0.3505, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.07035014455509155, |
|
"grad_norm": 0.1708446592092514, |
|
"learning_rate": 5.265265955876879e-05, |
|
"loss": 0.347, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.0704572224006853, |
|
"grad_norm": 0.1849069744348526, |
|
"learning_rate": 5.237617963320608e-05, |
|
"loss": 0.3015, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.07056430024627905, |
|
"grad_norm": 0.18246212601661682, |
|
"learning_rate": 5.210016973552391e-05, |
|
"loss": 0.3702, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.0706713780918728, |
|
"grad_norm": 0.1265881061553955, |
|
"learning_rate": 5.182463258982846e-05, |
|
"loss": 0.2589, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.07077845593746654, |
|
"grad_norm": 0.1724609136581421, |
|
"learning_rate": 5.1549570915560206e-05, |
|
"loss": 0.3557, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.07088553378306028, |
|
"grad_norm": 0.15820424258708954, |
|
"learning_rate": 5.127498742746675e-05, |
|
"loss": 0.3181, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.07099261162865403, |
|
"grad_norm": 0.33776232600212097, |
|
"learning_rate": 5.100088483557634e-05, |
|
"loss": 0.4203, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.07109968947424777, |
|
"grad_norm": 0.19054165482521057, |
|
"learning_rate": 5.072726584517086e-05, |
|
"loss": 0.3089, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.07120676731984152, |
|
"grad_norm": 0.16495317220687866, |
|
"learning_rate": 5.045413315675924e-05, |
|
"loss": 0.3246, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.07131384516543528, |
|
"grad_norm": 0.16578929126262665, |
|
"learning_rate": 5.018148946605092e-05, |
|
"loss": 0.2536, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.07142092301102902, |
|
"grad_norm": 0.1600061058998108, |
|
"learning_rate": 4.990933746392899e-05, |
|
"loss": 0.3994, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.07152800085662277, |
|
"grad_norm": 0.2062436193227768, |
|
"learning_rate": 4.9637679836423924e-05, |
|
"loss": 0.3789, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.07163507870221651, |
|
"grad_norm": 0.197468101978302, |
|
"learning_rate": 4.9366519264686725e-05, |
|
"loss": 0.3784, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.07174215654781026, |
|
"grad_norm": 0.19289591908454895, |
|
"learning_rate": 4.909585842496287e-05, |
|
"loss": 0.3574, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.071849234393404, |
|
"grad_norm": 0.13273729383945465, |
|
"learning_rate": 4.8825699988565485e-05, |
|
"loss": 0.205, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.07195631223899775, |
|
"grad_norm": 0.16252535581588745, |
|
"learning_rate": 4.8556046621849346e-05, |
|
"loss": 0.2912, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.07206339008459149, |
|
"grad_norm": 0.17688702046871185, |
|
"learning_rate": 4.828690098618429e-05, |
|
"loss": 0.3397, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.07217046793018525, |
|
"grad_norm": 0.1794818937778473, |
|
"learning_rate": 4.8018265737929044e-05, |
|
"loss": 0.4702, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.072277545775779, |
|
"grad_norm": 0.17864678800106049, |
|
"learning_rate": 4.7750143528405126e-05, |
|
"loss": 0.3496, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.07238462362137274, |
|
"grad_norm": 0.15607388317584991, |
|
"learning_rate": 4.748253700387042e-05, |
|
"loss": 0.3233, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.07249170146696648, |
|
"grad_norm": 0.18187350034713745, |
|
"learning_rate": 4.721544880549337e-05, |
|
"loss": 0.3302, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.07259877931256023, |
|
"grad_norm": 0.16444875299930573, |
|
"learning_rate": 4.694888156932658e-05, |
|
"loss": 0.283, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.07270585715815397, |
|
"grad_norm": 0.15640141069889069, |
|
"learning_rate": 4.668283792628114e-05, |
|
"loss": 0.3251, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.07281293500374772, |
|
"grad_norm": 0.16079206764698029, |
|
"learning_rate": 4.6417320502100316e-05, |
|
"loss": 0.2819, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.07292001284934146, |
|
"grad_norm": 0.18440526723861694, |
|
"learning_rate": 4.615233191733398e-05, |
|
"loss": 0.3787, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.07302709069493522, |
|
"grad_norm": 0.1407841593027115, |
|
"learning_rate": 4.588787478731242e-05, |
|
"loss": 0.2777, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.07313416854052897, |
|
"grad_norm": 0.18285134434700012, |
|
"learning_rate": 4.5623951722120736e-05, |
|
"loss": 0.3911, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.07324124638612271, |
|
"grad_norm": 0.18212471902370453, |
|
"learning_rate": 4.5360565326573104e-05, |
|
"loss": 0.3466, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.07334832423171646, |
|
"grad_norm": 0.16741225123405457, |
|
"learning_rate": 4.5097718200186814e-05, |
|
"loss": 0.382, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.0734554020773102, |
|
"grad_norm": 0.160224050283432, |
|
"learning_rate": 4.483541293715698e-05, |
|
"loss": 0.3652, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.07356247992290395, |
|
"grad_norm": 0.1788359433412552, |
|
"learning_rate": 4.457365212633058e-05, |
|
"loss": 0.3589, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.07366955776849769, |
|
"grad_norm": 0.19194957613945007, |
|
"learning_rate": 4.431243835118124e-05, |
|
"loss": 0.366, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.07377663561409144, |
|
"grad_norm": 0.15268036723136902, |
|
"learning_rate": 4.4051774189783315e-05, |
|
"loss": 0.2694, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.0738837134596852, |
|
"grad_norm": 0.1825036257505417, |
|
"learning_rate": 4.379166221478697e-05, |
|
"loss": 0.291, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.07399079130527894, |
|
"grad_norm": 0.14247457683086395, |
|
"learning_rate": 4.3532104993392306e-05, |
|
"loss": 0.277, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.07409786915087269, |
|
"grad_norm": 0.19379214942455292, |
|
"learning_rate": 4.327310508732437e-05, |
|
"loss": 0.4844, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.07420494699646643, |
|
"grad_norm": 0.1550220251083374, |
|
"learning_rate": 4.301466505280762e-05, |
|
"loss": 0.3188, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.07431202484206018, |
|
"grad_norm": 0.16109517216682434, |
|
"learning_rate": 4.2756787440540936e-05, |
|
"loss": 0.3566, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.07441910268765392, |
|
"grad_norm": 0.15639358758926392, |
|
"learning_rate": 4.249947479567218e-05, |
|
"loss": 0.3562, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.07452618053324767, |
|
"grad_norm": 0.17039039731025696, |
|
"learning_rate": 4.224272965777326e-05, |
|
"loss": 0.2699, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.07463325837884142, |
|
"grad_norm": 0.1703154742717743, |
|
"learning_rate": 4.1986554560815096e-05, |
|
"loss": 0.3501, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.07474033622443517, |
|
"grad_norm": 0.1571662575006485, |
|
"learning_rate": 4.173095203314241e-05, |
|
"loss": 0.3103, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.07484741407002891, |
|
"grad_norm": 0.18659600615501404, |
|
"learning_rate": 4.1475924597449024e-05, |
|
"loss": 0.3283, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.07495449191562266, |
|
"grad_norm": 0.18269579112529755, |
|
"learning_rate": 4.12214747707527e-05, |
|
"loss": 0.3737, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.0750615697612164, |
|
"grad_norm": 0.17083945870399475, |
|
"learning_rate": 4.096760506437057e-05, |
|
"loss": 0.338, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.07516864760681015, |
|
"grad_norm": 0.16920264065265656, |
|
"learning_rate": 4.071431798389408e-05, |
|
"loss": 0.3694, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.0752757254524039, |
|
"grad_norm": 0.22974881529808044, |
|
"learning_rate": 4.0461616029164526e-05, |
|
"loss": 0.3473, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.07538280329799764, |
|
"grad_norm": 0.1423478126525879, |
|
"learning_rate": 4.020950169424815e-05, |
|
"loss": 0.3492, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.0754898811435914, |
|
"grad_norm": 0.15745437145233154, |
|
"learning_rate": 3.9957977467411615e-05, |
|
"loss": 0.2361, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.07559695898918514, |
|
"grad_norm": 0.13066445291042328, |
|
"learning_rate": 3.9707045831097555e-05, |
|
"loss": 0.2374, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.07570403683477889, |
|
"grad_norm": 0.18829156458377838, |
|
"learning_rate": 3.945670926189987e-05, |
|
"loss": 0.3874, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.07581111468037263, |
|
"grad_norm": 0.19448167085647583, |
|
"learning_rate": 3.920697023053949e-05, |
|
"loss": 0.4145, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.07591819252596638, |
|
"grad_norm": 0.15791544318199158, |
|
"learning_rate": 3.895783120183976e-05, |
|
"loss": 0.3423, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.07602527037156012, |
|
"grad_norm": 0.16304408013820648, |
|
"learning_rate": 3.8709294634702376e-05, |
|
"loss": 0.3469, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.07613234821715387, |
|
"grad_norm": 0.15038977563381195, |
|
"learning_rate": 3.846136298208285e-05, |
|
"loss": 0.2923, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.07623942606274761, |
|
"grad_norm": 0.1510382443666458, |
|
"learning_rate": 3.821403869096658e-05, |
|
"loss": 0.2773, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.07634650390834137, |
|
"grad_norm": 0.2000676840543747, |
|
"learning_rate": 3.796732420234443e-05, |
|
"loss": 0.3363, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.07645358175393512, |
|
"grad_norm": 0.1835888922214508, |
|
"learning_rate": 3.7721221951188765e-05, |
|
"loss": 0.3695, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.07656065959952886, |
|
"grad_norm": 0.1762959063053131, |
|
"learning_rate": 3.747573436642951e-05, |
|
"loss": 0.3929, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.0766677374451226, |
|
"grad_norm": 0.15682466328144073, |
|
"learning_rate": 3.7230863870929964e-05, |
|
"loss": 0.3103, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.07677481529071635, |
|
"grad_norm": 0.17416490614414215, |
|
"learning_rate": 3.698661288146311e-05, |
|
"loss": 0.4379, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.0768818931363101, |
|
"grad_norm": 0.23010078072547913, |
|
"learning_rate": 3.674298380868756e-05, |
|
"loss": 0.5982, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.07698897098190384, |
|
"grad_norm": 0.1550900787115097, |
|
"learning_rate": 3.649997905712396e-05, |
|
"loss": 0.2699, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.07709604882749758, |
|
"grad_norm": 0.14471879601478577, |
|
"learning_rate": 3.6257601025131026e-05, |
|
"loss": 0.2861, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.07720312667309134, |
|
"grad_norm": 0.18543073534965515, |
|
"learning_rate": 3.601585210488218e-05, |
|
"loss": 0.288, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.07731020451868509, |
|
"grad_norm": 0.21810288727283478, |
|
"learning_rate": 3.577473468234156e-05, |
|
"loss": 0.3492, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.07741728236427883, |
|
"grad_norm": 0.18458350002765656, |
|
"learning_rate": 3.553425113724088e-05, |
|
"loss": 0.3745, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.07752436020987258, |
|
"grad_norm": 0.17252208292484283, |
|
"learning_rate": 3.52944038430556e-05, |
|
"loss": 0.376, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.07763143805546632, |
|
"grad_norm": 0.15418677031993866, |
|
"learning_rate": 3.5055195166981645e-05, |
|
"loss": 0.3308, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.07773851590106007, |
|
"grad_norm": 0.17200735211372375, |
|
"learning_rate": 3.481662746991214e-05, |
|
"loss": 0.3643, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.07784559374665381, |
|
"grad_norm": 0.13179121911525726, |
|
"learning_rate": 3.4578703106413904e-05, |
|
"loss": 0.2808, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.07795267159224756, |
|
"grad_norm": 0.16644497215747833, |
|
"learning_rate": 3.4341424424704375e-05, |
|
"loss": 0.3007, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.07805974943784132, |
|
"grad_norm": 0.17268535494804382, |
|
"learning_rate": 3.4104793766628304e-05, |
|
"loss": 0.3881, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.07816682728343506, |
|
"grad_norm": 0.15279394388198853, |
|
"learning_rate": 3.386881346763483e-05, |
|
"loss": 0.2837, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.0782739051290288, |
|
"grad_norm": 0.155588760972023, |
|
"learning_rate": 3.363348585675414e-05, |
|
"loss": 0.2995, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.07838098297462255, |
|
"grad_norm": 0.1607223004102707, |
|
"learning_rate": 3.339881325657484e-05, |
|
"loss": 0.3795, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.0784880608202163, |
|
"grad_norm": 0.15511462092399597, |
|
"learning_rate": 3.316479798322072e-05, |
|
"loss": 0.3791, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.07859513866581004, |
|
"grad_norm": 0.13938754796981812, |
|
"learning_rate": 3.2931442346328004e-05, |
|
"loss": 0.2123, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.07870221651140379, |
|
"grad_norm": 0.18411380052566528, |
|
"learning_rate": 3.269874864902269e-05, |
|
"loss": 0.2817, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.07880929435699753, |
|
"grad_norm": 0.13664402067661285, |
|
"learning_rate": 3.246671918789755e-05, |
|
"loss": 0.2564, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.07891637220259129, |
|
"grad_norm": 0.1579146832227707, |
|
"learning_rate": 3.223535625298979e-05, |
|
"loss": 0.2813, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.07902345004818503, |
|
"grad_norm": 0.17424800992012024, |
|
"learning_rate": 3.200466212775808e-05, |
|
"loss": 0.2554, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.07913052789377878, |
|
"grad_norm": 0.15112292766571045, |
|
"learning_rate": 3.1774639089060363e-05, |
|
"loss": 0.3183, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.07923760573937252, |
|
"grad_norm": 0.2046094685792923, |
|
"learning_rate": 3.154528940713113e-05, |
|
"loss": 0.3629, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.07934468358496627, |
|
"grad_norm": 0.22456932067871094, |
|
"learning_rate": 3.1316615345559185e-05, |
|
"loss": 0.363, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.07945176143056001, |
|
"grad_norm": 0.17923475801944733, |
|
"learning_rate": 3.108861916126518e-05, |
|
"loss": 0.4022, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.07955883927615376, |
|
"grad_norm": 0.18033398687839508, |
|
"learning_rate": 3.086130310447937e-05, |
|
"loss": 0.3728, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.0796659171217475, |
|
"grad_norm": 0.14223916828632355, |
|
"learning_rate": 3.063466941871952e-05, |
|
"loss": 0.2802, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.07977299496734126, |
|
"grad_norm": 0.156300351023674, |
|
"learning_rate": 3.0408720340768572e-05, |
|
"loss": 0.3538, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.07988007281293501, |
|
"grad_norm": 0.13515561819076538, |
|
"learning_rate": 3.018345810065275e-05, |
|
"loss": 0.2488, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.07998715065852875, |
|
"grad_norm": 0.16864290833473206, |
|
"learning_rate": 2.9958884921619367e-05, |
|
"loss": 0.358, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.0800942285041225, |
|
"grad_norm": 0.20047788321971893, |
|
"learning_rate": 2.9735003020115092e-05, |
|
"loss": 0.4827, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.08020130634971624, |
|
"grad_norm": 0.18187908828258514, |
|
"learning_rate": 2.9511814605763855e-05, |
|
"loss": 0.322, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.08030838419530999, |
|
"grad_norm": 0.16408580541610718, |
|
"learning_rate": 2.9289321881345254e-05, |
|
"loss": 0.3125, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.08041546204090373, |
|
"grad_norm": 0.19623160362243652, |
|
"learning_rate": 2.9067527042772636e-05, |
|
"loss": 0.4812, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.08052253988649748, |
|
"grad_norm": 0.1534736603498459, |
|
"learning_rate": 2.8846432279071467e-05, |
|
"loss": 0.3785, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.08062961773209124, |
|
"grad_norm": 0.1560440957546234, |
|
"learning_rate": 2.8626039772357882e-05, |
|
"loss": 0.3226, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.08073669557768498, |
|
"grad_norm": 0.21636521816253662, |
|
"learning_rate": 2.840635169781688e-05, |
|
"loss": 0.4498, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.08084377342327873, |
|
"grad_norm": 0.1631963849067688, |
|
"learning_rate": 2.8187370223681132e-05, |
|
"loss": 0.3713, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.08095085126887247, |
|
"grad_norm": 0.13680961728096008, |
|
"learning_rate": 2.7969097511209308e-05, |
|
"loss": 0.2974, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.08105792911446622, |
|
"grad_norm": 0.17071770131587982, |
|
"learning_rate": 2.775153571466502e-05, |
|
"loss": 0.2913, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.08116500696005996, |
|
"grad_norm": 0.14643071591854095, |
|
"learning_rate": 2.753468698129533e-05, |
|
"loss": 0.2833, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.0812720848056537, |
|
"grad_norm": 0.1774536669254303, |
|
"learning_rate": 2.7318553451309726e-05, |
|
"loss": 0.3131, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.08137916265124746, |
|
"grad_norm": 0.1635051816701889, |
|
"learning_rate": 2.7103137257858868e-05, |
|
"loss": 0.354, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.08148624049684121, |
|
"grad_norm": 0.20249997079372406, |
|
"learning_rate": 2.688844052701359e-05, |
|
"loss": 0.3487, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.08159331834243495, |
|
"grad_norm": 0.14786763489246368, |
|
"learning_rate": 2.6674465377744017e-05, |
|
"loss": 0.3071, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.0817003961880287, |
|
"grad_norm": 0.13844721019268036, |
|
"learning_rate": 2.646121392189841e-05, |
|
"loss": 0.2367, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.08180747403362244, |
|
"grad_norm": 0.1526380181312561, |
|
"learning_rate": 2.624868826418262e-05, |
|
"loss": 0.2976, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.08191455187921619, |
|
"grad_norm": 0.1874617487192154, |
|
"learning_rate": 2.603689050213902e-05, |
|
"loss": 0.2991, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.08202162972480993, |
|
"grad_norm": 0.1745980829000473, |
|
"learning_rate": 2.582582272612609e-05, |
|
"loss": 0.2624, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.08212870757040368, |
|
"grad_norm": 0.15471521019935608, |
|
"learning_rate": 2.561548701929749e-05, |
|
"loss": 0.3706, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.08223578541599744, |
|
"grad_norm": 0.16713665425777435, |
|
"learning_rate": 2.540588545758179e-05, |
|
"loss": 0.2719, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.08234286326159118, |
|
"grad_norm": 0.14261217415332794, |
|
"learning_rate": 2.5197020109661772e-05, |
|
"loss": 0.2737, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.08244994110718493, |
|
"grad_norm": 0.16853374242782593, |
|
"learning_rate": 2.4988893036954043e-05, |
|
"loss": 0.3817, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.08255701895277867, |
|
"grad_norm": 0.17833545804023743, |
|
"learning_rate": 2.4781506293588873e-05, |
|
"loss": 0.3775, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.08266409679837242, |
|
"grad_norm": 0.18436114490032196, |
|
"learning_rate": 2.4574861926389615e-05, |
|
"loss": 0.244, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.08277117464396616, |
|
"grad_norm": 0.17707830667495728, |
|
"learning_rate": 2.436896197485282e-05, |
|
"loss": 0.4478, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.0828782524895599, |
|
"grad_norm": 0.19500042498111725, |
|
"learning_rate": 2.4163808471127812e-05, |
|
"loss": 0.3646, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.08298533033515365, |
|
"grad_norm": 0.1503261774778366, |
|
"learning_rate": 2.3959403439996907e-05, |
|
"loss": 0.3389, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.08309240818074741, |
|
"grad_norm": 0.1736772656440735, |
|
"learning_rate": 2.37557488988552e-05, |
|
"loss": 0.3369, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.08319948602634115, |
|
"grad_norm": 0.1980380415916443, |
|
"learning_rate": 2.3552846857690846e-05, |
|
"loss": 0.3544, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.0833065638719349, |
|
"grad_norm": 0.16803400218486786, |
|
"learning_rate": 2.3350699319065026e-05, |
|
"loss": 0.375, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.08341364171752864, |
|
"grad_norm": 0.17213231325149536, |
|
"learning_rate": 2.3149308278092342e-05, |
|
"loss": 0.3151, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.08352071956312239, |
|
"grad_norm": 0.18893983960151672, |
|
"learning_rate": 2.2948675722421086e-05, |
|
"loss": 0.3629, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.08362779740871613, |
|
"grad_norm": 0.20004568994045258, |
|
"learning_rate": 2.2748803632213557e-05, |
|
"loss": 0.3777, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.08373487525430988, |
|
"grad_norm": 0.1406610906124115, |
|
"learning_rate": 2.254969398012663e-05, |
|
"loss": 0.2712, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.08384195309990362, |
|
"grad_norm": 0.1625978797674179, |
|
"learning_rate": 2.235134873129213e-05, |
|
"loss": 0.3094, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.08394903094549738, |
|
"grad_norm": 0.1865498125553131, |
|
"learning_rate": 2.2153769843297667e-05, |
|
"loss": 0.3852, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.08405610879109113, |
|
"grad_norm": 0.16556480526924133, |
|
"learning_rate": 2.195695926616702e-05, |
|
"loss": 0.2227, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.08416318663668487, |
|
"grad_norm": 0.1859363615512848, |
|
"learning_rate": 2.1760918942341192e-05, |
|
"loss": 0.4757, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.08427026448227862, |
|
"grad_norm": 0.17114213109016418, |
|
"learning_rate": 2.1565650806658975e-05, |
|
"loss": 0.3861, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.08437734232787236, |
|
"grad_norm": 0.18057489395141602, |
|
"learning_rate": 2.137115678633811e-05, |
|
"loss": 0.3755, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.08448442017346611, |
|
"grad_norm": 0.16163988411426544, |
|
"learning_rate": 2.1177438800956007e-05, |
|
"loss": 0.3012, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.08459149801905985, |
|
"grad_norm": 0.17569121718406677, |
|
"learning_rate": 2.098449876243096e-05, |
|
"loss": 0.3466, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.0846985758646536, |
|
"grad_norm": 0.14997471868991852, |
|
"learning_rate": 2.07923385750033e-05, |
|
"loss": 0.2821, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.08480565371024736, |
|
"grad_norm": 0.17677253484725952, |
|
"learning_rate": 2.0600960135216462e-05, |
|
"loss": 0.3436, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.0849127315558411, |
|
"grad_norm": 0.14548727869987488, |
|
"learning_rate": 2.0410365331898416e-05, |
|
"loss": 0.2875, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.08501980940143485, |
|
"grad_norm": 0.18589341640472412, |
|
"learning_rate": 2.0220556046142893e-05, |
|
"loss": 0.3885, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.08512688724702859, |
|
"grad_norm": 0.16286922991275787, |
|
"learning_rate": 2.0031534151290943e-05, |
|
"loss": 0.2452, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.08523396509262234, |
|
"grad_norm": 0.14786767959594727, |
|
"learning_rate": 1.9843301512912327e-05, |
|
"loss": 0.3465, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.08534104293821608, |
|
"grad_norm": 0.15878774225711823, |
|
"learning_rate": 1.965585998878724e-05, |
|
"loss": 0.3714, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.08544812078380983, |
|
"grad_norm": 0.16029423475265503, |
|
"learning_rate": 1.946921142888781e-05, |
|
"loss": 0.3118, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.08555519862940357, |
|
"grad_norm": 0.11456546932458878, |
|
"learning_rate": 1.928335767535997e-05, |
|
"loss": 0.2231, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.08566227647499733, |
|
"grad_norm": 0.17691242694854736, |
|
"learning_rate": 1.9098300562505266e-05, |
|
"loss": 0.3831, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.08576935432059107, |
|
"grad_norm": 0.178320974111557, |
|
"learning_rate": 1.891404191676265e-05, |
|
"loss": 0.3814, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.08587643216618482, |
|
"grad_norm": 0.1586633324623108, |
|
"learning_rate": 1.8730583556690605e-05, |
|
"loss": 0.2953, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.08598351001177856, |
|
"grad_norm": 0.1737189143896103, |
|
"learning_rate": 1.854792729294905e-05, |
|
"loss": 0.1958, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.08609058785737231, |
|
"grad_norm": 0.17950400710105896, |
|
"learning_rate": 1.8366074928281607e-05, |
|
"loss": 0.2301, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.08619766570296605, |
|
"grad_norm": 0.1717943549156189, |
|
"learning_rate": 1.818502825749764e-05, |
|
"loss": 0.3452, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.0863047435485598, |
|
"grad_norm": 0.19653019309043884, |
|
"learning_rate": 1.8004789067454764e-05, |
|
"loss": 0.3789, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.08641182139415354, |
|
"grad_norm": 0.23031754791736603, |
|
"learning_rate": 1.7825359137040988e-05, |
|
"loss": 0.3241, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.0865188992397473, |
|
"grad_norm": 0.19152677059173584, |
|
"learning_rate": 1.7646740237157256e-05, |
|
"loss": 0.3634, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.08662597708534105, |
|
"grad_norm": 0.14082860946655273, |
|
"learning_rate": 1.7468934130700044e-05, |
|
"loss": 0.2035, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.08673305493093479, |
|
"grad_norm": 0.15468376874923706, |
|
"learning_rate": 1.7291942572543807e-05, |
|
"loss": 0.299, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.08684013277652854, |
|
"grad_norm": 0.1572645902633667, |
|
"learning_rate": 1.7115767309523812e-05, |
|
"loss": 0.3126, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.08694721062212228, |
|
"grad_norm": 0.19454143941402435, |
|
"learning_rate": 1.6940410080418723e-05, |
|
"loss": 0.4161, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.08705428846771603, |
|
"grad_norm": 0.14077642560005188, |
|
"learning_rate": 1.6765872615933677e-05, |
|
"loss": 0.249, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.08716136631330977, |
|
"grad_norm": 0.22318896651268005, |
|
"learning_rate": 1.6592156638682886e-05, |
|
"loss": 0.3444, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.08726844415890353, |
|
"grad_norm": 0.15059834718704224, |
|
"learning_rate": 1.6419263863172997e-05, |
|
"loss": 0.2598, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.08737552200449727, |
|
"grad_norm": 0.1700175404548645, |
|
"learning_rate": 1.6247195995785837e-05, |
|
"loss": 0.372, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.08748259985009102, |
|
"grad_norm": 0.19216172397136688, |
|
"learning_rate": 1.6075954734761845e-05, |
|
"loss": 0.3689, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.08758967769568476, |
|
"grad_norm": 0.1473630666732788, |
|
"learning_rate": 1.5905541770183096e-05, |
|
"loss": 0.2568, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.08769675554127851, |
|
"grad_norm": 0.17223362624645233, |
|
"learning_rate": 1.5735958783956794e-05, |
|
"loss": 0.3367, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.08780383338687225, |
|
"grad_norm": 0.18466192483901978, |
|
"learning_rate": 1.5567207449798515e-05, |
|
"loss": 0.3607, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.087910911232466, |
|
"grad_norm": 0.1918594390153885, |
|
"learning_rate": 1.539928943321579e-05, |
|
"loss": 0.4122, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.08801798907805974, |
|
"grad_norm": 0.17056573927402496, |
|
"learning_rate": 1.5232206391491699e-05, |
|
"loss": 0.2995, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.0881250669236535, |
|
"grad_norm": 0.16323423385620117, |
|
"learning_rate": 1.5065959973668353e-05, |
|
"loss": 0.3332, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.08823214476924725, |
|
"grad_norm": 0.14421838521957397, |
|
"learning_rate": 1.4900551820530828e-05, |
|
"loss": 0.3181, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.08833922261484099, |
|
"grad_norm": 0.21069642901420593, |
|
"learning_rate": 1.4735983564590783e-05, |
|
"loss": 0.4617, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.08844630046043474, |
|
"grad_norm": 0.16968613862991333, |
|
"learning_rate": 1.4572256830070497e-05, |
|
"loss": 0.4204, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.08855337830602848, |
|
"grad_norm": 0.16650810837745667, |
|
"learning_rate": 1.4409373232886702e-05, |
|
"loss": 0.3315, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.08866045615162223, |
|
"grad_norm": 0.1736612319946289, |
|
"learning_rate": 1.4247334380634792e-05, |
|
"loss": 0.3739, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.08876753399721597, |
|
"grad_norm": 0.24995604157447815, |
|
"learning_rate": 1.4086141872572789e-05, |
|
"loss": 0.4213, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.08887461184280972, |
|
"grad_norm": 0.1580914556980133, |
|
"learning_rate": 1.3925797299605647e-05, |
|
"loss": 0.2874, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.08898168968840348, |
|
"grad_norm": 0.19272121787071228, |
|
"learning_rate": 1.3766302244269624e-05, |
|
"loss": 0.4418, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.08908876753399722, |
|
"grad_norm": 0.15801607072353363, |
|
"learning_rate": 1.3607658280716473e-05, |
|
"loss": 0.3175, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.08919584537959097, |
|
"grad_norm": 0.1594337821006775, |
|
"learning_rate": 1.3449866974698122e-05, |
|
"loss": 0.2932, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.08930292322518471, |
|
"grad_norm": 0.13377048075199127, |
|
"learning_rate": 1.3292929883550998e-05, |
|
"loss": 0.2479, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.08941000107077846, |
|
"grad_norm": 0.16444596648216248, |
|
"learning_rate": 1.3136848556180892e-05, |
|
"loss": 0.3236, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.0895170789163722, |
|
"grad_norm": 0.1959199160337448, |
|
"learning_rate": 1.2981624533047432e-05, |
|
"loss": 0.4308, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.08962415676196595, |
|
"grad_norm": 0.16385279595851898, |
|
"learning_rate": 1.2827259346149122e-05, |
|
"loss": 0.3871, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.08973123460755969, |
|
"grad_norm": 0.12947721779346466, |
|
"learning_rate": 1.2673754519008008e-05, |
|
"loss": 0.2712, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.08983831245315345, |
|
"grad_norm": 0.16591428220272064, |
|
"learning_rate": 1.2521111566654731e-05, |
|
"loss": 0.2512, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.0899453902987472, |
|
"grad_norm": 0.14401817321777344, |
|
"learning_rate": 1.2369331995613665e-05, |
|
"loss": 0.2989, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.09005246814434094, |
|
"grad_norm": 0.18844856321811676, |
|
"learning_rate": 1.2218417303887842e-05, |
|
"loss": 0.3036, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.09015954598993468, |
|
"grad_norm": 0.17315539717674255, |
|
"learning_rate": 1.206836898094439e-05, |
|
"loss": 0.3203, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.09026662383552843, |
|
"grad_norm": 0.1839323788881302, |
|
"learning_rate": 1.191918850769964e-05, |
|
"loss": 0.3624, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.09037370168112217, |
|
"grad_norm": 0.19954797625541687, |
|
"learning_rate": 1.1770877356504683e-05, |
|
"loss": 0.4046, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.09048077952671592, |
|
"grad_norm": 0.17619667947292328, |
|
"learning_rate": 1.1623436991130654e-05, |
|
"loss": 0.328, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.09058785737230966, |
|
"grad_norm": 0.14954163134098053, |
|
"learning_rate": 1.1476868866754486e-05, |
|
"loss": 0.2961, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.09069493521790342, |
|
"grad_norm": 0.1400119662284851, |
|
"learning_rate": 1.1331174429944347e-05, |
|
"loss": 0.26, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.09080201306349717, |
|
"grad_norm": 0.15237313508987427, |
|
"learning_rate": 1.1186355118645554e-05, |
|
"loss": 0.3014, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.09090909090909091, |
|
"grad_norm": 0.1653870940208435, |
|
"learning_rate": 1.1042412362166222e-05, |
|
"loss": 0.2426, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.09101616875468466, |
|
"grad_norm": 0.18718279898166656, |
|
"learning_rate": 1.0899347581163221e-05, |
|
"loss": 0.3714, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.0911232466002784, |
|
"grad_norm": 0.1616310328245163, |
|
"learning_rate": 1.0757162187628222e-05, |
|
"loss": 0.3011, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.09123032444587215, |
|
"grad_norm": 0.17707855999469757, |
|
"learning_rate": 1.0615857584873623e-05, |
|
"loss": 0.3171, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.09133740229146589, |
|
"grad_norm": 0.16873344779014587, |
|
"learning_rate": 1.0475435167518843e-05, |
|
"loss": 0.2675, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.09144448013705964, |
|
"grad_norm": 0.17603348195552826, |
|
"learning_rate": 1.0335896321476413e-05, |
|
"loss": 0.4121, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.0915515579826534, |
|
"grad_norm": 0.1532546430826187, |
|
"learning_rate": 1.0197242423938446e-05, |
|
"loss": 0.2609, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.09165863582824714, |
|
"grad_norm": 0.23365063965320587, |
|
"learning_rate": 1.0059474843362892e-05, |
|
"loss": 0.3109, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.09176571367384088, |
|
"grad_norm": 0.20792338252067566, |
|
"learning_rate": 9.922594939460194e-06, |
|
"loss": 0.4347, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.09187279151943463, |
|
"grad_norm": 0.15432211756706238, |
|
"learning_rate": 9.786604063179728e-06, |
|
"loss": 0.343, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.09197986936502837, |
|
"grad_norm": 0.1996994912624359, |
|
"learning_rate": 9.651503556696516e-06, |
|
"loss": 0.3304, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.09208694721062212, |
|
"grad_norm": 0.15036849677562714, |
|
"learning_rate": 9.517294753398064e-06, |
|
"loss": 0.2859, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.09219402505621586, |
|
"grad_norm": 0.163441002368927, |
|
"learning_rate": 9.383978977871021e-06, |
|
"loss": 0.2546, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.09230110290180961, |
|
"grad_norm": 0.17922081053256989, |
|
"learning_rate": 9.251557545888312e-06, |
|
"loss": 0.435, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.09240818074740337, |
|
"grad_norm": 0.16963575780391693, |
|
"learning_rate": 9.120031764395987e-06, |
|
"loss": 0.3112, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.09251525859299711, |
|
"grad_norm": 0.1739087551832199, |
|
"learning_rate": 8.989402931500434e-06, |
|
"loss": 0.3357, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.09262233643859086, |
|
"grad_norm": 0.15350165963172913, |
|
"learning_rate": 8.85967233645547e-06, |
|
"loss": 0.2912, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.0927294142841846, |
|
"grad_norm": 0.14361527562141418, |
|
"learning_rate": 8.730841259649725e-06, |
|
"loss": 0.2633, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.09283649212977835, |
|
"grad_norm": 0.16577735543251038, |
|
"learning_rate": 8.602910972593892e-06, |
|
"loss": 0.3942, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.09294356997537209, |
|
"grad_norm": 0.1609407663345337, |
|
"learning_rate": 8.475882737908248e-06, |
|
"loss": 0.3476, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.09305064782096584, |
|
"grad_norm": 0.17923296988010406, |
|
"learning_rate": 8.34975780931021e-06, |
|
"loss": 0.3774, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.09315772566655958, |
|
"grad_norm": 0.14731475710868835, |
|
"learning_rate": 8.224537431601886e-06, |
|
"loss": 0.2685, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.09326480351215334, |
|
"grad_norm": 0.1950971633195877, |
|
"learning_rate": 8.100222840657878e-06, |
|
"loss": 0.3933, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.09337188135774709, |
|
"grad_norm": 0.2104734629392624, |
|
"learning_rate": 7.976815263412963e-06, |
|
"loss": 0.305, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.09347895920334083, |
|
"grad_norm": 0.15162058174610138, |
|
"learning_rate": 7.854315917850163e-06, |
|
"loss": 0.296, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.09358603704893458, |
|
"grad_norm": 0.17476332187652588, |
|
"learning_rate": 7.73272601298851e-06, |
|
"loss": 0.2836, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.09369311489452832, |
|
"grad_norm": 0.1703297197818756, |
|
"learning_rate": 7.612046748871327e-06, |
|
"loss": 0.2873, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.09380019274012207, |
|
"grad_norm": 0.16354651749134064, |
|
"learning_rate": 7.492279316554207e-06, |
|
"loss": 0.3141, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.09390727058571581, |
|
"grad_norm": 0.19520509243011475, |
|
"learning_rate": 7.3734248980933395e-06, |
|
"loss": 0.397, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.09401434843130957, |
|
"grad_norm": 0.15677590668201447, |
|
"learning_rate": 7.255484666533874e-06, |
|
"loss": 0.3549, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.09412142627690331, |
|
"grad_norm": 0.19243308901786804, |
|
"learning_rate": 7.138459785898266e-06, |
|
"loss": 0.4026, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.09422850412249706, |
|
"grad_norm": 0.15591202676296234, |
|
"learning_rate": 7.022351411174866e-06, |
|
"loss": 0.3088, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.0943355819680908, |
|
"grad_norm": 0.17848993837833405, |
|
"learning_rate": 6.907160688306425e-06, |
|
"loss": 0.2836, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.09444265981368455, |
|
"grad_norm": 0.18109548091888428, |
|
"learning_rate": 6.7928887541789055e-06, |
|
"loss": 0.3225, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.0945497376592783, |
|
"grad_norm": 0.13605234026908875, |
|
"learning_rate": 6.679536736610137e-06, |
|
"loss": 0.2776, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.09465681550487204, |
|
"grad_norm": 0.12738342583179474, |
|
"learning_rate": 6.5671057543387985e-06, |
|
"loss": 0.2155, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.09476389335046578, |
|
"grad_norm": 0.19172219932079315, |
|
"learning_rate": 6.455596917013273e-06, |
|
"loss": 0.2594, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.09487097119605954, |
|
"grad_norm": 0.16190144419670105, |
|
"learning_rate": 6.345011325180772e-06, |
|
"loss": 0.311, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.09497804904165329, |
|
"grad_norm": 0.136839359998703, |
|
"learning_rate": 6.235350070276447e-06, |
|
"loss": 0.2523, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.09508512688724703, |
|
"grad_norm": 0.12164784222841263, |
|
"learning_rate": 6.126614234612593e-06, |
|
"loss": 0.2399, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.09519220473284078, |
|
"grad_norm": 0.15871037542819977, |
|
"learning_rate": 6.018804891368035e-06, |
|
"loss": 0.3429, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.09529928257843452, |
|
"grad_norm": 0.15956546366214752, |
|
"learning_rate": 5.911923104577455e-06, |
|
"loss": 0.3148, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.09540636042402827, |
|
"grad_norm": 0.17873486876487732, |
|
"learning_rate": 5.805969929120947e-06, |
|
"loss": 0.3492, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.09551343826962201, |
|
"grad_norm": 0.18274949491024017, |
|
"learning_rate": 5.700946410713548e-06, |
|
"loss": 0.4022, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.09562051611521576, |
|
"grad_norm": 0.13471592962741852, |
|
"learning_rate": 5.5968535858950345e-06, |
|
"loss": 0.2356, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.09572759396080951, |
|
"grad_norm": 0.16178227961063385, |
|
"learning_rate": 5.49369248201953e-06, |
|
"loss": 0.2734, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.09583467180640326, |
|
"grad_norm": 0.23086000978946686, |
|
"learning_rate": 5.39146411724547e-06, |
|
"loss": 0.4373, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.095941749651997, |
|
"grad_norm": 0.1594802439212799, |
|
"learning_rate": 5.290169500525577e-06, |
|
"loss": 0.3122, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.09604882749759075, |
|
"grad_norm": 0.1993282437324524, |
|
"learning_rate": 5.189809631596798e-06, |
|
"loss": 0.4464, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.0961559053431845, |
|
"grad_norm": 0.1615990251302719, |
|
"learning_rate": 5.0903855009705514e-06, |
|
"loss": 0.2872, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.09626298318877824, |
|
"grad_norm": 0.1525048166513443, |
|
"learning_rate": 4.991898089922819e-06, |
|
"loss": 0.3361, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.09637006103437198, |
|
"grad_norm": 0.1328526735305786, |
|
"learning_rate": 4.8943483704846475e-06, |
|
"loss": 0.2255, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.09647713887996573, |
|
"grad_norm": 0.16284851729869843, |
|
"learning_rate": 4.797737305432337e-06, |
|
"loss": 0.3079, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.09658421672555949, |
|
"grad_norm": 0.16765619814395905, |
|
"learning_rate": 4.702065848278126e-06, |
|
"loss": 0.3512, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.09669129457115323, |
|
"grad_norm": 0.136622354388237, |
|
"learning_rate": 4.607334943260655e-06, |
|
"loss": 0.2413, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.09679837241674698, |
|
"grad_norm": 0.22073903679847717, |
|
"learning_rate": 4.513545525335705e-06, |
|
"loss": 0.4166, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.09690545026234072, |
|
"grad_norm": 0.15773172676563263, |
|
"learning_rate": 4.420698520166988e-06, |
|
"loss": 0.303, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.09701252810793447, |
|
"grad_norm": 0.17100340127944946, |
|
"learning_rate": 4.328794844116946e-06, |
|
"loss": 0.2892, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.09711960595352821, |
|
"grad_norm": 0.17342492938041687, |
|
"learning_rate": 4.237835404237778e-06, |
|
"loss": 0.2773, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.09722668379912196, |
|
"grad_norm": 0.1771211475133896, |
|
"learning_rate": 4.147821098262405e-06, |
|
"loss": 0.3796, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.0973337616447157, |
|
"grad_norm": 0.13509005308151245, |
|
"learning_rate": 4.0587528145957235e-06, |
|
"loss": 0.2839, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.09744083949030946, |
|
"grad_norm": 0.17076431214809418, |
|
"learning_rate": 3.970631432305694e-06, |
|
"loss": 0.3372, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.0975479173359032, |
|
"grad_norm": 0.16045229136943817, |
|
"learning_rate": 3.883457821114811e-06, |
|
"loss": 0.2402, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.09765499518149695, |
|
"grad_norm": 0.18019679188728333, |
|
"learning_rate": 3.797232841391407e-06, |
|
"loss": 0.4094, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.0977620730270907, |
|
"grad_norm": 0.19246861338615417, |
|
"learning_rate": 3.711957344141237e-06, |
|
"loss": 0.3307, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.09786915087268444, |
|
"grad_norm": 0.1618802696466446, |
|
"learning_rate": 3.627632170999029e-06, |
|
"loss": 0.2785, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.09797622871827819, |
|
"grad_norm": 0.17104919254779816, |
|
"learning_rate": 3.5442581542201923e-06, |
|
"loss": 0.3925, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.09808330656387193, |
|
"grad_norm": 0.15704232454299927, |
|
"learning_rate": 3.461836116672612e-06, |
|
"loss": 0.2819, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.09819038440946568, |
|
"grad_norm": 0.17847512662410736, |
|
"learning_rate": 3.380366871828522e-06, |
|
"loss": 0.4173, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.09829746225505943, |
|
"grad_norm": 0.16614320874214172, |
|
"learning_rate": 3.2998512237565005e-06, |
|
"loss": 0.3005, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.09840454010065318, |
|
"grad_norm": 0.1553928405046463, |
|
"learning_rate": 3.2202899671134546e-06, |
|
"loss": 0.3145, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.09851161794624692, |
|
"grad_norm": 0.15356747806072235, |
|
"learning_rate": 3.1416838871368924e-06, |
|
"loss": 0.2905, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.09861869579184067, |
|
"grad_norm": 0.1633521169424057, |
|
"learning_rate": 3.064033759637064e-06, |
|
"loss": 0.3753, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.09872577363743441, |
|
"grad_norm": 0.16967448592185974, |
|
"learning_rate": 2.9873403509894203e-06, |
|
"loss": 0.3764, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.09883285148302816, |
|
"grad_norm": 0.1915234923362732, |
|
"learning_rate": 2.9116044181269007e-06, |
|
"loss": 0.3296, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.0989399293286219, |
|
"grad_norm": 0.2066269963979721, |
|
"learning_rate": 2.836826708532603e-06, |
|
"loss": 0.3912, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.09904700717421565, |
|
"grad_norm": 0.12314484268426895, |
|
"learning_rate": 2.7630079602323442e-06, |
|
"loss": 0.2109, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.09915408501980941, |
|
"grad_norm": 0.2099624127149582, |
|
"learning_rate": 2.690148901787337e-06, |
|
"loss": 0.3809, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.09926116286540315, |
|
"grad_norm": 0.16384465992450714, |
|
"learning_rate": 2.618250252287113e-06, |
|
"loss": 0.3344, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.0993682407109969, |
|
"grad_norm": 0.2039898931980133, |
|
"learning_rate": 2.5473127213422763e-06, |
|
"loss": 0.3398, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.09947531855659064, |
|
"grad_norm": 0.17251908779144287, |
|
"learning_rate": 2.4773370090776626e-06, |
|
"loss": 0.2963, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.09958239640218439, |
|
"grad_norm": 0.19402608275413513, |
|
"learning_rate": 2.4083238061252567e-06, |
|
"loss": 0.5081, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.09968947424777813, |
|
"grad_norm": 0.15272395312786102, |
|
"learning_rate": 2.3402737936175425e-06, |
|
"loss": 0.3675, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.09979655209337188, |
|
"grad_norm": 0.16620352864265442, |
|
"learning_rate": 2.273187643180652e-06, |
|
"loss": 0.3654, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.09990362993896562, |
|
"grad_norm": 0.2133980691432953, |
|
"learning_rate": 2.2070660169278166e-06, |
|
"loss": 0.3945, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.10001070778455938, |
|
"grad_norm": 0.13938206434249878, |
|
"learning_rate": 2.141909567452793e-06, |
|
"loss": 0.2548, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.10011778563015312, |
|
"grad_norm": 0.1559179276227951, |
|
"learning_rate": 2.0777189378234143e-06, |
|
"loss": 0.3075, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.10022486347574687, |
|
"grad_norm": 0.14096151292324066, |
|
"learning_rate": 2.014494761575314e-06, |
|
"loss": 0.2429, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.10033194132134061, |
|
"grad_norm": 0.14772740006446838, |
|
"learning_rate": 1.9522376627055583e-06, |
|
"loss": 0.3397, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.10043901916693436, |
|
"grad_norm": 0.1633581519126892, |
|
"learning_rate": 1.8909482556666024e-06, |
|
"loss": 0.3296, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.1005460970125281, |
|
"grad_norm": 0.1756487637758255, |
|
"learning_rate": 1.8306271453601199e-06, |
|
"loss": 0.3387, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.10065317485812185, |
|
"grad_norm": 0.1336044818162918, |
|
"learning_rate": 1.771274927131139e-06, |
|
"loss": 0.2621, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.10076025270371561, |
|
"grad_norm": 0.1638735979795456, |
|
"learning_rate": 1.712892186762083e-06, |
|
"loss": 0.3286, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.10086733054930935, |
|
"grad_norm": 0.16432860493659973, |
|
"learning_rate": 1.6554795004670388e-06, |
|
"loss": 0.4136, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.1009744083949031, |
|
"grad_norm": 0.20629476010799408, |
|
"learning_rate": 1.5990374348860305e-06, |
|
"loss": 0.332, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.10108148624049684, |
|
"grad_norm": 0.14184875786304474, |
|
"learning_rate": 1.543566547079467e-06, |
|
"loss": 0.2722, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.10118856408609059, |
|
"grad_norm": 0.19237804412841797, |
|
"learning_rate": 1.4890673845226133e-06, |
|
"loss": 0.401, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.10129564193168433, |
|
"grad_norm": 0.17715074121952057, |
|
"learning_rate": 1.4355404851001952e-06, |
|
"loss": 0.4161, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.10140271977727808, |
|
"grad_norm": 0.16229744255542755, |
|
"learning_rate": 1.3829863771011253e-06, |
|
"loss": 0.2394, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.10150979762287182, |
|
"grad_norm": 0.2094830572605133, |
|
"learning_rate": 1.3314055792131964e-06, |
|
"loss": 0.3434, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.10161687546846558, |
|
"grad_norm": 0.1709808111190796, |
|
"learning_rate": 1.280798600518085e-06, |
|
"loss": 0.3196, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.10172395331405933, |
|
"grad_norm": 0.1492084413766861, |
|
"learning_rate": 1.231165940486234e-06, |
|
"loss": 0.2903, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.10183103115965307, |
|
"grad_norm": 0.1893269568681717, |
|
"learning_rate": 1.1825080889719563e-06, |
|
"loss": 0.3737, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.10193810900524682, |
|
"grad_norm": 0.15027467906475067, |
|
"learning_rate": 1.134825526208605e-06, |
|
"loss": 0.2198, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.10204518685084056, |
|
"grad_norm": 0.1478700190782547, |
|
"learning_rate": 1.0881187228038215e-06, |
|
"loss": 0.2796, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.1021522646964343, |
|
"grad_norm": 0.16411510109901428, |
|
"learning_rate": 1.0423881397349068e-06, |
|
"loss": 0.358, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.10225934254202805, |
|
"grad_norm": 0.1691845953464508, |
|
"learning_rate": 9.976342283442463e-07, |
|
"loss": 0.3231, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.1023664203876218, |
|
"grad_norm": 0.21002455055713654, |
|
"learning_rate": 9.538574303348813e-07, |
|
"loss": 0.3972, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.10247349823321555, |
|
"grad_norm": 0.148806631565094, |
|
"learning_rate": 9.110581777661331e-07, |
|
"loss": 0.2966, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.1025805760788093, |
|
"grad_norm": 0.1439158171415329, |
|
"learning_rate": 8.692368930493521e-07, |
|
"loss": 0.2605, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.10268765392440304, |
|
"grad_norm": 0.15770378708839417, |
|
"learning_rate": 8.283939889437209e-07, |
|
"loss": 0.3339, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.10279473176999679, |
|
"grad_norm": 0.16292540729045868, |
|
"learning_rate": 7.885298685522235e-07, |
|
"loss": 0.311, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.10290180961559053, |
|
"grad_norm": 0.20163190364837646, |
|
"learning_rate": 7.496449253176274e-07, |
|
"loss": 0.3665, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.10300888746118428, |
|
"grad_norm": 0.1703958362340927, |
|
"learning_rate": 7.117395430186414e-07, |
|
"loss": 0.3539, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.10311596530677802, |
|
"grad_norm": 0.2039760798215866, |
|
"learning_rate": 6.748140957660631e-07, |
|
"loss": 0.3793, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.10322304315237177, |
|
"grad_norm": 0.18234048783779144, |
|
"learning_rate": 6.388689479991605e-07, |
|
"loss": 0.3457, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.10333012099796553, |
|
"grad_norm": 0.14040732383728027, |
|
"learning_rate": 6.039044544820404e-07, |
|
"loss": 0.2796, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.10343719884355927, |
|
"grad_norm": 0.13894395530223846, |
|
"learning_rate": 5.699209603001076e-07, |
|
"loss": 0.3024, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.10354427668915302, |
|
"grad_norm": 0.17470873892307281, |
|
"learning_rate": 5.369188008567672e-07, |
|
"loss": 0.3419, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.10365135453474676, |
|
"grad_norm": 0.17506030201911926, |
|
"learning_rate": 5.048983018699827e-07, |
|
"loss": 0.3442, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.1037584323803405, |
|
"grad_norm": 0.17124956846237183, |
|
"learning_rate": 4.738597793691679e-07, |
|
"loss": 0.3419, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.10386551022593425, |
|
"grad_norm": 0.15624070167541504, |
|
"learning_rate": 4.438035396920004e-07, |
|
"loss": 0.2658, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.103972588071528, |
|
"grad_norm": 0.2304699569940567, |
|
"learning_rate": 4.1472987948143473e-07, |
|
"loss": 0.5446, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.10407966591712174, |
|
"grad_norm": 0.14617536962032318, |
|
"learning_rate": 3.866390856827495e-07, |
|
"loss": 0.3385, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.1041867437627155, |
|
"grad_norm": 0.17763929069042206, |
|
"learning_rate": 3.595314355407609e-07, |
|
"loss": 0.3553, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.10429382160830925, |
|
"grad_norm": 0.14324170351028442, |
|
"learning_rate": 3.3340719659701313e-07, |
|
"loss": 0.3078, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.10440089945390299, |
|
"grad_norm": 0.1634160429239273, |
|
"learning_rate": 3.0826662668720364e-07, |
|
"loss": 0.3353, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.10450797729949673, |
|
"grad_norm": 0.17448119819164276, |
|
"learning_rate": 2.841099739386066e-07, |
|
"loss": 0.3754, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.10461505514509048, |
|
"grad_norm": 0.15392538905143738, |
|
"learning_rate": 2.609374767676309e-07, |
|
"loss": 0.338, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.10472213299068422, |
|
"grad_norm": 0.2160252183675766, |
|
"learning_rate": 2.387493638774774e-07, |
|
"loss": 0.4477, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.10482921083627797, |
|
"grad_norm": 0.16572636365890503, |
|
"learning_rate": 2.175458542558517e-07, |
|
"loss": 0.3833, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.10493628868187171, |
|
"grad_norm": 0.17389637231826782, |
|
"learning_rate": 1.973271571728441e-07, |
|
"loss": 0.2881, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.10504336652746547, |
|
"grad_norm": 0.16537849605083466, |
|
"learning_rate": 1.7809347217881966e-07, |
|
"loss": 0.3809, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.10515044437305922, |
|
"grad_norm": 0.1813080906867981, |
|
"learning_rate": 1.598449891024978e-07, |
|
"loss": 0.3269, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.10525752221865296, |
|
"grad_norm": 0.1730542629957199, |
|
"learning_rate": 1.425818880490315e-07, |
|
"loss": 0.3606, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.10536460006424671, |
|
"grad_norm": 0.2064792513847351, |
|
"learning_rate": 1.2630433939825327e-07, |
|
"loss": 0.3393, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.10547167790984045, |
|
"grad_norm": 0.15513303875923157, |
|
"learning_rate": 1.1101250380300965e-07, |
|
"loss": 0.3034, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.1055787557554342, |
|
"grad_norm": 0.17312531173229218, |
|
"learning_rate": 9.670653218752934e-08, |
|
"loss": 0.3564, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.10568583360102794, |
|
"grad_norm": 0.16997037827968597, |
|
"learning_rate": 8.33865657459909e-08, |
|
"loss": 0.4089, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.10579291144662169, |
|
"grad_norm": 0.2006881982088089, |
|
"learning_rate": 7.105273594107953e-08, |
|
"loss": 0.3803, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.10589998929221545, |
|
"grad_norm": 0.17012691497802734, |
|
"learning_rate": 5.970516450271025e-08, |
|
"loss": 0.3733, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.10600706713780919, |
|
"grad_norm": 0.1657508760690689, |
|
"learning_rate": 4.934396342684e-08, |
|
"loss": 0.3675, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.10611414498340294, |
|
"grad_norm": 0.15509915351867676, |
|
"learning_rate": 3.996923497434635e-08, |
|
"loss": 0.367, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.10622122282899668, |
|
"grad_norm": 0.1795806735754013, |
|
"learning_rate": 3.1581071670006015e-08, |
|
"loss": 0.4073, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.10632830067459043, |
|
"grad_norm": 0.1823895275592804, |
|
"learning_rate": 2.417955630159563e-08, |
|
"loss": 0.3606, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.10643537852018417, |
|
"grad_norm": 0.17259643971920013, |
|
"learning_rate": 1.7764761919103477e-08, |
|
"loss": 0.3723, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.10654245636577792, |
|
"grad_norm": 0.15075139701366425, |
|
"learning_rate": 1.2336751833941229e-08, |
|
"loss": 0.2497, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.10664953421137166, |
|
"grad_norm": 0.1535639613866806, |
|
"learning_rate": 7.895579618388827e-09, |
|
"loss": 0.3089, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.10675661205696542, |
|
"grad_norm": 0.17250818014144897, |
|
"learning_rate": 4.4412891050171765e-09, |
|
"loss": 0.3135, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.10686368990255916, |
|
"grad_norm": 0.1399751603603363, |
|
"learning_rate": 1.973914386288467e-09, |
|
"loss": 0.2826, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.10697076774815291, |
|
"grad_norm": 0.2154352217912674, |
|
"learning_rate": 4.934798141786879e-10, |
|
"loss": 0.4276, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.10707784559374665, |
|
"grad_norm": 0.21260492503643036, |
|
"learning_rate": 0.0, |
|
"loss": 0.4272, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.10707784559374665, |
|
"eval_loss": 0.3318476676940918, |
|
"eval_runtime": 355.7741, |
|
"eval_samples_per_second": 5.529, |
|
"eval_steps_per_second": 5.529, |
|
"step": 1000 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.6666433267734938e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|