Jambatypus-v0.1 / checkpoint-336 /trainer_state.json
mlabonne's picture
Upload folder using huggingface_hub
7b17ba0 verified
raw
history blame
54.9 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9981433345711103,
"eval_steps": 84,
"global_step": 336,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 0.297607958316803,
"learning_rate": 2e-05,
"loss": 0.5915,
"step": 1
},
{
"epoch": 0.0,
"eval_loss": 1.027875304222107,
"eval_runtime": 319.0994,
"eval_samples_per_second": 0.78,
"eval_steps_per_second": 0.78,
"step": 1
},
{
"epoch": 0.01,
"grad_norm": 0.3444654643535614,
"learning_rate": 4e-05,
"loss": 0.5941,
"step": 2
},
{
"epoch": 0.01,
"grad_norm": 0.31323298811912537,
"learning_rate": 6e-05,
"loss": 0.5986,
"step": 3
},
{
"epoch": 0.01,
"grad_norm": 0.3083553612232208,
"learning_rate": 8e-05,
"loss": 0.589,
"step": 4
},
{
"epoch": 0.01,
"grad_norm": 0.3427445590496063,
"learning_rate": 0.0001,
"loss": 0.6559,
"step": 5
},
{
"epoch": 0.02,
"grad_norm": 0.26026275753974915,
"learning_rate": 0.00012,
"loss": 0.6444,
"step": 6
},
{
"epoch": 0.02,
"grad_norm": 0.22449900209903717,
"learning_rate": 0.00014,
"loss": 0.5621,
"step": 7
},
{
"epoch": 0.02,
"grad_norm": 0.18667733669281006,
"learning_rate": 0.00016,
"loss": 0.574,
"step": 8
},
{
"epoch": 0.03,
"grad_norm": 0.1876465231180191,
"learning_rate": 0.00018,
"loss": 0.6105,
"step": 9
},
{
"epoch": 0.03,
"grad_norm": 0.2330338954925537,
"learning_rate": 0.0002,
"loss": 0.5793,
"step": 10
},
{
"epoch": 0.03,
"grad_norm": 0.18704406917095184,
"learning_rate": 0.00019999535665248002,
"loss": 0.5431,
"step": 11
},
{
"epoch": 0.04,
"grad_norm": 0.20125611126422882,
"learning_rate": 0.0001999814270411335,
"loss": 0.5219,
"step": 12
},
{
"epoch": 0.04,
"grad_norm": 0.17054004967212677,
"learning_rate": 0.000199958212459561,
"loss": 0.5022,
"step": 13
},
{
"epoch": 0.04,
"grad_norm": 0.20665132999420166,
"learning_rate": 0.00019992571506363,
"loss": 0.5907,
"step": 14
},
{
"epoch": 0.04,
"grad_norm": 0.21675613522529602,
"learning_rate": 0.00019988393787127441,
"loss": 0.4667,
"step": 15
},
{
"epoch": 0.05,
"grad_norm": 0.20350293815135956,
"learning_rate": 0.0001998328847622148,
"loss": 0.5533,
"step": 16
},
{
"epoch": 0.05,
"grad_norm": 0.18382889032363892,
"learning_rate": 0.00019977256047759765,
"loss": 0.5022,
"step": 17
},
{
"epoch": 0.05,
"grad_norm": 0.1696760207414627,
"learning_rate": 0.00019970297061955533,
"loss": 0.5381,
"step": 18
},
{
"epoch": 0.06,
"grad_norm": 0.19592055678367615,
"learning_rate": 0.00019962412165068573,
"loss": 0.5733,
"step": 19
},
{
"epoch": 0.06,
"grad_norm": 0.17972029745578766,
"learning_rate": 0.00019953602089345217,
"loss": 0.5891,
"step": 20
},
{
"epoch": 0.06,
"grad_norm": 0.19014880061149597,
"learning_rate": 0.0001994386765295032,
"loss": 0.5629,
"step": 21
},
{
"epoch": 0.07,
"grad_norm": 0.15824897587299347,
"learning_rate": 0.00019933209759891317,
"loss": 0.5627,
"step": 22
},
{
"epoch": 0.07,
"grad_norm": 0.17133353650569916,
"learning_rate": 0.00019921629399934223,
"loss": 0.5542,
"step": 23
},
{
"epoch": 0.07,
"grad_norm": 0.13738161325454712,
"learning_rate": 0.00019909127648511755,
"loss": 0.4443,
"step": 24
},
{
"epoch": 0.07,
"grad_norm": 0.16617180407047272,
"learning_rate": 0.0001989570566662345,
"loss": 0.5469,
"step": 25
},
{
"epoch": 0.08,
"grad_norm": 0.13969440758228302,
"learning_rate": 0.00019881364700727823,
"loss": 0.5141,
"step": 26
},
{
"epoch": 0.08,
"grad_norm": 0.1598738580942154,
"learning_rate": 0.0001986610608262665,
"loss": 0.5598,
"step": 27
},
{
"epoch": 0.08,
"grad_norm": 0.14080321788787842,
"learning_rate": 0.00019849931229341258,
"loss": 0.4978,
"step": 28
},
{
"epoch": 0.09,
"grad_norm": 0.16124503314495087,
"learning_rate": 0.00019832841642980945,
"loss": 0.613,
"step": 29
},
{
"epoch": 0.09,
"grad_norm": 0.1537931114435196,
"learning_rate": 0.00019814838910603481,
"loss": 0.4913,
"step": 30
},
{
"epoch": 0.09,
"grad_norm": 0.16199100017547607,
"learning_rate": 0.00019795924704067721,
"loss": 0.4872,
"step": 31
},
{
"epoch": 0.1,
"grad_norm": 0.13554096221923828,
"learning_rate": 0.00019776100779878345,
"loss": 0.4609,
"step": 32
},
{
"epoch": 0.1,
"grad_norm": 0.18090932071208954,
"learning_rate": 0.00019755368979022732,
"loss": 0.4582,
"step": 33
},
{
"epoch": 0.1,
"grad_norm": 0.14535771310329437,
"learning_rate": 0.00019733731226800015,
"loss": 0.5359,
"step": 34
},
{
"epoch": 0.1,
"grad_norm": 0.1427253782749176,
"learning_rate": 0.00019711189532642243,
"loss": 0.5324,
"step": 35
},
{
"epoch": 0.11,
"grad_norm": 0.13935472071170807,
"learning_rate": 0.00019687745989927823,
"loss": 0.5106,
"step": 36
},
{
"epoch": 0.11,
"grad_norm": 0.1388338804244995,
"learning_rate": 0.00019663402775787066,
"loss": 0.5634,
"step": 37
},
{
"epoch": 0.11,
"grad_norm": 0.13948731124401093,
"learning_rate": 0.00019638162150900027,
"loss": 0.5058,
"step": 38
},
{
"epoch": 0.12,
"grad_norm": 0.13687050342559814,
"learning_rate": 0.00019612026459286578,
"loss": 0.5806,
"step": 39
},
{
"epoch": 0.12,
"grad_norm": 0.1311887949705124,
"learning_rate": 0.00019584998128088684,
"loss": 0.4552,
"step": 40
},
{
"epoch": 0.12,
"grad_norm": 0.15644784271717072,
"learning_rate": 0.0001955707966734505,
"loss": 0.4424,
"step": 41
},
{
"epoch": 0.12,
"grad_norm": 0.14039994776248932,
"learning_rate": 0.00019528273669757972,
"loss": 0.4905,
"step": 42
},
{
"epoch": 0.13,
"grad_norm": 0.14709696173667908,
"learning_rate": 0.0001949858281045261,
"loss": 0.5655,
"step": 43
},
{
"epoch": 0.13,
"grad_norm": 0.1661250740289688,
"learning_rate": 0.00019468009846728513,
"loss": 0.5141,
"step": 44
},
{
"epoch": 0.13,
"grad_norm": 0.3046470582485199,
"learning_rate": 0.00019436557617803595,
"loss": 0.4792,
"step": 45
},
{
"epoch": 0.14,
"grad_norm": 0.15159915387630463,
"learning_rate": 0.00019404229044550433,
"loss": 0.5077,
"step": 46
},
{
"epoch": 0.14,
"grad_norm": 0.15115606784820557,
"learning_rate": 0.00019371027129225042,
"loss": 0.4915,
"step": 47
},
{
"epoch": 0.14,
"grad_norm": 0.14040274918079376,
"learning_rate": 0.0001933695495518804,
"loss": 0.5048,
"step": 48
},
{
"epoch": 0.15,
"grad_norm": 0.14947518706321716,
"learning_rate": 0.00019302015686618326,
"loss": 0.5124,
"step": 49
},
{
"epoch": 0.15,
"grad_norm": 0.13805006444454193,
"learning_rate": 0.0001926621256821922,
"loss": 0.4455,
"step": 50
},
{
"epoch": 0.15,
"grad_norm": 0.14188264310359955,
"learning_rate": 0.00019229548924917146,
"loss": 0.5244,
"step": 51
},
{
"epoch": 0.15,
"grad_norm": 0.1314387321472168,
"learning_rate": 0.00019192028161552847,
"loss": 0.5251,
"step": 52
},
{
"epoch": 0.16,
"grad_norm": 0.1659722477197647,
"learning_rate": 0.0001915365376256519,
"loss": 0.4851,
"step": 53
},
{
"epoch": 0.16,
"grad_norm": 0.1405472755432129,
"learning_rate": 0.00019114429291667583,
"loss": 0.463,
"step": 54
},
{
"epoch": 0.16,
"grad_norm": 0.14347247779369354,
"learning_rate": 0.00019074358391517023,
"loss": 0.4491,
"step": 55
},
{
"epoch": 0.17,
"grad_norm": 0.14897900819778442,
"learning_rate": 0.00019033444783375804,
"loss": 0.4942,
"step": 56
},
{
"epoch": 0.17,
"grad_norm": 0.13556955754756927,
"learning_rate": 0.00018991692266765947,
"loss": 0.4755,
"step": 57
},
{
"epoch": 0.17,
"grad_norm": 0.16432009637355804,
"learning_rate": 0.00018949104719116332,
"loss": 0.5747,
"step": 58
},
{
"epoch": 0.18,
"grad_norm": 0.15541689097881317,
"learning_rate": 0.00018905686095402647,
"loss": 0.533,
"step": 59
},
{
"epoch": 0.18,
"grad_norm": 0.1607791632413864,
"learning_rate": 0.0001886144042778006,
"loss": 0.556,
"step": 60
},
{
"epoch": 0.18,
"grad_norm": 0.15379983186721802,
"learning_rate": 0.00018816371825208789,
"loss": 0.4549,
"step": 61
},
{
"epoch": 0.18,
"grad_norm": 0.15419824421405792,
"learning_rate": 0.0001877048447307252,
"loss": 0.4933,
"step": 62
},
{
"epoch": 0.19,
"grad_norm": 0.1353635936975479,
"learning_rate": 0.00018723782632789701,
"loss": 0.393,
"step": 63
},
{
"epoch": 0.19,
"grad_norm": 0.15725383162498474,
"learning_rate": 0.00018676270641417822,
"loss": 0.4671,
"step": 64
},
{
"epoch": 0.19,
"grad_norm": 0.12909093499183655,
"learning_rate": 0.0001862795291125063,
"loss": 0.484,
"step": 65
},
{
"epoch": 0.2,
"grad_norm": 0.14394541084766388,
"learning_rate": 0.0001857883392940837,
"loss": 0.5232,
"step": 66
},
{
"epoch": 0.2,
"grad_norm": 0.1671515554189682,
"learning_rate": 0.000185289182574211,
"loss": 0.4683,
"step": 67
},
{
"epoch": 0.2,
"grad_norm": 0.1658957600593567,
"learning_rate": 0.0001847821053080505,
"loss": 0.5072,
"step": 68
},
{
"epoch": 0.2,
"grad_norm": 0.15142019093036652,
"learning_rate": 0.00018426715458632153,
"loss": 0.5328,
"step": 69
},
{
"epoch": 0.21,
"grad_norm": 0.141921728849411,
"learning_rate": 0.00018374437823092724,
"loss": 0.4387,
"step": 70
},
{
"epoch": 0.21,
"grad_norm": 0.1730872541666031,
"learning_rate": 0.00018321382479051347,
"loss": 0.5323,
"step": 71
},
{
"epoch": 0.21,
"grad_norm": 0.13589338958263397,
"learning_rate": 0.00018267554353596025,
"loss": 0.4217,
"step": 72
},
{
"epoch": 0.22,
"grad_norm": 0.16519547998905182,
"learning_rate": 0.0001821295844558062,
"loss": 0.4498,
"step": 73
},
{
"epoch": 0.22,
"grad_norm": 0.16089589893817902,
"learning_rate": 0.0001815759982516061,
"loss": 0.5204,
"step": 74
},
{
"epoch": 0.22,
"grad_norm": 0.16409483551979065,
"learning_rate": 0.00018101483633322255,
"loss": 0.4511,
"step": 75
},
{
"epoch": 0.23,
"grad_norm": 0.16179721057415009,
"learning_rate": 0.00018044615081405153,
"loss": 0.5866,
"step": 76
},
{
"epoch": 0.23,
"grad_norm": 0.15137337148189545,
"learning_rate": 0.00017986999450618295,
"loss": 0.4728,
"step": 77
},
{
"epoch": 0.23,
"grad_norm": 0.14265336096286774,
"learning_rate": 0.00017928642091549613,
"loss": 0.4491,
"step": 78
},
{
"epoch": 0.23,
"grad_norm": 0.14860045909881592,
"learning_rate": 0.00017869548423669077,
"loss": 0.4734,
"step": 79
},
{
"epoch": 0.24,
"grad_norm": 0.14790605008602142,
"learning_rate": 0.00017809723934825405,
"loss": 0.533,
"step": 80
},
{
"epoch": 0.24,
"grad_norm": 0.1581815779209137,
"learning_rate": 0.00017749174180736442,
"loss": 0.4661,
"step": 81
},
{
"epoch": 0.24,
"grad_norm": 0.1520436853170395,
"learning_rate": 0.00017687904784473188,
"loss": 0.5292,
"step": 82
},
{
"epoch": 0.25,
"grad_norm": 0.15341107547283173,
"learning_rate": 0.00017625921435937637,
"loss": 0.5062,
"step": 83
},
{
"epoch": 0.25,
"grad_norm": 0.14248648285865784,
"learning_rate": 0.00017563229891334338,
"loss": 0.4249,
"step": 84
},
{
"epoch": 0.25,
"eval_loss": 0.9711415767669678,
"eval_runtime": 320.5008,
"eval_samples_per_second": 0.777,
"eval_steps_per_second": 0.777,
"step": 84
},
{
"epoch": 0.25,
"grad_norm": 0.16506022214889526,
"learning_rate": 0.00017499835972635856,
"loss": 0.4873,
"step": 85
},
{
"epoch": 0.26,
"grad_norm": 0.14895877242088318,
"learning_rate": 0.00017435745567042095,
"loss": 0.4774,
"step": 86
},
{
"epoch": 0.26,
"grad_norm": 0.25245463848114014,
"learning_rate": 0.00017370964626433567,
"loss": 0.5324,
"step": 87
},
{
"epoch": 0.26,
"grad_norm": 0.14706483483314514,
"learning_rate": 0.0001730549916681868,
"loss": 0.4536,
"step": 88
},
{
"epoch": 0.26,
"grad_norm": 0.14856815338134766,
"learning_rate": 0.00017239355267775018,
"loss": 0.4576,
"step": 89
},
{
"epoch": 0.27,
"grad_norm": 0.1513276994228363,
"learning_rate": 0.0001717253907188477,
"loss": 0.4286,
"step": 90
},
{
"epoch": 0.27,
"grad_norm": 0.1549597978591919,
"learning_rate": 0.00017105056784164294,
"loss": 0.4554,
"step": 91
},
{
"epoch": 0.27,
"grad_norm": 0.15850332379341125,
"learning_rate": 0.00017036914671487852,
"loss": 0.454,
"step": 92
},
{
"epoch": 0.28,
"grad_norm": 0.15920379757881165,
"learning_rate": 0.00016968119062005642,
"loss": 0.5333,
"step": 93
},
{
"epoch": 0.28,
"grad_norm": 0.1652740240097046,
"learning_rate": 0.00016898676344556118,
"loss": 0.5272,
"step": 94
},
{
"epoch": 0.28,
"grad_norm": 0.15505805611610413,
"learning_rate": 0.00016828592968072678,
"loss": 0.5429,
"step": 95
},
{
"epoch": 0.29,
"grad_norm": 0.16118398308753967,
"learning_rate": 0.00016757875440984768,
"loss": 0.4782,
"step": 96
},
{
"epoch": 0.29,
"grad_norm": 0.1549789011478424,
"learning_rate": 0.0001668653033061347,
"loss": 0.4594,
"step": 97
},
{
"epoch": 0.29,
"grad_norm": 0.1643325537443161,
"learning_rate": 0.00016614564262561608,
"loss": 0.5086,
"step": 98
},
{
"epoch": 0.29,
"grad_norm": 0.1949179321527481,
"learning_rate": 0.0001654198392009846,
"loss": 0.4052,
"step": 99
},
{
"epoch": 0.3,
"grad_norm": 0.1303270012140274,
"learning_rate": 0.0001646879604353908,
"loss": 0.4845,
"step": 100
},
{
"epoch": 0.3,
"grad_norm": 0.13640670478343964,
"learning_rate": 0.00016395007429618382,
"loss": 0.4459,
"step": 101
},
{
"epoch": 0.3,
"grad_norm": 0.18252772092819214,
"learning_rate": 0.00016320624930859904,
"loss": 0.523,
"step": 102
},
{
"epoch": 0.31,
"grad_norm": 0.16342373192310333,
"learning_rate": 0.00016245655454939474,
"loss": 0.4714,
"step": 103
},
{
"epoch": 0.31,
"grad_norm": 0.1418078988790512,
"learning_rate": 0.00016170105964043695,
"loss": 0.4841,
"step": 104
},
{
"epoch": 0.31,
"grad_norm": 0.15758009254932404,
"learning_rate": 0.0001609398347422339,
"loss": 0.3911,
"step": 105
},
{
"epoch": 0.31,
"grad_norm": 0.1476665586233139,
"learning_rate": 0.00016017295054742046,
"loss": 0.3962,
"step": 106
},
{
"epoch": 0.32,
"grad_norm": 0.14230811595916748,
"learning_rate": 0.00015940047827419303,
"loss": 0.4391,
"step": 107
},
{
"epoch": 0.32,
"grad_norm": 0.15269385278224945,
"learning_rate": 0.00015862248965969604,
"loss": 0.4992,
"step": 108
},
{
"epoch": 0.32,
"grad_norm": 0.14437031745910645,
"learning_rate": 0.00015783905695335946,
"loss": 0.4768,
"step": 109
},
{
"epoch": 0.33,
"grad_norm": 0.1448802649974823,
"learning_rate": 0.0001570502529101896,
"loss": 0.4469,
"step": 110
},
{
"epoch": 0.33,
"grad_norm": 0.1571648269891739,
"learning_rate": 0.00015625615078401244,
"loss": 0.4048,
"step": 111
},
{
"epoch": 0.33,
"grad_norm": 0.1563793420791626,
"learning_rate": 0.00015545682432067067,
"loss": 0.4518,
"step": 112
},
{
"epoch": 0.34,
"grad_norm": 0.1441984921693802,
"learning_rate": 0.0001546523477511754,
"loss": 0.497,
"step": 113
},
{
"epoch": 0.34,
"grad_norm": 0.13513916730880737,
"learning_rate": 0.00015384279578481221,
"loss": 0.4923,
"step": 114
},
{
"epoch": 0.34,
"grad_norm": 0.1501740962266922,
"learning_rate": 0.00015302824360220353,
"loss": 0.4701,
"step": 115
},
{
"epoch": 0.34,
"grad_norm": 0.2020605355501175,
"learning_rate": 0.00015220876684832638,
"loss": 0.4122,
"step": 116
},
{
"epoch": 0.35,
"grad_norm": 0.15191146731376648,
"learning_rate": 0.0001513844416254879,
"loss": 0.4432,
"step": 117
},
{
"epoch": 0.35,
"grad_norm": 0.1627548784017563,
"learning_rate": 0.00015055534448625766,
"loss": 0.4779,
"step": 118
},
{
"epoch": 0.35,
"grad_norm": 0.14043453335762024,
"learning_rate": 0.00014972155242635852,
"loss": 0.471,
"step": 119
},
{
"epoch": 0.36,
"grad_norm": 0.16092036664485931,
"learning_rate": 0.0001488831428775164,
"loss": 0.4543,
"step": 120
},
{
"epoch": 0.36,
"grad_norm": 0.1478460729122162,
"learning_rate": 0.00014804019370026926,
"loss": 0.4521,
"step": 121
},
{
"epoch": 0.36,
"grad_norm": 0.15067800879478455,
"learning_rate": 0.00014719278317673655,
"loss": 0.5665,
"step": 122
},
{
"epoch": 0.37,
"grad_norm": 0.16037791967391968,
"learning_rate": 0.0001463409900033493,
"loss": 0.5198,
"step": 123
},
{
"epoch": 0.37,
"grad_norm": 0.15850840508937836,
"learning_rate": 0.00014548489328354195,
"loss": 0.5183,
"step": 124
},
{
"epoch": 0.37,
"grad_norm": 0.13782666623592377,
"learning_rate": 0.00014462457252040607,
"loss": 0.4415,
"step": 125
},
{
"epoch": 0.37,
"grad_norm": 0.17566777765750885,
"learning_rate": 0.00014376010760930728,
"loss": 0.5166,
"step": 126
},
{
"epoch": 0.38,
"grad_norm": 0.14671465754508972,
"learning_rate": 0.00014289157883046568,
"loss": 0.454,
"step": 127
},
{
"epoch": 0.38,
"grad_norm": 0.14827768504619598,
"learning_rate": 0.0001420190668415002,
"loss": 0.3743,
"step": 128
},
{
"epoch": 0.38,
"grad_norm": 0.16858412325382233,
"learning_rate": 0.00014114265266993846,
"loss": 0.5204,
"step": 129
},
{
"epoch": 0.39,
"grad_norm": 0.14838969707489014,
"learning_rate": 0.00014026241770569197,
"loss": 0.5072,
"step": 130
},
{
"epoch": 0.39,
"grad_norm": 0.1676333248615265,
"learning_rate": 0.00013937844369349734,
"loss": 0.5026,
"step": 131
},
{
"epoch": 0.39,
"grad_norm": 0.1548876315355301,
"learning_rate": 0.00013849081272532544,
"loss": 0.4393,
"step": 132
},
{
"epoch": 0.4,
"grad_norm": 0.16387146711349487,
"learning_rate": 0.00013759960723275732,
"loss": 0.4899,
"step": 133
},
{
"epoch": 0.4,
"grad_norm": 0.15013404190540314,
"learning_rate": 0.00013670490997932922,
"loss": 0.4398,
"step": 134
},
{
"epoch": 0.4,
"grad_norm": 0.1466226577758789,
"learning_rate": 0.00013580680405284664,
"loss": 0.5087,
"step": 135
},
{
"epoch": 0.4,
"grad_norm": 0.15578770637512207,
"learning_rate": 0.00013490537285766808,
"loss": 0.4662,
"step": 136
},
{
"epoch": 0.41,
"grad_norm": 0.16825653612613678,
"learning_rate": 0.00013400070010695966,
"loss": 0.4826,
"step": 137
},
{
"epoch": 0.41,
"grad_norm": 0.15624648332595825,
"learning_rate": 0.00013309286981492085,
"loss": 0.4818,
"step": 138
},
{
"epoch": 0.41,
"grad_norm": 0.16185308992862701,
"learning_rate": 0.00013218196628898233,
"loss": 0.4848,
"step": 139
},
{
"epoch": 0.42,
"grad_norm": 0.15799778699874878,
"learning_rate": 0.00013126807412197665,
"loss": 0.4456,
"step": 140
},
{
"epoch": 0.42,
"grad_norm": 0.15358638763427734,
"learning_rate": 0.0001303512781842824,
"loss": 0.434,
"step": 141
},
{
"epoch": 0.42,
"grad_norm": 0.14526773989200592,
"learning_rate": 0.00012943166361594242,
"loss": 0.4944,
"step": 142
},
{
"epoch": 0.42,
"grad_norm": 0.1717992126941681,
"learning_rate": 0.00012850931581875723,
"loss": 0.4966,
"step": 143
},
{
"epoch": 0.43,
"grad_norm": 0.14694097638130188,
"learning_rate": 0.00012758432044835392,
"loss": 0.4055,
"step": 144
},
{
"epoch": 0.43,
"grad_norm": 0.17159771919250488,
"learning_rate": 0.0001266567634062317,
"loss": 0.4455,
"step": 145
},
{
"epoch": 0.43,
"grad_norm": 0.16740797460079193,
"learning_rate": 0.0001257267308317845,
"loss": 0.3732,
"step": 146
},
{
"epoch": 0.44,
"grad_norm": 0.14464697241783142,
"learning_rate": 0.00012479430909430108,
"loss": 0.4976,
"step": 147
},
{
"epoch": 0.44,
"grad_norm": 0.15393030643463135,
"learning_rate": 0.00012385958478494487,
"loss": 0.3653,
"step": 148
},
{
"epoch": 0.44,
"grad_norm": 0.15888682007789612,
"learning_rate": 0.00012292264470871182,
"loss": 0.4966,
"step": 149
},
{
"epoch": 0.45,
"grad_norm": 0.14429794251918793,
"learning_rate": 0.00012198357587636957,
"loss": 0.4821,
"step": 150
},
{
"epoch": 0.45,
"grad_norm": 0.15093272924423218,
"learning_rate": 0.00012104246549637683,
"loss": 0.4431,
"step": 151
},
{
"epoch": 0.45,
"grad_norm": 0.16575871407985687,
"learning_rate": 0.00012009940096678452,
"loss": 0.4903,
"step": 152
},
{
"epoch": 0.45,
"grad_norm": 0.15208427608013153,
"learning_rate": 0.00011915446986711953,
"loss": 0.4465,
"step": 153
},
{
"epoch": 0.46,
"grad_norm": 0.16902336478233337,
"learning_rate": 0.00011820775995025147,
"loss": 0.4997,
"step": 154
},
{
"epoch": 0.46,
"grad_norm": 0.1612280309200287,
"learning_rate": 0.0001172593591342432,
"loss": 0.4258,
"step": 155
},
{
"epoch": 0.46,
"grad_norm": 0.14114782214164734,
"learning_rate": 0.00011630935549418627,
"loss": 0.5262,
"step": 156
},
{
"epoch": 0.47,
"grad_norm": 0.16275247931480408,
"learning_rate": 0.00011535783725402163,
"loss": 0.4403,
"step": 157
},
{
"epoch": 0.47,
"grad_norm": 0.1442529857158661,
"learning_rate": 0.00011440489277834645,
"loss": 0.479,
"step": 158
},
{
"epoch": 0.47,
"grad_norm": 0.15231800079345703,
"learning_rate": 0.0001134506105642081,
"loss": 0.4856,
"step": 159
},
{
"epoch": 0.48,
"grad_norm": 0.14390550553798676,
"learning_rate": 0.00011249507923288562,
"loss": 0.3665,
"step": 160
},
{
"epoch": 0.48,
"grad_norm": 0.19076046347618103,
"learning_rate": 0.0001115383875216598,
"loss": 0.4423,
"step": 161
},
{
"epoch": 0.48,
"grad_norm": 0.16523972153663635,
"learning_rate": 0.00011058062427557229,
"loss": 0.5816,
"step": 162
},
{
"epoch": 0.48,
"grad_norm": 0.1553642749786377,
"learning_rate": 0.00010962187843917497,
"loss": 0.4793,
"step": 163
},
{
"epoch": 0.49,
"grad_norm": 0.14419597387313843,
"learning_rate": 0.0001086622390482699,
"loss": 0.4862,
"step": 164
},
{
"epoch": 0.49,
"grad_norm": 0.14601647853851318,
"learning_rate": 0.00010770179522164079,
"loss": 0.4589,
"step": 165
},
{
"epoch": 0.49,
"grad_norm": 0.15500588715076447,
"learning_rate": 0.0001067406361527768,
"loss": 0.4754,
"step": 166
},
{
"epoch": 0.5,
"grad_norm": 0.16032655537128448,
"learning_rate": 0.00010577885110158958,
"loss": 0.4644,
"step": 167
},
{
"epoch": 0.5,
"grad_norm": 0.15342958271503448,
"learning_rate": 0.00010481652938612374,
"loss": 0.3911,
"step": 168
},
{
"epoch": 0.5,
"eval_loss": 0.9589568972587585,
"eval_runtime": 324.8473,
"eval_samples_per_second": 0.767,
"eval_steps_per_second": 0.767,
"step": 168
},
{
"epoch": 0.5,
"grad_norm": 0.15607774257659912,
"learning_rate": 0.00010385376037426226,
"loss": 0.492,
"step": 169
},
{
"epoch": 0.51,
"grad_norm": 0.15384891629219055,
"learning_rate": 0.00010289063347542726,
"loss": 0.487,
"step": 170
},
{
"epoch": 0.51,
"grad_norm": 0.15275317430496216,
"learning_rate": 0.00010192723813227672,
"loss": 0.4475,
"step": 171
},
{
"epoch": 0.51,
"grad_norm": 0.1687576025724411,
"learning_rate": 0.00010096366381239808,
"loss": 0.4491,
"step": 172
},
{
"epoch": 0.51,
"grad_norm": 0.17867827415466309,
"learning_rate": 0.0001,
"loss": 0.4197,
"step": 173
},
{
"epoch": 0.52,
"grad_norm": 0.14261119067668915,
"learning_rate": 9.903633618760195e-05,
"loss": 0.4734,
"step": 174
},
{
"epoch": 0.52,
"grad_norm": 0.14787495136260986,
"learning_rate": 9.807276186772333e-05,
"loss": 0.404,
"step": 175
},
{
"epoch": 0.52,
"grad_norm": 0.15715274214744568,
"learning_rate": 9.710936652457276e-05,
"loss": 0.4583,
"step": 176
},
{
"epoch": 0.53,
"grad_norm": 0.1578068882226944,
"learning_rate": 9.614623962573776e-05,
"loss": 0.4343,
"step": 177
},
{
"epoch": 0.53,
"grad_norm": 0.15597595274448395,
"learning_rate": 9.518347061387628e-05,
"loss": 0.4454,
"step": 178
},
{
"epoch": 0.53,
"grad_norm": 0.142894446849823,
"learning_rate": 9.422114889841044e-05,
"loss": 0.4443,
"step": 179
},
{
"epoch": 0.53,
"grad_norm": 0.15279816091060638,
"learning_rate": 9.325936384722321e-05,
"loss": 0.4428,
"step": 180
},
{
"epoch": 0.54,
"grad_norm": 0.1651957929134369,
"learning_rate": 9.229820477835927e-05,
"loss": 0.3857,
"step": 181
},
{
"epoch": 0.54,
"grad_norm": 0.22555674612522125,
"learning_rate": 9.133776095173015e-05,
"loss": 0.4578,
"step": 182
},
{
"epoch": 0.54,
"grad_norm": 0.14163288474082947,
"learning_rate": 9.037812156082504e-05,
"loss": 0.4409,
"step": 183
},
{
"epoch": 0.55,
"grad_norm": 0.1683199107646942,
"learning_rate": 8.941937572442773e-05,
"loss": 0.4829,
"step": 184
},
{
"epoch": 0.55,
"grad_norm": 0.17766785621643066,
"learning_rate": 8.846161247834024e-05,
"loss": 0.4357,
"step": 185
},
{
"epoch": 0.55,
"grad_norm": 0.15518875420093536,
"learning_rate": 8.750492076711439e-05,
"loss": 0.4681,
"step": 186
},
{
"epoch": 0.56,
"grad_norm": 0.13920673727989197,
"learning_rate": 8.654938943579194e-05,
"loss": 0.4258,
"step": 187
},
{
"epoch": 0.56,
"grad_norm": 0.17552931606769562,
"learning_rate": 8.55951072216536e-05,
"loss": 0.3932,
"step": 188
},
{
"epoch": 0.56,
"grad_norm": 0.15486915409564972,
"learning_rate": 8.464216274597838e-05,
"loss": 0.4679,
"step": 189
},
{
"epoch": 0.56,
"grad_norm": 0.1437031477689743,
"learning_rate": 8.369064450581373e-05,
"loss": 0.3984,
"step": 190
},
{
"epoch": 0.57,
"grad_norm": 0.15499216318130493,
"learning_rate": 8.274064086575681e-05,
"loss": 0.5216,
"step": 191
},
{
"epoch": 0.57,
"grad_norm": 0.18269003927707672,
"learning_rate": 8.179224004974857e-05,
"loss": 0.4853,
"step": 192
},
{
"epoch": 0.57,
"grad_norm": 0.1596907526254654,
"learning_rate": 8.084553013288048e-05,
"loss": 0.4784,
"step": 193
},
{
"epoch": 0.58,
"grad_norm": 0.15115347504615784,
"learning_rate": 7.990059903321553e-05,
"loss": 0.4306,
"step": 194
},
{
"epoch": 0.58,
"grad_norm": 0.19720065593719482,
"learning_rate": 7.89575345036232e-05,
"loss": 0.4782,
"step": 195
},
{
"epoch": 0.58,
"grad_norm": 0.14170032739639282,
"learning_rate": 7.801642412363041e-05,
"loss": 0.4318,
"step": 196
},
{
"epoch": 0.59,
"grad_norm": 0.16026908159255981,
"learning_rate": 7.707735529128819e-05,
"loss": 0.4656,
"step": 197
},
{
"epoch": 0.59,
"grad_norm": 0.16910114884376526,
"learning_rate": 7.614041521505517e-05,
"loss": 0.5014,
"step": 198
},
{
"epoch": 0.59,
"grad_norm": 0.18795110285282135,
"learning_rate": 7.520569090569893e-05,
"loss": 0.4512,
"step": 199
},
{
"epoch": 0.59,
"grad_norm": 0.15903539955615997,
"learning_rate": 7.427326916821557e-05,
"loss": 0.5036,
"step": 200
},
{
"epoch": 0.6,
"grad_norm": 0.15247230231761932,
"learning_rate": 7.334323659376829e-05,
"loss": 0.4369,
"step": 201
},
{
"epoch": 0.6,
"grad_norm": 0.1472063809633255,
"learning_rate": 7.24156795516461e-05,
"loss": 0.426,
"step": 202
},
{
"epoch": 0.6,
"grad_norm": 0.15710321068763733,
"learning_rate": 7.149068418124281e-05,
"loss": 0.4034,
"step": 203
},
{
"epoch": 0.61,
"grad_norm": 0.15857233107089996,
"learning_rate": 7.056833638405762e-05,
"loss": 0.4783,
"step": 204
},
{
"epoch": 0.61,
"grad_norm": 0.16299530863761902,
"learning_rate": 6.964872181571764e-05,
"loss": 0.4719,
"step": 205
},
{
"epoch": 0.61,
"grad_norm": 0.14976787567138672,
"learning_rate": 6.87319258780234e-05,
"loss": 0.4543,
"step": 206
},
{
"epoch": 0.61,
"grad_norm": 0.14311917126178741,
"learning_rate": 6.781803371101774e-05,
"loss": 0.4422,
"step": 207
},
{
"epoch": 0.62,
"grad_norm": 0.14494828879833221,
"learning_rate": 6.690713018507918e-05,
"loss": 0.4078,
"step": 208
},
{
"epoch": 0.62,
"grad_norm": 0.1489972174167633,
"learning_rate": 6.599929989304035e-05,
"loss": 0.3908,
"step": 209
},
{
"epoch": 0.62,
"grad_norm": 0.15475499629974365,
"learning_rate": 6.509462714233195e-05,
"loss": 0.5416,
"step": 210
},
{
"epoch": 0.63,
"grad_norm": 0.1536937654018402,
"learning_rate": 6.419319594715339e-05,
"loss": 0.4267,
"step": 211
},
{
"epoch": 0.63,
"grad_norm": 0.1654786467552185,
"learning_rate": 6.32950900206708e-05,
"loss": 0.4622,
"step": 212
},
{
"epoch": 0.63,
"grad_norm": 0.14727674424648285,
"learning_rate": 6.240039276724272e-05,
"loss": 0.441,
"step": 213
},
{
"epoch": 0.64,
"grad_norm": 0.1774049550294876,
"learning_rate": 6.150918727467455e-05,
"loss": 0.4956,
"step": 214
},
{
"epoch": 0.64,
"grad_norm": 0.14867380261421204,
"learning_rate": 6.062155630650265e-05,
"loss": 0.5788,
"step": 215
},
{
"epoch": 0.64,
"grad_norm": 0.16996605694293976,
"learning_rate": 5.973758229430806e-05,
"loss": 0.588,
"step": 216
},
{
"epoch": 0.64,
"grad_norm": 0.15790195763111115,
"learning_rate": 5.885734733006154e-05,
"loss": 0.425,
"step": 217
},
{
"epoch": 0.65,
"grad_norm": 0.15656080842018127,
"learning_rate": 5.798093315849984e-05,
"loss": 0.4193,
"step": 218
},
{
"epoch": 0.65,
"grad_norm": 0.143330916762352,
"learning_rate": 5.710842116953438e-05,
"loss": 0.4427,
"step": 219
},
{
"epoch": 0.65,
"grad_norm": 0.15389469265937805,
"learning_rate": 5.623989239069275e-05,
"loss": 0.4265,
"step": 220
},
{
"epoch": 0.66,
"grad_norm": 0.16425251960754395,
"learning_rate": 5.537542747959394e-05,
"loss": 0.53,
"step": 221
},
{
"epoch": 0.66,
"grad_norm": 0.1583469659090042,
"learning_rate": 5.451510671645807e-05,
"loss": 0.5009,
"step": 222
},
{
"epoch": 0.66,
"grad_norm": 0.16075047850608826,
"learning_rate": 5.36590099966507e-05,
"loss": 0.5189,
"step": 223
},
{
"epoch": 0.67,
"grad_norm": 0.14854145050048828,
"learning_rate": 5.2807216823263484e-05,
"loss": 0.463,
"step": 224
},
{
"epoch": 0.67,
"grad_norm": 0.15285861492156982,
"learning_rate": 5.1959806299730774e-05,
"loss": 0.4413,
"step": 225
},
{
"epoch": 0.67,
"grad_norm": 0.14538119733333588,
"learning_rate": 5.111685712248364e-05,
"loss": 0.4736,
"step": 226
},
{
"epoch": 0.67,
"grad_norm": 0.15160393714904785,
"learning_rate": 5.0278447573641495e-05,
"loss": 0.4298,
"step": 227
},
{
"epoch": 0.68,
"grad_norm": 0.14497333765029907,
"learning_rate": 4.944465551374238e-05,
"loss": 0.3981,
"step": 228
},
{
"epoch": 0.68,
"grad_norm": 0.14342719316482544,
"learning_rate": 4.861555837451213e-05,
"loss": 0.4502,
"step": 229
},
{
"epoch": 0.68,
"grad_norm": 0.1621183604001999,
"learning_rate": 4.779123315167362e-05,
"loss": 0.4322,
"step": 230
},
{
"epoch": 0.69,
"grad_norm": 0.15252463519573212,
"learning_rate": 4.6971756397796504e-05,
"loss": 0.5072,
"step": 231
},
{
"epoch": 0.69,
"grad_norm": 0.17667719721794128,
"learning_rate": 4.61572042151878e-05,
"loss": 0.4708,
"step": 232
},
{
"epoch": 0.69,
"grad_norm": 0.15273234248161316,
"learning_rate": 4.5347652248824624e-05,
"loss": 0.5212,
"step": 233
},
{
"epoch": 0.7,
"grad_norm": 0.16372352838516235,
"learning_rate": 4.4543175679329344e-05,
"loss": 0.5303,
"step": 234
},
{
"epoch": 0.7,
"grad_norm": 0.14327646791934967,
"learning_rate": 4.3743849215987595e-05,
"loss": 0.4,
"step": 235
},
{
"epoch": 0.7,
"grad_norm": 0.1604086011648178,
"learning_rate": 4.294974708981041e-05,
"loss": 0.4508,
"step": 236
},
{
"epoch": 0.7,
"grad_norm": 0.23445452749729156,
"learning_rate": 4.216094304664056e-05,
"loss": 0.4363,
"step": 237
},
{
"epoch": 0.71,
"grad_norm": 0.15834979712963104,
"learning_rate": 4.137751034030399e-05,
"loss": 0.5153,
"step": 238
},
{
"epoch": 0.71,
"grad_norm": 0.14937692880630493,
"learning_rate": 4.059952172580694e-05,
"loss": 0.4423,
"step": 239
},
{
"epoch": 0.71,
"grad_norm": 0.16776776313781738,
"learning_rate": 3.982704945257957e-05,
"loss": 0.4089,
"step": 240
},
{
"epoch": 0.72,
"grad_norm": 0.16948916018009186,
"learning_rate": 3.906016525776611e-05,
"loss": 0.562,
"step": 241
},
{
"epoch": 0.72,
"grad_norm": 0.16811306774616241,
"learning_rate": 3.829894035956306e-05,
"loss": 0.407,
"step": 242
},
{
"epoch": 0.72,
"grad_norm": 0.13980595767498016,
"learning_rate": 3.7543445450605285e-05,
"loss": 0.4329,
"step": 243
},
{
"epoch": 0.72,
"grad_norm": 0.16961558163166046,
"learning_rate": 3.6793750691400994e-05,
"loss": 0.4476,
"step": 244
},
{
"epoch": 0.73,
"grad_norm": 0.16474933922290802,
"learning_rate": 3.6049925703816214e-05,
"loss": 0.4629,
"step": 245
},
{
"epoch": 0.73,
"grad_norm": 0.1619129627943039,
"learning_rate": 3.53120395646092e-05,
"loss": 0.4648,
"step": 246
},
{
"epoch": 0.73,
"grad_norm": 0.20941300690174103,
"learning_rate": 3.458016079901544e-05,
"loss": 0.4777,
"step": 247
},
{
"epoch": 0.74,
"grad_norm": 0.1691591888666153,
"learning_rate": 3.38543573743839e-05,
"loss": 0.4544,
"step": 248
},
{
"epoch": 0.74,
"grad_norm": 0.1482505053281784,
"learning_rate": 3.3134696693865316e-05,
"loss": 0.4295,
"step": 249
},
{
"epoch": 0.74,
"grad_norm": 0.16560567915439606,
"learning_rate": 3.242124559015234e-05,
"loss": 0.4568,
"step": 250
},
{
"epoch": 0.75,
"grad_norm": 0.15260429680347443,
"learning_rate": 3.171407031927325e-05,
"loss": 0.4311,
"step": 251
},
{
"epoch": 0.75,
"grad_norm": 0.15556661784648895,
"learning_rate": 3.101323655443882e-05,
"loss": 0.4655,
"step": 252
},
{
"epoch": 0.75,
"eval_loss": 0.9627882838249207,
"eval_runtime": 320.8688,
"eval_samples_per_second": 0.776,
"eval_steps_per_second": 0.776,
"step": 252
},
{
"epoch": 0.75,
"grad_norm": 0.16530689597129822,
"learning_rate": 3.031880937994359e-05,
"loss": 0.4532,
"step": 253
},
{
"epoch": 0.75,
"grad_norm": 0.1501285582780838,
"learning_rate": 2.9630853285121508e-05,
"loss": 0.4538,
"step": 254
},
{
"epoch": 0.76,
"grad_norm": 0.16318055987358093,
"learning_rate": 2.894943215835708e-05,
"loss": 0.4258,
"step": 255
},
{
"epoch": 0.76,
"grad_norm": 0.15718339383602142,
"learning_rate": 2.827460928115232e-05,
"loss": 0.4724,
"step": 256
},
{
"epoch": 0.76,
"grad_norm": 0.18563856184482574,
"learning_rate": 2.7606447322249872e-05,
"loss": 0.4064,
"step": 257
},
{
"epoch": 0.77,
"grad_norm": 0.15688276290893555,
"learning_rate": 2.6945008331813226e-05,
"loss": 0.4399,
"step": 258
},
{
"epoch": 0.77,
"grad_norm": 0.15704520046710968,
"learning_rate": 2.629035373566433e-05,
"loss": 0.4367,
"step": 259
},
{
"epoch": 0.77,
"grad_norm": 0.1557130515575409,
"learning_rate": 2.5642544329579088e-05,
"loss": 0.4937,
"step": 260
},
{
"epoch": 0.78,
"grad_norm": 0.16318149864673615,
"learning_rate": 2.500164027364147e-05,
"loss": 0.4891,
"step": 261
},
{
"epoch": 0.78,
"grad_norm": 0.16402341425418854,
"learning_rate": 2.4367701086656624e-05,
"loss": 0.5084,
"step": 262
},
{
"epoch": 0.78,
"grad_norm": 0.15349122881889343,
"learning_rate": 2.3740785640623643e-05,
"loss": 0.465,
"step": 263
},
{
"epoch": 0.78,
"grad_norm": 0.16672901809215546,
"learning_rate": 2.312095215526814e-05,
"loss": 0.4428,
"step": 264
},
{
"epoch": 0.79,
"grad_norm": 0.14176689088344574,
"learning_rate": 2.2508258192635612e-05,
"loss": 0.4012,
"step": 265
},
{
"epoch": 0.79,
"grad_norm": 0.15022584795951843,
"learning_rate": 2.1902760651745958e-05,
"loss": 0.413,
"step": 266
},
{
"epoch": 0.79,
"grad_norm": 0.15624596178531647,
"learning_rate": 2.1304515763309253e-05,
"loss": 0.4438,
"step": 267
},
{
"epoch": 0.8,
"grad_norm": 0.15674558281898499,
"learning_rate": 2.0713579084503876e-05,
"loss": 0.3967,
"step": 268
},
{
"epoch": 0.8,
"grad_norm": 0.16516539454460144,
"learning_rate": 2.013000549381706e-05,
"loss": 0.5009,
"step": 269
},
{
"epoch": 0.8,
"grad_norm": 0.15163196623325348,
"learning_rate": 1.9553849185948512e-05,
"loss": 0.4626,
"step": 270
},
{
"epoch": 0.81,
"grad_norm": 0.15535032749176025,
"learning_rate": 1.8985163666777473e-05,
"loss": 0.5294,
"step": 271
},
{
"epoch": 0.81,
"grad_norm": 0.1474699229001999,
"learning_rate": 1.8424001748393905e-05,
"loss": 0.4898,
"step": 272
},
{
"epoch": 0.81,
"grad_norm": 0.14586691558361053,
"learning_rate": 1.787041554419381e-05,
"loss": 0.3786,
"step": 273
},
{
"epoch": 0.81,
"grad_norm": 0.14684820175170898,
"learning_rate": 1.7324456464039752e-05,
"loss": 0.3862,
"step": 274
},
{
"epoch": 0.82,
"grad_norm": 0.15825559198856354,
"learning_rate": 1.6786175209486566e-05,
"loss": 0.4631,
"step": 275
},
{
"epoch": 0.82,
"grad_norm": 0.1755644828081131,
"learning_rate": 1.6255621769072805e-05,
"loss": 0.3825,
"step": 276
},
{
"epoch": 0.82,
"grad_norm": 0.15722620487213135,
"learning_rate": 1.5732845413678477e-05,
"loss": 0.4219,
"step": 277
},
{
"epoch": 0.83,
"grad_norm": 0.1658666431903839,
"learning_rate": 1.521789469194952e-05,
"loss": 0.45,
"step": 278
},
{
"epoch": 0.83,
"grad_norm": 0.13978326320648193,
"learning_rate": 1.4710817425789014e-05,
"loss": 0.4063,
"step": 279
},
{
"epoch": 0.83,
"grad_norm": 0.17951220273971558,
"learning_rate": 1.4211660705916285e-05,
"loss": 0.4697,
"step": 280
},
{
"epoch": 0.83,
"grad_norm": 0.1507434844970703,
"learning_rate": 1.3720470887493719e-05,
"loss": 0.4618,
"step": 281
},
{
"epoch": 0.84,
"grad_norm": 0.1491430401802063,
"learning_rate": 1.3237293585821786e-05,
"loss": 0.4294,
"step": 282
},
{
"epoch": 0.84,
"grad_norm": 0.15942510962486267,
"learning_rate": 1.2762173672102996e-05,
"loss": 0.4503,
"step": 283
},
{
"epoch": 0.84,
"grad_norm": 0.16852998733520508,
"learning_rate": 1.2295155269274827e-05,
"loss": 0.5253,
"step": 284
},
{
"epoch": 0.85,
"grad_norm": 0.15177421271800995,
"learning_rate": 1.1836281747912125e-05,
"loss": 0.4395,
"step": 285
},
{
"epoch": 0.85,
"grad_norm": 0.14927661418914795,
"learning_rate": 1.1385595722199438e-05,
"loss": 0.463,
"step": 286
},
{
"epoch": 0.85,
"grad_norm": 0.186951145529747,
"learning_rate": 1.0943139045973549e-05,
"loss": 0.4683,
"step": 287
},
{
"epoch": 0.86,
"grad_norm": 0.16321802139282227,
"learning_rate": 1.050895280883668e-05,
"loss": 0.478,
"step": 288
},
{
"epoch": 0.86,
"grad_norm": 0.15494760870933533,
"learning_rate": 1.0083077332340562e-05,
"loss": 0.4524,
"step": 289
},
{
"epoch": 0.86,
"grad_norm": 0.1643322855234146,
"learning_rate": 9.665552166241964e-06,
"loss": 0.514,
"step": 290
},
{
"epoch": 0.86,
"grad_norm": 0.1540500372648239,
"learning_rate": 9.256416084829778e-06,
"loss": 0.5143,
"step": 291
},
{
"epoch": 0.87,
"grad_norm": 0.17111855745315552,
"learning_rate": 8.855707083324183e-06,
"loss": 0.4618,
"step": 292
},
{
"epoch": 0.87,
"grad_norm": 0.1491898000240326,
"learning_rate": 8.46346237434813e-06,
"loss": 0.413,
"step": 293
},
{
"epoch": 0.87,
"grad_norm": 0.16696412861347198,
"learning_rate": 8.079718384471557e-06,
"loss": 0.4021,
"step": 294
},
{
"epoch": 0.88,
"grad_norm": 0.16858446598052979,
"learning_rate": 7.704510750828542e-06,
"loss": 0.4751,
"step": 295
},
{
"epoch": 0.88,
"grad_norm": 0.1504489928483963,
"learning_rate": 7.337874317807802e-06,
"loss": 0.398,
"step": 296
},
{
"epoch": 0.88,
"grad_norm": 0.183710977435112,
"learning_rate": 6.979843133816743e-06,
"loss": 0.4214,
"step": 297
},
{
"epoch": 0.89,
"grad_norm": 0.16924622654914856,
"learning_rate": 6.630450448119618e-06,
"loss": 0.4698,
"step": 298
},
{
"epoch": 0.89,
"grad_norm": 0.16073620319366455,
"learning_rate": 6.289728707749609e-06,
"loss": 0.509,
"step": 299
},
{
"epoch": 0.89,
"grad_norm": 0.1766895204782486,
"learning_rate": 5.957709554495683e-06,
"loss": 0.4978,
"step": 300
},
{
"epoch": 0.89,
"grad_norm": 0.17092542350292206,
"learning_rate": 5.634423821964074e-06,
"loss": 0.4194,
"step": 301
},
{
"epoch": 0.9,
"grad_norm": 0.1549624800682068,
"learning_rate": 5.319901532714877e-06,
"loss": 0.4529,
"step": 302
},
{
"epoch": 0.9,
"grad_norm": 0.16290757060050964,
"learning_rate": 5.014171895473929e-06,
"loss": 0.4134,
"step": 303
},
{
"epoch": 0.9,
"grad_norm": 0.1581151932477951,
"learning_rate": 4.717263302420283e-06,
"loss": 0.4456,
"step": 304
},
{
"epoch": 0.91,
"grad_norm": 0.16641291975975037,
"learning_rate": 4.429203326549525e-06,
"loss": 0.463,
"step": 305
},
{
"epoch": 0.91,
"grad_norm": 0.1655256152153015,
"learning_rate": 4.1500187191131466e-06,
"loss": 0.4648,
"step": 306
},
{
"epoch": 0.91,
"grad_norm": 0.16075529158115387,
"learning_rate": 3.879735407134244e-06,
"loss": 0.4838,
"step": 307
},
{
"epoch": 0.91,
"grad_norm": 0.17388100922107697,
"learning_rate": 3.6183784909997187e-06,
"loss": 0.4453,
"step": 308
},
{
"epoch": 0.92,
"grad_norm": 0.15238863229751587,
"learning_rate": 3.3659722421293783e-06,
"loss": 0.492,
"step": 309
},
{
"epoch": 0.92,
"grad_norm": 0.1615055352449417,
"learning_rate": 3.1225401007217936e-06,
"loss": 0.4442,
"step": 310
},
{
"epoch": 0.92,
"grad_norm": 0.14086954295635223,
"learning_rate": 2.8881046735775742e-06,
"loss": 0.3575,
"step": 311
},
{
"epoch": 0.93,
"grad_norm": 0.2627798914909363,
"learning_rate": 2.66268773199988e-06,
"loss": 0.4517,
"step": 312
},
{
"epoch": 0.93,
"grad_norm": 0.16790565848350525,
"learning_rate": 2.446310209772684e-06,
"loss": 0.4654,
"step": 313
},
{
"epoch": 0.93,
"grad_norm": 0.16735273599624634,
"learning_rate": 2.2389922012165944e-06,
"loss": 0.4681,
"step": 314
},
{
"epoch": 0.94,
"grad_norm": 0.14851590991020203,
"learning_rate": 2.0407529593228116e-06,
"loss": 0.4367,
"step": 315
},
{
"epoch": 0.94,
"grad_norm": 0.15848883986473083,
"learning_rate": 1.8516108939651945e-06,
"loss": 0.4876,
"step": 316
},
{
"epoch": 0.94,
"grad_norm": 0.16141897439956665,
"learning_rate": 1.6715835701905603e-06,
"loss": 0.4772,
"step": 317
},
{
"epoch": 0.94,
"grad_norm": 0.1615990400314331,
"learning_rate": 1.5006877065874336e-06,
"loss": 0.5161,
"step": 318
},
{
"epoch": 0.95,
"grad_norm": 0.14088964462280273,
"learning_rate": 1.3389391737335112e-06,
"loss": 0.4474,
"step": 319
},
{
"epoch": 0.95,
"grad_norm": 0.15512192249298096,
"learning_rate": 1.1863529927217732e-06,
"loss": 0.4444,
"step": 320
},
{
"epoch": 0.95,
"grad_norm": 0.14642339944839478,
"learning_rate": 1.0429433337655115e-06,
"loss": 0.4234,
"step": 321
},
{
"epoch": 0.96,
"grad_norm": 0.16831575334072113,
"learning_rate": 9.087235148824368e-07,
"loss": 0.4631,
"step": 322
},
{
"epoch": 0.96,
"grad_norm": 0.16404962539672852,
"learning_rate": 7.837060006577801e-07,
"loss": 0.5365,
"step": 323
},
{
"epoch": 0.96,
"grad_norm": 0.15236851572990417,
"learning_rate": 6.679024010868618e-07,
"loss": 0.4675,
"step": 324
},
{
"epoch": 0.97,
"grad_norm": 0.1572081297636032,
"learning_rate": 5.613234704967996e-07,
"loss": 0.4364,
"step": 325
},
{
"epoch": 0.97,
"grad_norm": 0.15015996992588043,
"learning_rate": 4.639791065478738e-07,
"loss": 0.4466,
"step": 326
},
{
"epoch": 0.97,
"grad_norm": 0.1579872965812683,
"learning_rate": 3.758783493142737e-07,
"loss": 0.5068,
"step": 327
},
{
"epoch": 0.97,
"grad_norm": 0.1496421992778778,
"learning_rate": 2.9702938044468e-07,
"loss": 0.3996,
"step": 328
},
{
"epoch": 0.98,
"grad_norm": 0.14979000389575958,
"learning_rate": 2.2743952240236176e-07,
"loss": 0.4681,
"step": 329
},
{
"epoch": 0.98,
"grad_norm": 0.1566384732723236,
"learning_rate": 1.6711523778520921e-07,
"loss": 0.4222,
"step": 330
},
{
"epoch": 0.98,
"grad_norm": 0.15907008945941925,
"learning_rate": 1.1606212872559141e-07,
"loss": 0.4386,
"step": 331
},
{
"epoch": 0.99,
"grad_norm": 0.1623019427061081,
"learning_rate": 7.428493637002821e-08,
"loss": 0.4347,
"step": 332
},
{
"epoch": 0.99,
"grad_norm": 0.1848601996898651,
"learning_rate": 4.178754043898669e-08,
"loss": 0.4431,
"step": 333
},
{
"epoch": 0.99,
"grad_norm": 0.1645398586988449,
"learning_rate": 1.8572958866514e-08,
"loss": 0.4556,
"step": 334
},
{
"epoch": 1.0,
"grad_norm": 0.1536179780960083,
"learning_rate": 4.643347520005836e-09,
"loss": 0.4282,
"step": 335
},
{
"epoch": 1.0,
"grad_norm": 0.15975765883922577,
"learning_rate": 0.0,
"loss": 0.4569,
"step": 336
},
{
"epoch": 1.0,
"eval_loss": 0.9572544097900391,
"eval_runtime": 320.7126,
"eval_samples_per_second": 0.776,
"eval_steps_per_second": 0.776,
"step": 336
}
],
"logging_steps": 1,
"max_steps": 336,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 84,
"total_flos": 3.406567632525066e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}