ToastyPigeon's picture
Training in progress, step 708, checkpoint
716463b verified
raw
history blame
125 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.7531914893617021,
"eval_steps": 189,
"global_step": 708,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0010638297872340426,
"grad_norm": 0.15719425678253174,
"learning_rate": 2.5e-05,
"loss": 1.9422,
"step": 1
},
{
"epoch": 0.0010638297872340426,
"eval_loss": 2.3947696685791016,
"eval_runtime": 228.0833,
"eval_samples_per_second": 0.403,
"eval_steps_per_second": 0.403,
"step": 1
},
{
"epoch": 0.002127659574468085,
"grad_norm": 0.12337572872638702,
"learning_rate": 5e-05,
"loss": 1.8454,
"step": 2
},
{
"epoch": 0.0031914893617021275,
"grad_norm": 0.13498862087726593,
"learning_rate": 7.5e-05,
"loss": 2.0205,
"step": 3
},
{
"epoch": 0.00425531914893617,
"grad_norm": 0.14641264081001282,
"learning_rate": 0.0001,
"loss": 2.2241,
"step": 4
},
{
"epoch": 0.005319148936170213,
"grad_norm": 0.221758633852005,
"learning_rate": 0.000125,
"loss": 2.1881,
"step": 5
},
{
"epoch": 0.006382978723404255,
"grad_norm": 0.16080522537231445,
"learning_rate": 0.00015,
"loss": 2.3418,
"step": 6
},
{
"epoch": 0.007446808510638298,
"grad_norm": 0.200699582695961,
"learning_rate": 0.000175,
"loss": 2.2462,
"step": 7
},
{
"epoch": 0.00851063829787234,
"grad_norm": 0.22244133055210114,
"learning_rate": 0.0002,
"loss": 2.1351,
"step": 8
},
{
"epoch": 0.009574468085106383,
"grad_norm": 0.12635783851146698,
"learning_rate": 0.00022500000000000002,
"loss": 2.1829,
"step": 9
},
{
"epoch": 0.010638297872340425,
"grad_norm": 0.23889027535915375,
"learning_rate": 0.00025,
"loss": 2.285,
"step": 10
},
{
"epoch": 0.011702127659574468,
"grad_norm": 0.16700100898742676,
"learning_rate": 0.0002499993636030197,
"loss": 2.0751,
"step": 11
},
{
"epoch": 0.01276595744680851,
"grad_norm": 0.13016349077224731,
"learning_rate": 0.00024999745441927886,
"loss": 2.1398,
"step": 12
},
{
"epoch": 0.013829787234042552,
"grad_norm": 0.22968527674674988,
"learning_rate": 0.00024999427247037733,
"loss": 2.0359,
"step": 13
},
{
"epoch": 0.014893617021276596,
"grad_norm": 0.145517960190773,
"learning_rate": 0.000249989817792315,
"loss": 1.9857,
"step": 14
},
{
"epoch": 0.015957446808510637,
"grad_norm": 0.12325005978345871,
"learning_rate": 0.0002499840904354907,
"loss": 2.1137,
"step": 15
},
{
"epoch": 0.01702127659574468,
"grad_norm": 0.10462953895330429,
"learning_rate": 0.0002499770904647022,
"loss": 2.1306,
"step": 16
},
{
"epoch": 0.018085106382978722,
"grad_norm": 0.10856030881404877,
"learning_rate": 0.0002499688179591453,
"loss": 1.9974,
"step": 17
},
{
"epoch": 0.019148936170212766,
"grad_norm": 0.12662340700626373,
"learning_rate": 0.0002499592730124128,
"loss": 2.3052,
"step": 18
},
{
"epoch": 0.02021276595744681,
"grad_norm": 0.10757914185523987,
"learning_rate": 0.0002499484557324936,
"loss": 2.1465,
"step": 19
},
{
"epoch": 0.02127659574468085,
"grad_norm": 0.1157526895403862,
"learning_rate": 0.0002499363662417714,
"loss": 2.2331,
"step": 20
},
{
"epoch": 0.022340425531914895,
"grad_norm": 0.12368866801261902,
"learning_rate": 0.00024992300467702346,
"loss": 2.1574,
"step": 21
},
{
"epoch": 0.023404255319148935,
"grad_norm": 0.10928671061992645,
"learning_rate": 0.00024990837118941874,
"loss": 2.0207,
"step": 22
},
{
"epoch": 0.02446808510638298,
"grad_norm": 0.1046021580696106,
"learning_rate": 0.00024989246594451646,
"loss": 2.1064,
"step": 23
},
{
"epoch": 0.02553191489361702,
"grad_norm": 0.09606686979532242,
"learning_rate": 0.0002498752891222643,
"loss": 1.9726,
"step": 24
},
{
"epoch": 0.026595744680851064,
"grad_norm": 0.10333864390850067,
"learning_rate": 0.00024985684091699594,
"loss": 1.9889,
"step": 25
},
{
"epoch": 0.027659574468085105,
"grad_norm": 0.09974610805511475,
"learning_rate": 0.00024983712153742936,
"loss": 2.1849,
"step": 26
},
{
"epoch": 0.02872340425531915,
"grad_norm": 0.09508953243494034,
"learning_rate": 0.00024981613120666417,
"loss": 1.8267,
"step": 27
},
{
"epoch": 0.029787234042553193,
"grad_norm": 0.11703687906265259,
"learning_rate": 0.0002497938701621792,
"loss": 2.1924,
"step": 28
},
{
"epoch": 0.030851063829787233,
"grad_norm": 0.10372091084718704,
"learning_rate": 0.0002497703386558297,
"loss": 1.8565,
"step": 29
},
{
"epoch": 0.031914893617021274,
"grad_norm": 0.10386721789836884,
"learning_rate": 0.0002497455369538447,
"loss": 2.0355,
"step": 30
},
{
"epoch": 0.03297872340425532,
"grad_norm": 0.09948979318141937,
"learning_rate": 0.0002497194653368238,
"loss": 2.088,
"step": 31
},
{
"epoch": 0.03404255319148936,
"grad_norm": 0.0912819355726242,
"learning_rate": 0.0002496921240997341,
"loss": 2.1658,
"step": 32
},
{
"epoch": 0.035106382978723406,
"grad_norm": 0.09557418525218964,
"learning_rate": 0.0002496635135519068,
"loss": 2.1933,
"step": 33
},
{
"epoch": 0.036170212765957444,
"grad_norm": 0.10222162306308746,
"learning_rate": 0.0002496336340170337,
"loss": 2.0631,
"step": 34
},
{
"epoch": 0.03723404255319149,
"grad_norm": 0.08941858261823654,
"learning_rate": 0.0002496024858331637,
"loss": 1.9997,
"step": 35
},
{
"epoch": 0.03829787234042553,
"grad_norm": 0.0999622568488121,
"learning_rate": 0.00024957006935269867,
"loss": 2.2196,
"step": 36
},
{
"epoch": 0.039361702127659576,
"grad_norm": 0.10653268545866013,
"learning_rate": 0.0002495363849423898,
"loss": 2.1727,
"step": 37
},
{
"epoch": 0.04042553191489362,
"grad_norm": 0.1327868402004242,
"learning_rate": 0.0002495014329833331,
"loss": 2.0095,
"step": 38
},
{
"epoch": 0.04148936170212766,
"grad_norm": 0.08846122026443481,
"learning_rate": 0.00024946521387096557,
"loss": 1.8497,
"step": 39
},
{
"epoch": 0.0425531914893617,
"grad_norm": 0.1001238003373146,
"learning_rate": 0.0002494277280150601,
"loss": 2.0145,
"step": 40
},
{
"epoch": 0.043617021276595745,
"grad_norm": 0.1113252341747284,
"learning_rate": 0.0002493889758397214,
"loss": 1.7905,
"step": 41
},
{
"epoch": 0.04468085106382979,
"grad_norm": 0.10579642653465271,
"learning_rate": 0.00024934895778338086,
"loss": 1.9684,
"step": 42
},
{
"epoch": 0.045744680851063826,
"grad_norm": 0.09689021110534668,
"learning_rate": 0.0002493076742987917,
"loss": 1.9877,
"step": 43
},
{
"epoch": 0.04680851063829787,
"grad_norm": 0.10904544591903687,
"learning_rate": 0.000249265125853024,
"loss": 1.9456,
"step": 44
},
{
"epoch": 0.047872340425531915,
"grad_norm": 0.08989599347114563,
"learning_rate": 0.000249221312927459,
"loss": 1.9915,
"step": 45
},
{
"epoch": 0.04893617021276596,
"grad_norm": 0.10626763105392456,
"learning_rate": 0.0002491762360177841,
"loss": 2.1049,
"step": 46
},
{
"epoch": 0.05,
"grad_norm": 0.10239193588495255,
"learning_rate": 0.00024912989563398696,
"loss": 2.1835,
"step": 47
},
{
"epoch": 0.05106382978723404,
"grad_norm": 0.09280844777822495,
"learning_rate": 0.00024908229230035,
"loss": 1.9502,
"step": 48
},
{
"epoch": 0.052127659574468084,
"grad_norm": 0.10963878780603409,
"learning_rate": 0.00024903342655544414,
"loss": 1.9337,
"step": 49
},
{
"epoch": 0.05319148936170213,
"grad_norm": 0.10434435307979584,
"learning_rate": 0.00024898329895212297,
"loss": 1.9479,
"step": 50
},
{
"epoch": 0.05425531914893617,
"grad_norm": 0.11127671599388123,
"learning_rate": 0.00024893191005751636,
"loss": 2.1103,
"step": 51
},
{
"epoch": 0.05531914893617021,
"grad_norm": 0.09825640171766281,
"learning_rate": 0.00024887926045302403,
"loss": 1.777,
"step": 52
},
{
"epoch": 0.05638297872340425,
"grad_norm": 0.10062053054571152,
"learning_rate": 0.00024882535073430915,
"loss": 2.0612,
"step": 53
},
{
"epoch": 0.0574468085106383,
"grad_norm": 0.1014753133058548,
"learning_rate": 0.00024877018151129134,
"loss": 2.015,
"step": 54
},
{
"epoch": 0.05851063829787234,
"grad_norm": 0.10736796259880066,
"learning_rate": 0.00024871375340814004,
"loss": 2.1078,
"step": 55
},
{
"epoch": 0.059574468085106386,
"grad_norm": 0.09837566316127777,
"learning_rate": 0.00024865606706326714,
"loss": 2.0138,
"step": 56
},
{
"epoch": 0.06063829787234042,
"grad_norm": 0.10129820555448532,
"learning_rate": 0.00024859712312932014,
"loss": 1.8079,
"step": 57
},
{
"epoch": 0.06170212765957447,
"grad_norm": 0.11590743064880371,
"learning_rate": 0.0002485369222731744,
"loss": 1.8639,
"step": 58
},
{
"epoch": 0.0627659574468085,
"grad_norm": 0.10071361064910889,
"learning_rate": 0.0002484754651759258,
"loss": 1.9809,
"step": 59
},
{
"epoch": 0.06382978723404255,
"grad_norm": 0.16242088377475739,
"learning_rate": 0.0002484127525328829,
"loss": 2.0969,
"step": 60
},
{
"epoch": 0.06489361702127659,
"grad_norm": 0.09952322393655777,
"learning_rate": 0.0002483487850535594,
"loss": 1.7684,
"step": 61
},
{
"epoch": 0.06595744680851064,
"grad_norm": 0.0950712263584137,
"learning_rate": 0.0002482835634616656,
"loss": 1.86,
"step": 62
},
{
"epoch": 0.06702127659574468,
"grad_norm": 0.1205645427107811,
"learning_rate": 0.00024821708849510053,
"loss": 1.9127,
"step": 63
},
{
"epoch": 0.06808510638297872,
"grad_norm": 0.10416848212480545,
"learning_rate": 0.00024814936090594383,
"loss": 2.1245,
"step": 64
},
{
"epoch": 0.06914893617021277,
"grad_norm": 0.10299328714609146,
"learning_rate": 0.00024808038146044664,
"loss": 2.2193,
"step": 65
},
{
"epoch": 0.07021276595744681,
"grad_norm": 0.10638066381216049,
"learning_rate": 0.00024801015093902333,
"loss": 1.8795,
"step": 66
},
{
"epoch": 0.07127659574468086,
"grad_norm": 0.13867318630218506,
"learning_rate": 0.0002479386701362427,
"loss": 2.0245,
"step": 67
},
{
"epoch": 0.07234042553191489,
"grad_norm": 0.09890972822904587,
"learning_rate": 0.00024786593986081884,
"loss": 2.1797,
"step": 68
},
{
"epoch": 0.07340425531914893,
"grad_norm": 0.09402583539485931,
"learning_rate": 0.00024779196093560195,
"loss": 2.0093,
"step": 69
},
{
"epoch": 0.07446808510638298,
"grad_norm": 0.10384728014469147,
"learning_rate": 0.00024771673419756926,
"loss": 1.9576,
"step": 70
},
{
"epoch": 0.07553191489361702,
"grad_norm": 0.11337074637413025,
"learning_rate": 0.0002476402604978153,
"loss": 2.1267,
"step": 71
},
{
"epoch": 0.07659574468085106,
"grad_norm": 0.12024985998868942,
"learning_rate": 0.00024756254070154235,
"loss": 2.1193,
"step": 72
},
{
"epoch": 0.07765957446808511,
"grad_norm": 0.09909618645906448,
"learning_rate": 0.0002474835756880509,
"loss": 2.0732,
"step": 73
},
{
"epoch": 0.07872340425531915,
"grad_norm": 0.10968328267335892,
"learning_rate": 0.00024740336635072915,
"loss": 1.8176,
"step": 74
},
{
"epoch": 0.0797872340425532,
"grad_norm": 0.10667924582958221,
"learning_rate": 0.00024732191359704345,
"loss": 1.7713,
"step": 75
},
{
"epoch": 0.08085106382978724,
"grad_norm": 0.10140973329544067,
"learning_rate": 0.0002472392183485279,
"loss": 1.9644,
"step": 76
},
{
"epoch": 0.08191489361702127,
"grad_norm": 0.1087941825389862,
"learning_rate": 0.0002471552815407736,
"loss": 2.0052,
"step": 77
},
{
"epoch": 0.08297872340425531,
"grad_norm": 0.1062784343957901,
"learning_rate": 0.0002470701041234185,
"loss": 1.9551,
"step": 78
},
{
"epoch": 0.08404255319148936,
"grad_norm": 0.10445404797792435,
"learning_rate": 0.0002469836870601363,
"loss": 2.0273,
"step": 79
},
{
"epoch": 0.0851063829787234,
"grad_norm": 0.09620675444602966,
"learning_rate": 0.000246896031328626,
"loss": 1.837,
"step": 80
},
{
"epoch": 0.08617021276595745,
"grad_norm": 0.09469189494848251,
"learning_rate": 0.00024680713792060016,
"loss": 2.0745,
"step": 81
},
{
"epoch": 0.08723404255319149,
"grad_norm": 0.10994596034288406,
"learning_rate": 0.00024671700784177435,
"loss": 1.8983,
"step": 82
},
{
"epoch": 0.08829787234042553,
"grad_norm": 0.11240627616643906,
"learning_rate": 0.0002466256421118555,
"loss": 2.0791,
"step": 83
},
{
"epoch": 0.08936170212765958,
"grad_norm": 0.11063706874847412,
"learning_rate": 0.0002465330417645301,
"loss": 1.9128,
"step": 84
},
{
"epoch": 0.09042553191489362,
"grad_norm": 0.09807933121919632,
"learning_rate": 0.0002464392078474531,
"loss": 1.9822,
"step": 85
},
{
"epoch": 0.09148936170212765,
"grad_norm": 0.11714861541986465,
"learning_rate": 0.0002463441414222355,
"loss": 2.149,
"step": 86
},
{
"epoch": 0.0925531914893617,
"grad_norm": 0.11007135361433029,
"learning_rate": 0.0002462478435644325,
"loss": 1.7036,
"step": 87
},
{
"epoch": 0.09361702127659574,
"grad_norm": 0.1519717276096344,
"learning_rate": 0.0002461503153635316,
"loss": 2.1358,
"step": 88
},
{
"epoch": 0.09468085106382979,
"grad_norm": 0.10105334222316742,
"learning_rate": 0.00024605155792293995,
"loss": 1.8889,
"step": 89
},
{
"epoch": 0.09574468085106383,
"grad_norm": 0.09844442456960678,
"learning_rate": 0.000245951572359972,
"loss": 2.0052,
"step": 90
},
{
"epoch": 0.09680851063829787,
"grad_norm": 0.1045021116733551,
"learning_rate": 0.0002458503598058367,
"loss": 2.2272,
"step": 91
},
{
"epoch": 0.09787234042553192,
"grad_norm": 0.09748804569244385,
"learning_rate": 0.0002457479214056251,
"loss": 2.2324,
"step": 92
},
{
"epoch": 0.09893617021276596,
"grad_norm": 0.09661906957626343,
"learning_rate": 0.0002456442583182969,
"loss": 2.0503,
"step": 93
},
{
"epoch": 0.1,
"grad_norm": 0.10234526544809341,
"learning_rate": 0.0002455393717166678,
"loss": 2.1088,
"step": 94
},
{
"epoch": 0.10106382978723404,
"grad_norm": 0.10246799141168594,
"learning_rate": 0.00024543326278739566,
"loss": 2.2405,
"step": 95
},
{
"epoch": 0.10212765957446808,
"grad_norm": 0.1000024825334549,
"learning_rate": 0.0002453259327309679,
"loss": 1.86,
"step": 96
},
{
"epoch": 0.10319148936170212,
"grad_norm": 0.10678999125957489,
"learning_rate": 0.0002452173827616869,
"loss": 2.134,
"step": 97
},
{
"epoch": 0.10425531914893617,
"grad_norm": 0.10660276561975479,
"learning_rate": 0.00024510761410765725,
"loss": 1.8102,
"step": 98
},
{
"epoch": 0.10531914893617021,
"grad_norm": 0.10170196741819382,
"learning_rate": 0.00024499662801077114,
"loss": 1.9482,
"step": 99
},
{
"epoch": 0.10638297872340426,
"grad_norm": 0.11082155257463455,
"learning_rate": 0.0002448844257266947,
"loss": 2.2016,
"step": 100
},
{
"epoch": 0.1074468085106383,
"grad_norm": 0.09914924204349518,
"learning_rate": 0.0002447710085248535,
"loss": 1.7777,
"step": 101
},
{
"epoch": 0.10851063829787234,
"grad_norm": 0.0965084433555603,
"learning_rate": 0.00024465637768841857,
"loss": 2.0662,
"step": 102
},
{
"epoch": 0.10957446808510639,
"grad_norm": 0.10745298117399216,
"learning_rate": 0.0002445405345142915,
"loss": 1.8013,
"step": 103
},
{
"epoch": 0.11063829787234042,
"grad_norm": 0.11509711295366287,
"learning_rate": 0.0002444234803130901,
"loss": 1.9372,
"step": 104
},
{
"epoch": 0.11170212765957446,
"grad_norm": 0.09749408811330795,
"learning_rate": 0.00024430521640913323,
"loss": 2.1201,
"step": 105
},
{
"epoch": 0.1127659574468085,
"grad_norm": 0.10483244061470032,
"learning_rate": 0.0002441857441404261,
"loss": 1.9827,
"step": 106
},
{
"epoch": 0.11382978723404255,
"grad_norm": 0.11765095591545105,
"learning_rate": 0.00024406506485864497,
"loss": 1.8852,
"step": 107
},
{
"epoch": 0.1148936170212766,
"grad_norm": 0.09942599385976791,
"learning_rate": 0.00024394317992912194,
"loss": 1.8814,
"step": 108
},
{
"epoch": 0.11595744680851064,
"grad_norm": 0.1128399670124054,
"learning_rate": 0.0002438200907308295,
"loss": 1.9887,
"step": 109
},
{
"epoch": 0.11702127659574468,
"grad_norm": 0.1034490242600441,
"learning_rate": 0.00024369579865636485,
"loss": 2.087,
"step": 110
},
{
"epoch": 0.11808510638297873,
"grad_norm": 0.10408236086368561,
"learning_rate": 0.00024357030511193423,
"loss": 1.4333,
"step": 111
},
{
"epoch": 0.11914893617021277,
"grad_norm": 0.10553466528654099,
"learning_rate": 0.00024344361151733702,
"loss": 2.1482,
"step": 112
},
{
"epoch": 0.1202127659574468,
"grad_norm": 0.10812198370695114,
"learning_rate": 0.00024331571930594954,
"loss": 1.7569,
"step": 113
},
{
"epoch": 0.12127659574468085,
"grad_norm": 0.3904220759868622,
"learning_rate": 0.00024318662992470896,
"loss": 2.0672,
"step": 114
},
{
"epoch": 0.12234042553191489,
"grad_norm": 0.1045014038681984,
"learning_rate": 0.00024305634483409703,
"loss": 1.7733,
"step": 115
},
{
"epoch": 0.12340425531914893,
"grad_norm": 0.11234873533248901,
"learning_rate": 0.00024292486550812323,
"loss": 1.8588,
"step": 116
},
{
"epoch": 0.12446808510638298,
"grad_norm": 0.12042441219091415,
"learning_rate": 0.00024279219343430842,
"loss": 2.3323,
"step": 117
},
{
"epoch": 0.125531914893617,
"grad_norm": 0.13395822048187256,
"learning_rate": 0.00024265833011366777,
"loss": 1.9681,
"step": 118
},
{
"epoch": 0.12659574468085105,
"grad_norm": 0.13613761961460114,
"learning_rate": 0.00024252327706069407,
"loss": 1.8812,
"step": 119
},
{
"epoch": 0.1276595744680851,
"grad_norm": 0.11202660948038101,
"learning_rate": 0.0002423870358033402,
"loss": 2.1017,
"step": 120
},
{
"epoch": 0.12872340425531914,
"grad_norm": 0.10813065618276596,
"learning_rate": 0.00024224960788300217,
"loss": 1.6332,
"step": 121
},
{
"epoch": 0.12978723404255318,
"grad_norm": 0.12841644883155823,
"learning_rate": 0.00024211099485450167,
"loss": 2.0098,
"step": 122
},
{
"epoch": 0.13085106382978723,
"grad_norm": 0.10981354117393494,
"learning_rate": 0.0002419711982860681,
"loss": 2.0437,
"step": 123
},
{
"epoch": 0.13191489361702127,
"grad_norm": 0.10300987213850021,
"learning_rate": 0.0002418302197593213,
"loss": 1.9898,
"step": 124
},
{
"epoch": 0.13297872340425532,
"grad_norm": 0.09915132820606232,
"learning_rate": 0.00024168806086925348,
"loss": 1.7063,
"step": 125
},
{
"epoch": 0.13404255319148936,
"grad_norm": 0.10637573152780533,
"learning_rate": 0.00024154472322421113,
"loss": 1.8626,
"step": 126
},
{
"epoch": 0.1351063829787234,
"grad_norm": 0.10706152766942978,
"learning_rate": 0.00024140020844587667,
"loss": 2.009,
"step": 127
},
{
"epoch": 0.13617021276595745,
"grad_norm": 0.10487835109233856,
"learning_rate": 0.00024125451816925058,
"loss": 2.0771,
"step": 128
},
{
"epoch": 0.1372340425531915,
"grad_norm": 0.11320215463638306,
"learning_rate": 0.0002411076540426324,
"loss": 2.111,
"step": 129
},
{
"epoch": 0.13829787234042554,
"grad_norm": 0.10687907040119171,
"learning_rate": 0.00024095961772760237,
"loss": 1.7782,
"step": 130
},
{
"epoch": 0.13936170212765958,
"grad_norm": 0.10387088358402252,
"learning_rate": 0.0002408104108990024,
"loss": 1.947,
"step": 131
},
{
"epoch": 0.14042553191489363,
"grad_norm": 0.11471838504076004,
"learning_rate": 0.00024066003524491752,
"loss": 2.0569,
"step": 132
},
{
"epoch": 0.14148936170212767,
"grad_norm": 0.10412081331014633,
"learning_rate": 0.00024050849246665628,
"loss": 1.9277,
"step": 133
},
{
"epoch": 0.1425531914893617,
"grad_norm": 0.11945554614067078,
"learning_rate": 0.00024035578427873192,
"loss": 2.0863,
"step": 134
},
{
"epoch": 0.14361702127659576,
"grad_norm": 0.12696759402751923,
"learning_rate": 0.00024020191240884276,
"loss": 2.0757,
"step": 135
},
{
"epoch": 0.14468085106382977,
"grad_norm": 0.11579248309135437,
"learning_rate": 0.00024004687859785265,
"loss": 1.9771,
"step": 136
},
{
"epoch": 0.14574468085106382,
"grad_norm": 0.1291639804840088,
"learning_rate": 0.00023989068459977139,
"loss": 1.9655,
"step": 137
},
{
"epoch": 0.14680851063829786,
"grad_norm": 0.10244561731815338,
"learning_rate": 0.0002397333321817347,
"loss": 2.0573,
"step": 138
},
{
"epoch": 0.1478723404255319,
"grad_norm": 0.10870030522346497,
"learning_rate": 0.00023957482312398452,
"loss": 1.9953,
"step": 139
},
{
"epoch": 0.14893617021276595,
"grad_norm": 0.11521364748477936,
"learning_rate": 0.00023941515921984855,
"loss": 1.7835,
"step": 140
},
{
"epoch": 0.15,
"grad_norm": 0.1189892515540123,
"learning_rate": 0.00023925434227572013,
"loss": 1.8137,
"step": 141
},
{
"epoch": 0.15106382978723404,
"grad_norm": 0.11443563550710678,
"learning_rate": 0.00023909237411103785,
"loss": 2.1109,
"step": 142
},
{
"epoch": 0.15212765957446808,
"grad_norm": 0.1882992833852768,
"learning_rate": 0.00023892925655826476,
"loss": 2.1079,
"step": 143
},
{
"epoch": 0.15319148936170213,
"grad_norm": 0.11956554651260376,
"learning_rate": 0.0002387649914628679,
"loss": 2.0648,
"step": 144
},
{
"epoch": 0.15425531914893617,
"grad_norm": 0.13715718686580658,
"learning_rate": 0.00023859958068329725,
"loss": 2.004,
"step": 145
},
{
"epoch": 0.15531914893617021,
"grad_norm": 0.1169043555855751,
"learning_rate": 0.00023843302609096467,
"loss": 2.1193,
"step": 146
},
{
"epoch": 0.15638297872340426,
"grad_norm": 0.11039727926254272,
"learning_rate": 0.00023826532957022293,
"loss": 1.9419,
"step": 147
},
{
"epoch": 0.1574468085106383,
"grad_norm": 0.11504384130239487,
"learning_rate": 0.00023809649301834415,
"loss": 1.9714,
"step": 148
},
{
"epoch": 0.15851063829787235,
"grad_norm": 0.11435811966657639,
"learning_rate": 0.00023792651834549856,
"loss": 1.7879,
"step": 149
},
{
"epoch": 0.1595744680851064,
"grad_norm": 0.27376243472099304,
"learning_rate": 0.00023775540747473262,
"loss": 2.0011,
"step": 150
},
{
"epoch": 0.16063829787234044,
"grad_norm": 0.12073350697755814,
"learning_rate": 0.0002375831623419476,
"loss": 1.9285,
"step": 151
},
{
"epoch": 0.16170212765957448,
"grad_norm": 0.11144834011793137,
"learning_rate": 0.00023740978489587733,
"loss": 1.9147,
"step": 152
},
{
"epoch": 0.16276595744680852,
"grad_norm": 0.1074347198009491,
"learning_rate": 0.00023723527709806647,
"loss": 2.092,
"step": 153
},
{
"epoch": 0.16382978723404254,
"grad_norm": 0.11253698170185089,
"learning_rate": 0.00023705964092284805,
"loss": 1.9452,
"step": 154
},
{
"epoch": 0.16489361702127658,
"grad_norm": 0.1153697595000267,
"learning_rate": 0.00023688287835732134,
"loss": 2.008,
"step": 155
},
{
"epoch": 0.16595744680851063,
"grad_norm": 0.11541475355625153,
"learning_rate": 0.00023670499140132918,
"loss": 1.5319,
"step": 156
},
{
"epoch": 0.16702127659574467,
"grad_norm": 0.11363336443901062,
"learning_rate": 0.00023652598206743553,
"loss": 2.1214,
"step": 157
},
{
"epoch": 0.16808510638297872,
"grad_norm": 0.10930821299552917,
"learning_rate": 0.00023634585238090265,
"loss": 1.8862,
"step": 158
},
{
"epoch": 0.16914893617021276,
"grad_norm": 0.13756263256072998,
"learning_rate": 0.000236164604379668,
"loss": 2.001,
"step": 159
},
{
"epoch": 0.1702127659574468,
"grad_norm": 0.11600874364376068,
"learning_rate": 0.00023598224011432163,
"loss": 2.0017,
"step": 160
},
{
"epoch": 0.17127659574468085,
"grad_norm": 0.14339403808116913,
"learning_rate": 0.00023579876164808234,
"loss": 1.8875,
"step": 161
},
{
"epoch": 0.1723404255319149,
"grad_norm": 0.11734171211719513,
"learning_rate": 0.000235614171056775,
"loss": 2.025,
"step": 162
},
{
"epoch": 0.17340425531914894,
"grad_norm": 0.16881157457828522,
"learning_rate": 0.00023542847042880662,
"loss": 2.1279,
"step": 163
},
{
"epoch": 0.17446808510638298,
"grad_norm": 0.10519936680793762,
"learning_rate": 0.00023524166186514284,
"loss": 1.8798,
"step": 164
},
{
"epoch": 0.17553191489361702,
"grad_norm": 0.11043398827314377,
"learning_rate": 0.00023505374747928419,
"loss": 1.9368,
"step": 165
},
{
"epoch": 0.17659574468085107,
"grad_norm": 0.11440404504537582,
"learning_rate": 0.00023486472939724217,
"loss": 1.9948,
"step": 166
},
{
"epoch": 0.1776595744680851,
"grad_norm": 0.11574520170688629,
"learning_rate": 0.0002346746097575153,
"loss": 2.1822,
"step": 167
},
{
"epoch": 0.17872340425531916,
"grad_norm": 0.11897675693035126,
"learning_rate": 0.00023448339071106461,
"loss": 1.9755,
"step": 168
},
{
"epoch": 0.1797872340425532,
"grad_norm": 0.116481252014637,
"learning_rate": 0.00023429107442128974,
"loss": 1.751,
"step": 169
},
{
"epoch": 0.18085106382978725,
"grad_norm": 0.11168333142995834,
"learning_rate": 0.0002340976630640041,
"loss": 2.1915,
"step": 170
},
{
"epoch": 0.1819148936170213,
"grad_norm": 0.1481730043888092,
"learning_rate": 0.00023390315882741038,
"loss": 1.9164,
"step": 171
},
{
"epoch": 0.1829787234042553,
"grad_norm": 0.10625225305557251,
"learning_rate": 0.00023370756391207588,
"loss": 1.8064,
"step": 172
},
{
"epoch": 0.18404255319148935,
"grad_norm": 0.11492953449487686,
"learning_rate": 0.0002335108805309075,
"loss": 2.0202,
"step": 173
},
{
"epoch": 0.1851063829787234,
"grad_norm": 0.11886078119277954,
"learning_rate": 0.00023331311090912677,
"loss": 1.994,
"step": 174
},
{
"epoch": 0.18617021276595744,
"grad_norm": 0.12113317847251892,
"learning_rate": 0.0002331142572842446,
"loss": 2.061,
"step": 175
},
{
"epoch": 0.18723404255319148,
"grad_norm": 0.11451803892850876,
"learning_rate": 0.00023291432190603603,
"loss": 2.0846,
"step": 176
},
{
"epoch": 0.18829787234042553,
"grad_norm": 0.13504713773727417,
"learning_rate": 0.00023271330703651482,
"loss": 2.0917,
"step": 177
},
{
"epoch": 0.18936170212765957,
"grad_norm": 0.11458391696214676,
"learning_rate": 0.00023251121494990764,
"loss": 1.704,
"step": 178
},
{
"epoch": 0.19042553191489361,
"grad_norm": 0.13398422300815582,
"learning_rate": 0.00023230804793262873,
"loss": 2.103,
"step": 179
},
{
"epoch": 0.19148936170212766,
"grad_norm": 0.11917758733034134,
"learning_rate": 0.00023210380828325353,
"loss": 1.7718,
"step": 180
},
{
"epoch": 0.1925531914893617,
"grad_norm": 0.1172708049416542,
"learning_rate": 0.0002318984983124932,
"loss": 2.0172,
"step": 181
},
{
"epoch": 0.19361702127659575,
"grad_norm": 0.1266704946756363,
"learning_rate": 0.00023169212034316805,
"loss": 1.9212,
"step": 182
},
{
"epoch": 0.1946808510638298,
"grad_norm": 0.13580508530139923,
"learning_rate": 0.0002314846767101815,
"loss": 2.0556,
"step": 183
},
{
"epoch": 0.19574468085106383,
"grad_norm": 0.1328660249710083,
"learning_rate": 0.00023127616976049357,
"loss": 1.9746,
"step": 184
},
{
"epoch": 0.19680851063829788,
"grad_norm": 0.1272975504398346,
"learning_rate": 0.00023106660185309434,
"loss": 1.9995,
"step": 185
},
{
"epoch": 0.19787234042553192,
"grad_norm": 0.11610808223485947,
"learning_rate": 0.0002308559753589773,
"loss": 1.9755,
"step": 186
},
{
"epoch": 0.19893617021276597,
"grad_norm": 0.11477465182542801,
"learning_rate": 0.0002306442926611125,
"loss": 1.8594,
"step": 187
},
{
"epoch": 0.2,
"grad_norm": 0.11912880837917328,
"learning_rate": 0.00023043155615441963,
"loss": 2.2537,
"step": 188
},
{
"epoch": 0.20106382978723406,
"grad_norm": 0.12370310723781586,
"learning_rate": 0.00023021776824574077,
"loss": 1.8427,
"step": 189
},
{
"epoch": 0.20106382978723406,
"eval_loss": 2.2439846992492676,
"eval_runtime": 228.9331,
"eval_samples_per_second": 0.402,
"eval_steps_per_second": 0.402,
"step": 189
},
{
"epoch": 0.20212765957446807,
"grad_norm": 0.12219749391078949,
"learning_rate": 0.0002300029313538133,
"loss": 1.6372,
"step": 190
},
{
"epoch": 0.20319148936170212,
"grad_norm": 0.12997935712337494,
"learning_rate": 0.00022978704790924264,
"loss": 1.7587,
"step": 191
},
{
"epoch": 0.20425531914893616,
"grad_norm": 0.1191156730055809,
"learning_rate": 0.00022957012035447445,
"loss": 2.1128,
"step": 192
},
{
"epoch": 0.2053191489361702,
"grad_norm": 0.1545037180185318,
"learning_rate": 0.00022935215114376723,
"loss": 1.6692,
"step": 193
},
{
"epoch": 0.20638297872340425,
"grad_norm": 0.1366199105978012,
"learning_rate": 0.00022913314274316455,
"loss": 1.9685,
"step": 194
},
{
"epoch": 0.2074468085106383,
"grad_norm": 0.12138072401285172,
"learning_rate": 0.00022891309763046697,
"loss": 1.8484,
"step": 195
},
{
"epoch": 0.20851063829787234,
"grad_norm": 0.1285848766565323,
"learning_rate": 0.0002286920182952042,
"loss": 1.7935,
"step": 196
},
{
"epoch": 0.20957446808510638,
"grad_norm": 0.11619207262992859,
"learning_rate": 0.00022846990723860686,
"loss": 2.0731,
"step": 197
},
{
"epoch": 0.21063829787234042,
"grad_norm": 0.4214017689228058,
"learning_rate": 0.00022824676697357814,
"loss": 2.0675,
"step": 198
},
{
"epoch": 0.21170212765957447,
"grad_norm": 0.11827094852924347,
"learning_rate": 0.0002280226000246654,
"loss": 1.7509,
"step": 199
},
{
"epoch": 0.2127659574468085,
"grad_norm": 0.132975772023201,
"learning_rate": 0.00022779740892803175,
"loss": 1.6893,
"step": 200
},
{
"epoch": 0.21382978723404256,
"grad_norm": 0.12743262946605682,
"learning_rate": 0.00022757119623142706,
"loss": 1.9338,
"step": 201
},
{
"epoch": 0.2148936170212766,
"grad_norm": 0.13391292095184326,
"learning_rate": 0.00022734396449415934,
"loss": 2.1821,
"step": 202
},
{
"epoch": 0.21595744680851064,
"grad_norm": 0.12177909165620804,
"learning_rate": 0.0002271157162870658,
"loss": 1.8998,
"step": 203
},
{
"epoch": 0.2170212765957447,
"grad_norm": 0.3537590503692627,
"learning_rate": 0.00022688645419248374,
"loss": 1.8342,
"step": 204
},
{
"epoch": 0.21808510638297873,
"grad_norm": 0.14446629583835602,
"learning_rate": 0.00022665618080422114,
"loss": 1.9379,
"step": 205
},
{
"epoch": 0.21914893617021278,
"grad_norm": 0.1220061182975769,
"learning_rate": 0.00022642489872752765,
"loss": 1.8389,
"step": 206
},
{
"epoch": 0.22021276595744682,
"grad_norm": 0.10652719438076019,
"learning_rate": 0.00022619261057906485,
"loss": 1.7434,
"step": 207
},
{
"epoch": 0.22127659574468084,
"grad_norm": 0.1321529746055603,
"learning_rate": 0.00022595931898687673,
"loss": 2.2176,
"step": 208
},
{
"epoch": 0.22234042553191488,
"grad_norm": 0.11512637138366699,
"learning_rate": 0.00022572502659036007,
"loss": 1.8772,
"step": 209
},
{
"epoch": 0.22340425531914893,
"grad_norm": 0.15195037424564362,
"learning_rate": 0.00022548973604023437,
"loss": 2.1445,
"step": 210
},
{
"epoch": 0.22446808510638297,
"grad_norm": 0.1171952560544014,
"learning_rate": 0.00022525344999851197,
"loss": 2.0841,
"step": 211
},
{
"epoch": 0.225531914893617,
"grad_norm": 0.11501558870077133,
"learning_rate": 0.00022501617113846798,
"loss": 1.8845,
"step": 212
},
{
"epoch": 0.22659574468085106,
"grad_norm": 0.14285384118556976,
"learning_rate": 0.00022477790214460994,
"loss": 1.7457,
"step": 213
},
{
"epoch": 0.2276595744680851,
"grad_norm": 0.1312599778175354,
"learning_rate": 0.0002245386457126475,
"loss": 2.12,
"step": 214
},
{
"epoch": 0.22872340425531915,
"grad_norm": 0.12865905463695526,
"learning_rate": 0.00022429840454946196,
"loss": 1.9595,
"step": 215
},
{
"epoch": 0.2297872340425532,
"grad_norm": 0.18248826265335083,
"learning_rate": 0.00022405718137307552,
"loss": 2.0199,
"step": 216
},
{
"epoch": 0.23085106382978723,
"grad_norm": 0.129413902759552,
"learning_rate": 0.00022381497891262066,
"loss": 1.8833,
"step": 217
},
{
"epoch": 0.23191489361702128,
"grad_norm": 0.12320221215486526,
"learning_rate": 0.00022357179990830923,
"loss": 2.0043,
"step": 218
},
{
"epoch": 0.23297872340425532,
"grad_norm": 0.11922425031661987,
"learning_rate": 0.00022332764711140127,
"loss": 1.9641,
"step": 219
},
{
"epoch": 0.23404255319148937,
"grad_norm": 0.1320050060749054,
"learning_rate": 0.00022308252328417426,
"loss": 1.9602,
"step": 220
},
{
"epoch": 0.2351063829787234,
"grad_norm": 0.14773212373256683,
"learning_rate": 0.00022283643119989154,
"loss": 1.8785,
"step": 221
},
{
"epoch": 0.23617021276595745,
"grad_norm": 0.128052219748497,
"learning_rate": 0.00022258937364277105,
"loss": 1.8772,
"step": 222
},
{
"epoch": 0.2372340425531915,
"grad_norm": 0.12113828212022781,
"learning_rate": 0.00022234135340795378,
"loss": 2.0577,
"step": 223
},
{
"epoch": 0.23829787234042554,
"grad_norm": 0.1478564292192459,
"learning_rate": 0.0002220923733014723,
"loss": 1.7237,
"step": 224
},
{
"epoch": 0.2393617021276596,
"grad_norm": 0.12648318707942963,
"learning_rate": 0.0002218424361402188,
"loss": 1.5301,
"step": 225
},
{
"epoch": 0.2404255319148936,
"grad_norm": 0.13919681310653687,
"learning_rate": 0.00022159154475191343,
"loss": 1.8451,
"step": 226
},
{
"epoch": 0.24148936170212765,
"grad_norm": 0.162435844540596,
"learning_rate": 0.00022133970197507199,
"loss": 1.7585,
"step": 227
},
{
"epoch": 0.2425531914893617,
"grad_norm": 0.1612282693386078,
"learning_rate": 0.0002210869106589743,
"loss": 1.9223,
"step": 228
},
{
"epoch": 0.24361702127659574,
"grad_norm": 0.1214359775185585,
"learning_rate": 0.0002208331736636316,
"loss": 1.8637,
"step": 229
},
{
"epoch": 0.24468085106382978,
"grad_norm": 0.1307225078344345,
"learning_rate": 0.0002205784938597541,
"loss": 1.9397,
"step": 230
},
{
"epoch": 0.24574468085106382,
"grad_norm": 0.13562381267547607,
"learning_rate": 0.00022032287412871893,
"loss": 1.9584,
"step": 231
},
{
"epoch": 0.24680851063829787,
"grad_norm": 0.12984585762023926,
"learning_rate": 0.00022006631736253715,
"loss": 1.8084,
"step": 232
},
{
"epoch": 0.2478723404255319,
"grad_norm": 0.11881977319717407,
"learning_rate": 0.00021980882646382128,
"loss": 1.747,
"step": 233
},
{
"epoch": 0.24893617021276596,
"grad_norm": 0.13311228156089783,
"learning_rate": 0.0002195504043457523,
"loss": 1.9873,
"step": 234
},
{
"epoch": 0.25,
"grad_norm": 0.13149425387382507,
"learning_rate": 0.00021929105393204675,
"loss": 2.0715,
"step": 235
},
{
"epoch": 0.251063829787234,
"grad_norm": 0.1460556536912918,
"learning_rate": 0.00021903077815692358,
"loss": 1.7887,
"step": 236
},
{
"epoch": 0.2521276595744681,
"grad_norm": 0.13757912814617157,
"learning_rate": 0.0002187695799650712,
"loss": 1.7325,
"step": 237
},
{
"epoch": 0.2531914893617021,
"grad_norm": 0.13054710626602173,
"learning_rate": 0.0002185074623116138,
"loss": 1.9821,
"step": 238
},
{
"epoch": 0.2542553191489362,
"grad_norm": 0.11559314280748367,
"learning_rate": 0.00021824442816207828,
"loss": 1.6488,
"step": 239
},
{
"epoch": 0.2553191489361702,
"grad_norm": 0.12026959657669067,
"learning_rate": 0.00021798048049236036,
"loss": 2.0594,
"step": 240
},
{
"epoch": 0.25638297872340426,
"grad_norm": 0.12917517125606537,
"learning_rate": 0.00021771562228869122,
"loss": 2.0606,
"step": 241
},
{
"epoch": 0.2574468085106383,
"grad_norm": 0.1309247761964798,
"learning_rate": 0.00021744985654760345,
"loss": 1.816,
"step": 242
},
{
"epoch": 0.25851063829787235,
"grad_norm": 0.11951275914907455,
"learning_rate": 0.00021718318627589732,
"loss": 1.8091,
"step": 243
},
{
"epoch": 0.25957446808510637,
"grad_norm": 0.1549317091703415,
"learning_rate": 0.00021691561449060668,
"loss": 1.7875,
"step": 244
},
{
"epoch": 0.26063829787234044,
"grad_norm": 0.10621926933526993,
"learning_rate": 0.00021664714421896494,
"loss": 1.9503,
"step": 245
},
{
"epoch": 0.26170212765957446,
"grad_norm": 0.11808984726667404,
"learning_rate": 0.00021637777849837069,
"loss": 1.7337,
"step": 246
},
{
"epoch": 0.26276595744680853,
"grad_norm": 0.12691949307918549,
"learning_rate": 0.00021610752037635333,
"loss": 1.8108,
"step": 247
},
{
"epoch": 0.26382978723404255,
"grad_norm": 0.15534377098083496,
"learning_rate": 0.0002158363729105387,
"loss": 1.8423,
"step": 248
},
{
"epoch": 0.2648936170212766,
"grad_norm": 0.12336020171642303,
"learning_rate": 0.00021556433916861446,
"loss": 1.7249,
"step": 249
},
{
"epoch": 0.26595744680851063,
"grad_norm": 0.13493220508098602,
"learning_rate": 0.0002152914222282952,
"loss": 2.0487,
"step": 250
},
{
"epoch": 0.2670212765957447,
"grad_norm": 0.19705940783023834,
"learning_rate": 0.00021501762517728802,
"loss": 1.5196,
"step": 251
},
{
"epoch": 0.2680851063829787,
"grad_norm": 0.15215103328227997,
"learning_rate": 0.00021474295111325708,
"loss": 1.8737,
"step": 252
},
{
"epoch": 0.2691489361702128,
"grad_norm": 0.13895325362682343,
"learning_rate": 0.0002144674031437891,
"loss": 1.7027,
"step": 253
},
{
"epoch": 0.2702127659574468,
"grad_norm": 0.11689046770334244,
"learning_rate": 0.00021419098438635765,
"loss": 2.0475,
"step": 254
},
{
"epoch": 0.2712765957446808,
"grad_norm": 0.1213349923491478,
"learning_rate": 0.00021391369796828835,
"loss": 1.9444,
"step": 255
},
{
"epoch": 0.2723404255319149,
"grad_norm": 0.11021054536104202,
"learning_rate": 0.00021363554702672326,
"loss": 1.9382,
"step": 256
},
{
"epoch": 0.2734042553191489,
"grad_norm": 0.1282310038805008,
"learning_rate": 0.0002133565347085853,
"loss": 1.9628,
"step": 257
},
{
"epoch": 0.274468085106383,
"grad_norm": 0.11131104081869125,
"learning_rate": 0.0002130766641705429,
"loss": 1.8109,
"step": 258
},
{
"epoch": 0.275531914893617,
"grad_norm": 0.123484767973423,
"learning_rate": 0.00021279593857897411,
"loss": 2.1388,
"step": 259
},
{
"epoch": 0.2765957446808511,
"grad_norm": 0.12884195148944855,
"learning_rate": 0.00021251436110993085,
"loss": 1.7784,
"step": 260
},
{
"epoch": 0.2776595744680851,
"grad_norm": 0.14552558958530426,
"learning_rate": 0.00021223193494910293,
"loss": 2.0514,
"step": 261
},
{
"epoch": 0.27872340425531916,
"grad_norm": 0.1309530884027481,
"learning_rate": 0.00021194866329178193,
"loss": 1.9086,
"step": 262
},
{
"epoch": 0.2797872340425532,
"grad_norm": 0.12760141491889954,
"learning_rate": 0.00021166454934282521,
"loss": 1.8776,
"step": 263
},
{
"epoch": 0.28085106382978725,
"grad_norm": 0.1298709511756897,
"learning_rate": 0.0002113795963166197,
"loss": 1.9672,
"step": 264
},
{
"epoch": 0.28191489361702127,
"grad_norm": 0.13304266333580017,
"learning_rate": 0.0002110938074370453,
"loss": 1.9939,
"step": 265
},
{
"epoch": 0.28297872340425534,
"grad_norm": 0.12197323888540268,
"learning_rate": 0.00021080718593743846,
"loss": 2.041,
"step": 266
},
{
"epoch": 0.28404255319148936,
"grad_norm": 0.11913374811410904,
"learning_rate": 0.00021051973506055587,
"loss": 1.6231,
"step": 267
},
{
"epoch": 0.2851063829787234,
"grad_norm": 0.18801186978816986,
"learning_rate": 0.00021023145805853735,
"loss": 1.9754,
"step": 268
},
{
"epoch": 0.28617021276595744,
"grad_norm": 0.14811000227928162,
"learning_rate": 0.0002099423581928694,
"loss": 2.0761,
"step": 269
},
{
"epoch": 0.2872340425531915,
"grad_norm": 0.12679357826709747,
"learning_rate": 0.00020965243873434806,
"loss": 1.8865,
"step": 270
},
{
"epoch": 0.28829787234042553,
"grad_norm": 0.1385306566953659,
"learning_rate": 0.00020936170296304222,
"loss": 2.036,
"step": 271
},
{
"epoch": 0.28936170212765955,
"grad_norm": 0.11551671475172043,
"learning_rate": 0.00020907015416825607,
"loss": 2.0402,
"step": 272
},
{
"epoch": 0.2904255319148936,
"grad_norm": 0.15645508468151093,
"learning_rate": 0.00020877779564849232,
"loss": 1.7009,
"step": 273
},
{
"epoch": 0.29148936170212764,
"grad_norm": 0.1431574672460556,
"learning_rate": 0.0002084846307114145,
"loss": 1.7321,
"step": 274
},
{
"epoch": 0.2925531914893617,
"grad_norm": 0.14062929153442383,
"learning_rate": 0.00020819066267380995,
"loss": 1.9823,
"step": 275
},
{
"epoch": 0.2936170212765957,
"grad_norm": 0.12429283559322357,
"learning_rate": 0.00020789589486155187,
"loss": 1.9957,
"step": 276
},
{
"epoch": 0.2946808510638298,
"grad_norm": 0.12170275300741196,
"learning_rate": 0.00020760033060956192,
"loss": 1.8834,
"step": 277
},
{
"epoch": 0.2957446808510638,
"grad_norm": 0.12186326086521149,
"learning_rate": 0.00020730397326177257,
"loss": 1.821,
"step": 278
},
{
"epoch": 0.2968085106382979,
"grad_norm": 0.14743085205554962,
"learning_rate": 0.00020700682617108906,
"loss": 1.9366,
"step": 279
},
{
"epoch": 0.2978723404255319,
"grad_norm": 0.1397976577281952,
"learning_rate": 0.00020670889269935156,
"loss": 1.924,
"step": 280
},
{
"epoch": 0.298936170212766,
"grad_norm": 0.16039437055587769,
"learning_rate": 0.0002064101762172972,
"loss": 1.6082,
"step": 281
},
{
"epoch": 0.3,
"grad_norm": 0.12346199154853821,
"learning_rate": 0.00020611068010452187,
"loss": 2.0276,
"step": 282
},
{
"epoch": 0.30106382978723406,
"grad_norm": 0.12658649682998657,
"learning_rate": 0.0002058104077494419,
"loss": 1.8506,
"step": 283
},
{
"epoch": 0.3021276595744681,
"grad_norm": 0.13528642058372498,
"learning_rate": 0.00020550936254925588,
"loss": 1.7422,
"step": 284
},
{
"epoch": 0.30319148936170215,
"grad_norm": 0.14657703042030334,
"learning_rate": 0.00020520754790990617,
"loss": 1.6911,
"step": 285
},
{
"epoch": 0.30425531914893617,
"grad_norm": 0.14604607224464417,
"learning_rate": 0.0002049049672460403,
"loss": 1.8764,
"step": 286
},
{
"epoch": 0.30531914893617024,
"grad_norm": 0.15669088065624237,
"learning_rate": 0.00020460162398097242,
"loss": 1.9052,
"step": 287
},
{
"epoch": 0.30638297872340425,
"grad_norm": 0.1708899289369583,
"learning_rate": 0.0002042975215466446,
"loss": 1.9045,
"step": 288
},
{
"epoch": 0.3074468085106383,
"grad_norm": 0.15268519520759583,
"learning_rate": 0.00020399266338358776,
"loss": 1.9059,
"step": 289
},
{
"epoch": 0.30851063829787234,
"grad_norm": 0.13716553151607513,
"learning_rate": 0.00020368705294088313,
"loss": 1.9051,
"step": 290
},
{
"epoch": 0.30957446808510636,
"grad_norm": 0.14579275250434875,
"learning_rate": 0.00020338069367612294,
"loss": 1.7654,
"step": 291
},
{
"epoch": 0.31063829787234043,
"grad_norm": 0.14516612887382507,
"learning_rate": 0.00020307358905537133,
"loss": 1.7659,
"step": 292
},
{
"epoch": 0.31170212765957445,
"grad_norm": 0.13099181652069092,
"learning_rate": 0.00020276574255312529,
"loss": 1.8719,
"step": 293
},
{
"epoch": 0.3127659574468085,
"grad_norm": 0.14634104073047638,
"learning_rate": 0.00020245715765227523,
"loss": 1.8947,
"step": 294
},
{
"epoch": 0.31382978723404253,
"grad_norm": 0.13733884692192078,
"learning_rate": 0.00020214783784406554,
"loss": 1.9405,
"step": 295
},
{
"epoch": 0.3148936170212766,
"grad_norm": 0.1501719057559967,
"learning_rate": 0.00020183778662805516,
"loss": 2.1347,
"step": 296
},
{
"epoch": 0.3159574468085106,
"grad_norm": 0.1275307983160019,
"learning_rate": 0.0002015270075120781,
"loss": 1.9346,
"step": 297
},
{
"epoch": 0.3170212765957447,
"grad_norm": 0.2898395359516144,
"learning_rate": 0.00020121550401220342,
"loss": 1.8929,
"step": 298
},
{
"epoch": 0.3180851063829787,
"grad_norm": 0.13291271030902863,
"learning_rate": 0.00020090327965269583,
"loss": 2.0107,
"step": 299
},
{
"epoch": 0.3191489361702128,
"grad_norm": 0.13238263130187988,
"learning_rate": 0.00020059033796597563,
"loss": 2.0485,
"step": 300
},
{
"epoch": 0.3202127659574468,
"grad_norm": 0.1353788673877716,
"learning_rate": 0.0002002766824925786,
"loss": 1.9357,
"step": 301
},
{
"epoch": 0.32127659574468087,
"grad_norm": 0.16099432110786438,
"learning_rate": 0.00019996231678111635,
"loss": 1.8787,
"step": 302
},
{
"epoch": 0.3223404255319149,
"grad_norm": 0.22791248559951782,
"learning_rate": 0.00019964724438823573,
"loss": 1.5107,
"step": 303
},
{
"epoch": 0.32340425531914896,
"grad_norm": 0.181181401014328,
"learning_rate": 0.00019933146887857891,
"loss": 1.7504,
"step": 304
},
{
"epoch": 0.324468085106383,
"grad_norm": 0.14129158854484558,
"learning_rate": 0.0001990149938247428,
"loss": 1.6271,
"step": 305
},
{
"epoch": 0.32553191489361705,
"grad_norm": 0.16222205758094788,
"learning_rate": 0.00019869782280723897,
"loss": 1.7481,
"step": 306
},
{
"epoch": 0.32659574468085106,
"grad_norm": 0.13940641283988953,
"learning_rate": 0.0001983799594144527,
"loss": 2.0948,
"step": 307
},
{
"epoch": 0.3276595744680851,
"grad_norm": 0.14055512845516205,
"learning_rate": 0.00019806140724260276,
"loss": 1.7353,
"step": 308
},
{
"epoch": 0.32872340425531915,
"grad_norm": 0.14281584322452545,
"learning_rate": 0.00019774216989570055,
"loss": 1.8931,
"step": 309
},
{
"epoch": 0.32978723404255317,
"grad_norm": 0.14208434522151947,
"learning_rate": 0.00019742225098550925,
"loss": 1.7266,
"step": 310
},
{
"epoch": 0.33085106382978724,
"grad_norm": 0.14099860191345215,
"learning_rate": 0.00019710165413150318,
"loss": 1.7742,
"step": 311
},
{
"epoch": 0.33191489361702126,
"grad_norm": 0.12579743564128876,
"learning_rate": 0.00019678038296082676,
"loss": 2.0517,
"step": 312
},
{
"epoch": 0.33297872340425533,
"grad_norm": 0.14031557738780975,
"learning_rate": 0.00019645844110825333,
"loss": 2.0609,
"step": 313
},
{
"epoch": 0.33404255319148934,
"grad_norm": 0.1430150419473648,
"learning_rate": 0.00019613583221614415,
"loss": 1.8406,
"step": 314
},
{
"epoch": 0.3351063829787234,
"grad_norm": 0.140987828373909,
"learning_rate": 0.00019581255993440726,
"loss": 1.7759,
"step": 315
},
{
"epoch": 0.33617021276595743,
"grad_norm": 0.1489114910364151,
"learning_rate": 0.000195488627920456,
"loss": 1.8095,
"step": 316
},
{
"epoch": 0.3372340425531915,
"grad_norm": 0.13374760746955872,
"learning_rate": 0.00019516403983916788,
"loss": 2.0718,
"step": 317
},
{
"epoch": 0.3382978723404255,
"grad_norm": 0.2130061239004135,
"learning_rate": 0.0001948387993628429,
"loss": 1.9204,
"step": 318
},
{
"epoch": 0.3393617021276596,
"grad_norm": 0.14826621115207672,
"learning_rate": 0.000194512910171162,
"loss": 2.02,
"step": 319
},
{
"epoch": 0.3404255319148936,
"grad_norm": 0.18779098987579346,
"learning_rate": 0.0001941863759511456,
"loss": 2.01,
"step": 320
},
{
"epoch": 0.3414893617021277,
"grad_norm": 0.12427100539207458,
"learning_rate": 0.0001938592003971118,
"loss": 2.0124,
"step": 321
},
{
"epoch": 0.3425531914893617,
"grad_norm": 0.15826719999313354,
"learning_rate": 0.00019353138721063453,
"loss": 1.6313,
"step": 322
},
{
"epoch": 0.34361702127659577,
"grad_norm": 0.15692561864852905,
"learning_rate": 0.0001932029401005018,
"loss": 1.9172,
"step": 323
},
{
"epoch": 0.3446808510638298,
"grad_norm": 0.15546223521232605,
"learning_rate": 0.0001928738627826735,
"loss": 1.6639,
"step": 324
},
{
"epoch": 0.34574468085106386,
"grad_norm": 0.17297136783599854,
"learning_rate": 0.0001925441589802396,
"loss": 1.7724,
"step": 325
},
{
"epoch": 0.3468085106382979,
"grad_norm": 0.1639593541622162,
"learning_rate": 0.00019221383242337806,
"loss": 1.8565,
"step": 326
},
{
"epoch": 0.3478723404255319,
"grad_norm": 0.15639278292655945,
"learning_rate": 0.0001918828868493123,
"loss": 1.3549,
"step": 327
},
{
"epoch": 0.34893617021276596,
"grad_norm": 0.1649019867181778,
"learning_rate": 0.00019155132600226932,
"loss": 1.7564,
"step": 328
},
{
"epoch": 0.35,
"grad_norm": 0.15620344877243042,
"learning_rate": 0.000191219153633437,
"loss": 1.8208,
"step": 329
},
{
"epoch": 0.35106382978723405,
"grad_norm": 0.14155499637126923,
"learning_rate": 0.00019088637350092195,
"loss": 2.1168,
"step": 330
},
{
"epoch": 0.35212765957446807,
"grad_norm": 0.12891899049282074,
"learning_rate": 0.00019055298936970663,
"loss": 1.9618,
"step": 331
},
{
"epoch": 0.35319148936170214,
"grad_norm": 0.13389241695404053,
"learning_rate": 0.00019021900501160728,
"loss": 1.6969,
"step": 332
},
{
"epoch": 0.35425531914893615,
"grad_norm": 0.17621472477912903,
"learning_rate": 0.00018988442420523068,
"loss": 2.229,
"step": 333
},
{
"epoch": 0.3553191489361702,
"grad_norm": 0.13503223657608032,
"learning_rate": 0.00018954925073593178,
"loss": 1.9057,
"step": 334
},
{
"epoch": 0.35638297872340424,
"grad_norm": 0.1374453455209732,
"learning_rate": 0.0001892134883957708,
"loss": 2.1979,
"step": 335
},
{
"epoch": 0.3574468085106383,
"grad_norm": 0.14095570147037506,
"learning_rate": 0.00018887714098347013,
"loss": 1.9959,
"step": 336
},
{
"epoch": 0.35851063829787233,
"grad_norm": 0.16883546113967896,
"learning_rate": 0.00018854021230437157,
"loss": 1.4638,
"step": 337
},
{
"epoch": 0.3595744680851064,
"grad_norm": 0.1310475915670395,
"learning_rate": 0.00018820270617039335,
"loss": 1.7868,
"step": 338
},
{
"epoch": 0.3606382978723404,
"grad_norm": 0.19436661899089813,
"learning_rate": 0.0001878646263999867,
"loss": 1.6914,
"step": 339
},
{
"epoch": 0.3617021276595745,
"grad_norm": 0.13219347596168518,
"learning_rate": 0.0001875259768180928,
"loss": 1.6344,
"step": 340
},
{
"epoch": 0.3627659574468085,
"grad_norm": 0.16265998780727386,
"learning_rate": 0.00018718676125609968,
"loss": 1.9101,
"step": 341
},
{
"epoch": 0.3638297872340426,
"grad_norm": 0.13733355700969696,
"learning_rate": 0.00018684698355179855,
"loss": 1.8052,
"step": 342
},
{
"epoch": 0.3648936170212766,
"grad_norm": 0.13916534185409546,
"learning_rate": 0.0001865066475493406,
"loss": 1.7605,
"step": 343
},
{
"epoch": 0.3659574468085106,
"grad_norm": 0.14417007565498352,
"learning_rate": 0.00018616575709919355,
"loss": 1.9043,
"step": 344
},
{
"epoch": 0.3670212765957447,
"grad_norm": 0.136358380317688,
"learning_rate": 0.00018582431605809784,
"loss": 1.8972,
"step": 345
},
{
"epoch": 0.3680851063829787,
"grad_norm": 0.15074212849140167,
"learning_rate": 0.0001854823282890232,
"loss": 1.7418,
"step": 346
},
{
"epoch": 0.36914893617021277,
"grad_norm": 0.13453400135040283,
"learning_rate": 0.0001851397976611249,
"loss": 2.1205,
"step": 347
},
{
"epoch": 0.3702127659574468,
"grad_norm": 0.17259922623634338,
"learning_rate": 0.00018479672804969994,
"loss": 1.9015,
"step": 348
},
{
"epoch": 0.37127659574468086,
"grad_norm": 0.1601068377494812,
"learning_rate": 0.0001844531233361433,
"loss": 1.8593,
"step": 349
},
{
"epoch": 0.3723404255319149,
"grad_norm": 0.14334119856357574,
"learning_rate": 0.0001841089874079039,
"loss": 1.7331,
"step": 350
},
{
"epoch": 0.37340425531914895,
"grad_norm": 0.18101570010185242,
"learning_rate": 0.0001837643241584406,
"loss": 1.4718,
"step": 351
},
{
"epoch": 0.37446808510638296,
"grad_norm": 0.14393842220306396,
"learning_rate": 0.00018341913748717837,
"loss": 1.9668,
"step": 352
},
{
"epoch": 0.37553191489361704,
"grad_norm": 0.15244780480861664,
"learning_rate": 0.00018307343129946396,
"loss": 1.8434,
"step": 353
},
{
"epoch": 0.37659574468085105,
"grad_norm": 0.14516274631023407,
"learning_rate": 0.00018272720950652182,
"loss": 1.8403,
"step": 354
},
{
"epoch": 0.3776595744680851,
"grad_norm": 0.13441959023475647,
"learning_rate": 0.0001823804760254098,
"loss": 2.0176,
"step": 355
},
{
"epoch": 0.37872340425531914,
"grad_norm": 0.1408858597278595,
"learning_rate": 0.00018203323477897478,
"loss": 1.9152,
"step": 356
},
{
"epoch": 0.3797872340425532,
"grad_norm": 0.13329507410526276,
"learning_rate": 0.00018168548969580849,
"loss": 1.909,
"step": 357
},
{
"epoch": 0.38085106382978723,
"grad_norm": 0.14730651676654816,
"learning_rate": 0.00018133724471020273,
"loss": 2.0144,
"step": 358
},
{
"epoch": 0.3819148936170213,
"grad_norm": 0.13821519911289215,
"learning_rate": 0.0001809885037621053,
"loss": 2.0014,
"step": 359
},
{
"epoch": 0.3829787234042553,
"grad_norm": 0.17461977899074554,
"learning_rate": 0.00018063927079707507,
"loss": 1.9347,
"step": 360
},
{
"epoch": 0.3840425531914894,
"grad_norm": 0.13765928149223328,
"learning_rate": 0.00018028954976623743,
"loss": 1.9792,
"step": 361
},
{
"epoch": 0.3851063829787234,
"grad_norm": 0.17126794159412384,
"learning_rate": 0.00017993934462623957,
"loss": 1.8147,
"step": 362
},
{
"epoch": 0.3861702127659574,
"grad_norm": 0.16928210854530334,
"learning_rate": 0.00017958865933920588,
"loss": 1.9913,
"step": 363
},
{
"epoch": 0.3872340425531915,
"grad_norm": 0.16158626973628998,
"learning_rate": 0.00017923749787269297,
"loss": 1.6398,
"step": 364
},
{
"epoch": 0.3882978723404255,
"grad_norm": 0.14897343516349792,
"learning_rate": 0.0001788858641996447,
"loss": 1.9838,
"step": 365
},
{
"epoch": 0.3893617021276596,
"grad_norm": 0.13012531399726868,
"learning_rate": 0.00017853376229834753,
"loss": 1.7733,
"step": 366
},
{
"epoch": 0.3904255319148936,
"grad_norm": 0.14755995571613312,
"learning_rate": 0.00017818119615238513,
"loss": 1.8995,
"step": 367
},
{
"epoch": 0.39148936170212767,
"grad_norm": 0.15094389021396637,
"learning_rate": 0.00017782816975059362,
"loss": 1.7396,
"step": 368
},
{
"epoch": 0.3925531914893617,
"grad_norm": 0.16519273817539215,
"learning_rate": 0.00017747468708701633,
"loss": 1.8836,
"step": 369
},
{
"epoch": 0.39361702127659576,
"grad_norm": 0.14115650951862335,
"learning_rate": 0.00017712075216085862,
"loss": 1.9956,
"step": 370
},
{
"epoch": 0.3946808510638298,
"grad_norm": 0.13805969059467316,
"learning_rate": 0.00017676636897644255,
"loss": 1.5305,
"step": 371
},
{
"epoch": 0.39574468085106385,
"grad_norm": 0.13481782376766205,
"learning_rate": 0.00017641154154316173,
"loss": 1.9364,
"step": 372
},
{
"epoch": 0.39680851063829786,
"grad_norm": 0.15755616128444672,
"learning_rate": 0.00017605627387543574,
"loss": 1.8696,
"step": 373
},
{
"epoch": 0.39787234042553193,
"grad_norm": 0.13525258004665375,
"learning_rate": 0.00017570056999266506,
"loss": 1.9387,
"step": 374
},
{
"epoch": 0.39893617021276595,
"grad_norm": 0.12865746021270752,
"learning_rate": 0.00017534443391918522,
"loss": 1.7657,
"step": 375
},
{
"epoch": 0.4,
"grad_norm": 0.15333984792232513,
"learning_rate": 0.00017498786968422154,
"loss": 1.7918,
"step": 376
},
{
"epoch": 0.40106382978723404,
"grad_norm": 0.16212083399295807,
"learning_rate": 0.0001746308813218434,
"loss": 1.6527,
"step": 377
},
{
"epoch": 0.4021276595744681,
"grad_norm": 0.1855621039867401,
"learning_rate": 0.00017427347287091857,
"loss": 1.6786,
"step": 378
},
{
"epoch": 0.4021276595744681,
"eval_loss": 2.2142834663391113,
"eval_runtime": 228.7665,
"eval_samples_per_second": 0.402,
"eval_steps_per_second": 0.402,
"step": 378
},
{
"epoch": 0.4031914893617021,
"grad_norm": 0.1534307897090912,
"learning_rate": 0.0001739156483750677,
"loss": 1.9339,
"step": 379
},
{
"epoch": 0.40425531914893614,
"grad_norm": 0.15335072576999664,
"learning_rate": 0.00017355741188261842,
"loss": 1.8833,
"step": 380
},
{
"epoch": 0.4053191489361702,
"grad_norm": 0.1465091109275818,
"learning_rate": 0.0001731987674465595,
"loss": 1.842,
"step": 381
},
{
"epoch": 0.40638297872340423,
"grad_norm": 0.1900940239429474,
"learning_rate": 0.00017283971912449527,
"loss": 1.582,
"step": 382
},
{
"epoch": 0.4074468085106383,
"grad_norm": 0.20077620446681976,
"learning_rate": 0.00017248027097859923,
"loss": 1.7603,
"step": 383
},
{
"epoch": 0.4085106382978723,
"grad_norm": 0.13478918373584747,
"learning_rate": 0.0001721204270755686,
"loss": 1.9302,
"step": 384
},
{
"epoch": 0.4095744680851064,
"grad_norm": 0.12872326374053955,
"learning_rate": 0.00017176019148657805,
"loss": 1.9961,
"step": 385
},
{
"epoch": 0.4106382978723404,
"grad_norm": 0.13909921050071716,
"learning_rate": 0.00017139956828723357,
"loss": 1.7054,
"step": 386
},
{
"epoch": 0.4117021276595745,
"grad_norm": 0.15287229418754578,
"learning_rate": 0.0001710385615575266,
"loss": 1.8546,
"step": 387
},
{
"epoch": 0.4127659574468085,
"grad_norm": 0.13635213673114777,
"learning_rate": 0.00017067717538178767,
"loss": 2.0493,
"step": 388
},
{
"epoch": 0.41382978723404257,
"grad_norm": 0.14589855074882507,
"learning_rate": 0.0001703154138486402,
"loss": 1.7876,
"step": 389
},
{
"epoch": 0.4148936170212766,
"grad_norm": 0.17141617834568024,
"learning_rate": 0.00016995328105095446,
"loss": 1.8563,
"step": 390
},
{
"epoch": 0.41595744680851066,
"grad_norm": 0.13958558440208435,
"learning_rate": 0.0001695907810858009,
"loss": 1.7036,
"step": 391
},
{
"epoch": 0.41702127659574467,
"grad_norm": 0.22088086605072021,
"learning_rate": 0.0001692279180544042,
"loss": 1.5841,
"step": 392
},
{
"epoch": 0.41808510638297874,
"grad_norm": 0.16525325179100037,
"learning_rate": 0.00016886469606209657,
"loss": 1.7766,
"step": 393
},
{
"epoch": 0.41914893617021276,
"grad_norm": 0.14994482696056366,
"learning_rate": 0.00016850111921827134,
"loss": 1.8659,
"step": 394
},
{
"epoch": 0.42021276595744683,
"grad_norm": 0.16915826499462128,
"learning_rate": 0.00016813719163633672,
"loss": 2.0314,
"step": 395
},
{
"epoch": 0.42127659574468085,
"grad_norm": 0.16112112998962402,
"learning_rate": 0.00016777291743366886,
"loss": 1.9179,
"step": 396
},
{
"epoch": 0.4223404255319149,
"grad_norm": 0.13491110503673553,
"learning_rate": 0.00016740830073156565,
"loss": 1.8374,
"step": 397
},
{
"epoch": 0.42340425531914894,
"grad_norm": 0.15047532320022583,
"learning_rate": 0.00016704334565519985,
"loss": 1.9855,
"step": 398
},
{
"epoch": 0.42446808510638295,
"grad_norm": 0.19827599823474884,
"learning_rate": 0.0001666780563335725,
"loss": 1.9839,
"step": 399
},
{
"epoch": 0.425531914893617,
"grad_norm": 0.15697021782398224,
"learning_rate": 0.00016631243689946613,
"loss": 1.8814,
"step": 400
},
{
"epoch": 0.42659574468085104,
"grad_norm": 0.15276220440864563,
"learning_rate": 0.00016594649148939824,
"loss": 1.9774,
"step": 401
},
{
"epoch": 0.4276595744680851,
"grad_norm": 0.16622979938983917,
"learning_rate": 0.00016558022424357417,
"loss": 1.4551,
"step": 402
},
{
"epoch": 0.42872340425531913,
"grad_norm": 0.20551925897598267,
"learning_rate": 0.0001652136393058405,
"loss": 1.8946,
"step": 403
},
{
"epoch": 0.4297872340425532,
"grad_norm": 0.14936675131320953,
"learning_rate": 0.0001648467408236381,
"loss": 1.7902,
"step": 404
},
{
"epoch": 0.4308510638297872,
"grad_norm": 0.18554438650608063,
"learning_rate": 0.00016447953294795514,
"loss": 1.6981,
"step": 405
},
{
"epoch": 0.4319148936170213,
"grad_norm": 0.20670993626117706,
"learning_rate": 0.00016411201983328023,
"loss": 1.2872,
"step": 406
},
{
"epoch": 0.4329787234042553,
"grad_norm": 0.1911289244890213,
"learning_rate": 0.0001637442056375553,
"loss": 1.9631,
"step": 407
},
{
"epoch": 0.4340425531914894,
"grad_norm": 0.16014768183231354,
"learning_rate": 0.00016337609452212872,
"loss": 1.8782,
"step": 408
},
{
"epoch": 0.4351063829787234,
"grad_norm": 0.15993919968605042,
"learning_rate": 0.0001630076906517081,
"loss": 1.8805,
"step": 409
},
{
"epoch": 0.43617021276595747,
"grad_norm": 0.15530019998550415,
"learning_rate": 0.00016263899819431317,
"loss": 2.1479,
"step": 410
},
{
"epoch": 0.4372340425531915,
"grad_norm": 0.14382420480251312,
"learning_rate": 0.00016227002132122867,
"loss": 1.938,
"step": 411
},
{
"epoch": 0.43829787234042555,
"grad_norm": 0.15634652972221375,
"learning_rate": 0.00016190076420695716,
"loss": 1.3769,
"step": 412
},
{
"epoch": 0.43936170212765957,
"grad_norm": 0.16254384815692902,
"learning_rate": 0.00016153123102917167,
"loss": 1.8184,
"step": 413
},
{
"epoch": 0.44042553191489364,
"grad_norm": 0.1571929007768631,
"learning_rate": 0.00016116142596866867,
"loss": 1.809,
"step": 414
},
{
"epoch": 0.44148936170212766,
"grad_norm": 0.1451842188835144,
"learning_rate": 0.00016079135320932062,
"loss": 1.8872,
"step": 415
},
{
"epoch": 0.4425531914893617,
"grad_norm": 0.1494993418455124,
"learning_rate": 0.00016042101693802856,
"loss": 1.5983,
"step": 416
},
{
"epoch": 0.44361702127659575,
"grad_norm": 0.13900022208690643,
"learning_rate": 0.0001600504213446749,
"loss": 1.838,
"step": 417
},
{
"epoch": 0.44468085106382976,
"grad_norm": 0.14836052060127258,
"learning_rate": 0.00015967957062207593,
"loss": 1.7853,
"step": 418
},
{
"epoch": 0.44574468085106383,
"grad_norm": 0.1537352055311203,
"learning_rate": 0.00015930846896593449,
"loss": 1.7041,
"step": 419
},
{
"epoch": 0.44680851063829785,
"grad_norm": 0.1612105369567871,
"learning_rate": 0.00015893712057479215,
"loss": 1.9813,
"step": 420
},
{
"epoch": 0.4478723404255319,
"grad_norm": 0.16644567251205444,
"learning_rate": 0.00015856552964998232,
"loss": 1.907,
"step": 421
},
{
"epoch": 0.44893617021276594,
"grad_norm": 0.1353951394557953,
"learning_rate": 0.00015819370039558213,
"loss": 2.0852,
"step": 422
},
{
"epoch": 0.45,
"grad_norm": 0.15057289600372314,
"learning_rate": 0.0001578216370183651,
"loss": 2.0359,
"step": 423
},
{
"epoch": 0.451063829787234,
"grad_norm": 0.24671976268291473,
"learning_rate": 0.00015744934372775377,
"loss": 1.5343,
"step": 424
},
{
"epoch": 0.4521276595744681,
"grad_norm": 0.13813172280788422,
"learning_rate": 0.00015707682473577162,
"loss": 1.8472,
"step": 425
},
{
"epoch": 0.4531914893617021,
"grad_norm": 0.1736101508140564,
"learning_rate": 0.0001567040842569958,
"loss": 2.1779,
"step": 426
},
{
"epoch": 0.4542553191489362,
"grad_norm": 0.16814932227134705,
"learning_rate": 0.00015633112650850919,
"loss": 1.7505,
"step": 427
},
{
"epoch": 0.4553191489361702,
"grad_norm": 0.570379376411438,
"learning_rate": 0.0001559579557098529,
"loss": 1.9268,
"step": 428
},
{
"epoch": 0.4563829787234043,
"grad_norm": 0.15529078245162964,
"learning_rate": 0.00015558457608297837,
"loss": 1.609,
"step": 429
},
{
"epoch": 0.4574468085106383,
"grad_norm": 0.21124756336212158,
"learning_rate": 0.0001552109918521996,
"loss": 1.8287,
"step": 430
},
{
"epoch": 0.45851063829787236,
"grad_norm": 0.15143603086471558,
"learning_rate": 0.00015483720724414565,
"loss": 1.8677,
"step": 431
},
{
"epoch": 0.4595744680851064,
"grad_norm": 0.17100220918655396,
"learning_rate": 0.00015446322648771235,
"loss": 1.7661,
"step": 432
},
{
"epoch": 0.46063829787234045,
"grad_norm": 0.14617975056171417,
"learning_rate": 0.00015408905381401487,
"loss": 1.8792,
"step": 433
},
{
"epoch": 0.46170212765957447,
"grad_norm": 0.19449691474437714,
"learning_rate": 0.00015371469345633952,
"loss": 1.7281,
"step": 434
},
{
"epoch": 0.4627659574468085,
"grad_norm": 0.17838919162750244,
"learning_rate": 0.0001533401496500961,
"loss": 1.9533,
"step": 435
},
{
"epoch": 0.46382978723404256,
"grad_norm": 0.1702663004398346,
"learning_rate": 0.00015296542663276996,
"loss": 1.9352,
"step": 436
},
{
"epoch": 0.4648936170212766,
"grad_norm": 0.16134527325630188,
"learning_rate": 0.0001525905286438739,
"loss": 2.0971,
"step": 437
},
{
"epoch": 0.46595744680851064,
"grad_norm": 0.15274085104465485,
"learning_rate": 0.00015221545992490033,
"loss": 2.0342,
"step": 438
},
{
"epoch": 0.46702127659574466,
"grad_norm": 0.14991699159145355,
"learning_rate": 0.00015184022471927323,
"loss": 1.8673,
"step": 439
},
{
"epoch": 0.46808510638297873,
"grad_norm": 0.24493008852005005,
"learning_rate": 0.00015146482727230025,
"loss": 1.6853,
"step": 440
},
{
"epoch": 0.46914893617021275,
"grad_norm": 0.15089358389377594,
"learning_rate": 0.00015108927183112443,
"loss": 1.9943,
"step": 441
},
{
"epoch": 0.4702127659574468,
"grad_norm": 0.1807398796081543,
"learning_rate": 0.00015071356264467653,
"loss": 1.9267,
"step": 442
},
{
"epoch": 0.47127659574468084,
"grad_norm": 0.16581107676029205,
"learning_rate": 0.0001503377039636265,
"loss": 1.8497,
"step": 443
},
{
"epoch": 0.4723404255319149,
"grad_norm": 0.15481366217136383,
"learning_rate": 0.0001499617000403359,
"loss": 1.889,
"step": 444
},
{
"epoch": 0.4734042553191489,
"grad_norm": 0.2134893834590912,
"learning_rate": 0.00014958555512880923,
"loss": 1.9578,
"step": 445
},
{
"epoch": 0.474468085106383,
"grad_norm": 0.16061900556087494,
"learning_rate": 0.00014920927348464633,
"loss": 1.5451,
"step": 446
},
{
"epoch": 0.475531914893617,
"grad_norm": 0.1559426188468933,
"learning_rate": 0.00014883285936499388,
"loss": 1.7151,
"step": 447
},
{
"epoch": 0.4765957446808511,
"grad_norm": 0.1664920598268509,
"learning_rate": 0.0001484563170284974,
"loss": 1.9813,
"step": 448
},
{
"epoch": 0.4776595744680851,
"grad_norm": 0.14086946845054626,
"learning_rate": 0.00014807965073525297,
"loss": 1.8155,
"step": 449
},
{
"epoch": 0.4787234042553192,
"grad_norm": 0.15908385813236237,
"learning_rate": 0.00014770286474675908,
"loss": 1.8051,
"step": 450
},
{
"epoch": 0.4797872340425532,
"grad_norm": 0.14929479360580444,
"learning_rate": 0.00014732596332586847,
"loss": 1.9001,
"step": 451
},
{
"epoch": 0.4808510638297872,
"grad_norm": 0.15190398693084717,
"learning_rate": 0.00014694895073673987,
"loss": 1.7345,
"step": 452
},
{
"epoch": 0.4819148936170213,
"grad_norm": 0.1622495800256729,
"learning_rate": 0.00014657183124478962,
"loss": 1.8637,
"step": 453
},
{
"epoch": 0.4829787234042553,
"grad_norm": 0.1647133231163025,
"learning_rate": 0.00014619460911664367,
"loss": 1.6461,
"step": 454
},
{
"epoch": 0.48404255319148937,
"grad_norm": 0.15832483768463135,
"learning_rate": 0.00014581728862008905,
"loss": 1.8695,
"step": 455
},
{
"epoch": 0.4851063829787234,
"grad_norm": 0.21159999072551727,
"learning_rate": 0.00014543987402402576,
"loss": 1.4445,
"step": 456
},
{
"epoch": 0.48617021276595745,
"grad_norm": 0.3298247158527374,
"learning_rate": 0.00014506236959841841,
"loss": 2.0541,
"step": 457
},
{
"epoch": 0.48723404255319147,
"grad_norm": 0.16765986382961273,
"learning_rate": 0.00014468477961424798,
"loss": 1.766,
"step": 458
},
{
"epoch": 0.48829787234042554,
"grad_norm": 0.15714918076992035,
"learning_rate": 0.00014430710834346326,
"loss": 1.9625,
"step": 459
},
{
"epoch": 0.48936170212765956,
"grad_norm": 0.16249720752239227,
"learning_rate": 0.0001439293600589329,
"loss": 1.7994,
"step": 460
},
{
"epoch": 0.49042553191489363,
"grad_norm": 0.16646772623062134,
"learning_rate": 0.00014355153903439672,
"loss": 1.9068,
"step": 461
},
{
"epoch": 0.49148936170212765,
"grad_norm": 0.22211192548274994,
"learning_rate": 0.00014317364954441754,
"loss": 1.9102,
"step": 462
},
{
"epoch": 0.4925531914893617,
"grad_norm": 0.17935901880264282,
"learning_rate": 0.0001427956958643328,
"loss": 1.8746,
"step": 463
},
{
"epoch": 0.49361702127659574,
"grad_norm": 0.14291420578956604,
"learning_rate": 0.00014241768227020622,
"loss": 2.1468,
"step": 464
},
{
"epoch": 0.4946808510638298,
"grad_norm": 0.18176282942295074,
"learning_rate": 0.00014203961303877922,
"loss": 1.5823,
"step": 465
},
{
"epoch": 0.4957446808510638,
"grad_norm": 0.15742754936218262,
"learning_rate": 0.00014166149244742278,
"loss": 1.7668,
"step": 466
},
{
"epoch": 0.4968085106382979,
"grad_norm": 0.17805592715740204,
"learning_rate": 0.000141283324774089,
"loss": 1.8197,
"step": 467
},
{
"epoch": 0.4978723404255319,
"grad_norm": 0.14772458374500275,
"learning_rate": 0.00014090511429726255,
"loss": 1.7012,
"step": 468
},
{
"epoch": 0.498936170212766,
"grad_norm": 0.14443214237689972,
"learning_rate": 0.00014052686529591243,
"loss": 2.0839,
"step": 469
},
{
"epoch": 0.5,
"grad_norm": 0.1701783388853073,
"learning_rate": 0.00014014858204944351,
"loss": 1.6082,
"step": 470
},
{
"epoch": 0.5010638297872341,
"grad_norm": 0.17072685062885284,
"learning_rate": 0.00013977026883764805,
"loss": 1.8665,
"step": 471
},
{
"epoch": 0.502127659574468,
"grad_norm": 0.16589944064617157,
"learning_rate": 0.00013939192994065727,
"loss": 1.9802,
"step": 472
},
{
"epoch": 0.5031914893617021,
"grad_norm": 0.19318854808807373,
"learning_rate": 0.0001390135696388932,
"loss": 1.3777,
"step": 473
},
{
"epoch": 0.5042553191489362,
"grad_norm": 0.1829719990491867,
"learning_rate": 0.00013863519221301982,
"loss": 1.8002,
"step": 474
},
{
"epoch": 0.5053191489361702,
"grad_norm": 0.16888178884983063,
"learning_rate": 0.00013825680194389496,
"loss": 2.0893,
"step": 475
},
{
"epoch": 0.5063829787234042,
"grad_norm": 0.22794437408447266,
"learning_rate": 0.0001378784031125217,
"loss": 1.326,
"step": 476
},
{
"epoch": 0.5074468085106383,
"grad_norm": 0.15228323638439178,
"learning_rate": 0.0001375,
"loss": 1.7156,
"step": 477
},
{
"epoch": 0.5085106382978724,
"grad_norm": 0.1590748280286789,
"learning_rate": 0.00013712159688747833,
"loss": 1.8796,
"step": 478
},
{
"epoch": 0.5095744680851064,
"grad_norm": 0.17704014480113983,
"learning_rate": 0.00013674319805610506,
"loss": 2.0174,
"step": 479
},
{
"epoch": 0.5106382978723404,
"grad_norm": 0.1526685655117035,
"learning_rate": 0.00013636480778698022,
"loss": 1.9632,
"step": 480
},
{
"epoch": 0.5117021276595745,
"grad_norm": 0.1670306921005249,
"learning_rate": 0.0001359864303611068,
"loss": 1.9358,
"step": 481
},
{
"epoch": 0.5127659574468085,
"grad_norm": 0.1440451741218567,
"learning_rate": 0.00013560807005934272,
"loss": 1.6262,
"step": 482
},
{
"epoch": 0.5138297872340426,
"grad_norm": 0.14946603775024414,
"learning_rate": 0.000135229731162352,
"loss": 1.8391,
"step": 483
},
{
"epoch": 0.5148936170212766,
"grad_norm": 0.40189728140830994,
"learning_rate": 0.00013485141795055653,
"loss": 1.9325,
"step": 484
},
{
"epoch": 0.5159574468085106,
"grad_norm": 0.1524064540863037,
"learning_rate": 0.0001344731347040876,
"loss": 1.9053,
"step": 485
},
{
"epoch": 0.5170212765957447,
"grad_norm": 0.1905772089958191,
"learning_rate": 0.00013409488570273752,
"loss": 1.6801,
"step": 486
},
{
"epoch": 0.5180851063829788,
"grad_norm": 0.18460091948509216,
"learning_rate": 0.00013371667522591105,
"loss": 1.716,
"step": 487
},
{
"epoch": 0.5191489361702127,
"grad_norm": 0.1699136197566986,
"learning_rate": 0.00013333850755257727,
"loss": 1.955,
"step": 488
},
{
"epoch": 0.5202127659574468,
"grad_norm": 0.14414088428020477,
"learning_rate": 0.00013296038696122085,
"loss": 1.8916,
"step": 489
},
{
"epoch": 0.5212765957446809,
"grad_norm": 0.2500864267349243,
"learning_rate": 0.00013258231772979382,
"loss": 1.7159,
"step": 490
},
{
"epoch": 0.5223404255319148,
"grad_norm": 0.16285300254821777,
"learning_rate": 0.00013220430413566722,
"loss": 1.6846,
"step": 491
},
{
"epoch": 0.5234042553191489,
"grad_norm": 0.13979285955429077,
"learning_rate": 0.00013182635045558248,
"loss": 1.8205,
"step": 492
},
{
"epoch": 0.524468085106383,
"grad_norm": 0.174757719039917,
"learning_rate": 0.0001314484609656033,
"loss": 1.7025,
"step": 493
},
{
"epoch": 0.5255319148936171,
"grad_norm": 0.12527626752853394,
"learning_rate": 0.00013107063994106713,
"loss": 1.5829,
"step": 494
},
{
"epoch": 0.526595744680851,
"grad_norm": 0.16596892476081848,
"learning_rate": 0.00013069289165653676,
"loss": 1.5709,
"step": 495
},
{
"epoch": 0.5276595744680851,
"grad_norm": 0.16609230637550354,
"learning_rate": 0.00013031522038575206,
"loss": 1.8311,
"step": 496
},
{
"epoch": 0.5287234042553192,
"grad_norm": 0.15371407568454742,
"learning_rate": 0.00012993763040158158,
"loss": 1.7831,
"step": 497
},
{
"epoch": 0.5297872340425532,
"grad_norm": 0.14225727319717407,
"learning_rate": 0.00012956012597597428,
"loss": 1.8035,
"step": 498
},
{
"epoch": 0.5308510638297872,
"grad_norm": 0.15067608654499054,
"learning_rate": 0.000129182711379911,
"loss": 1.7671,
"step": 499
},
{
"epoch": 0.5319148936170213,
"grad_norm": 0.15496258437633514,
"learning_rate": 0.00012880539088335635,
"loss": 1.6137,
"step": 500
},
{
"epoch": 0.5329787234042553,
"grad_norm": 0.1736234724521637,
"learning_rate": 0.00012842816875521042,
"loss": 2.233,
"step": 501
},
{
"epoch": 0.5340425531914894,
"grad_norm": 0.1675282120704651,
"learning_rate": 0.00012805104926326018,
"loss": 1.9801,
"step": 502
},
{
"epoch": 0.5351063829787234,
"grad_norm": 0.17056336998939514,
"learning_rate": 0.00012767403667413154,
"loss": 1.8477,
"step": 503
},
{
"epoch": 0.5361702127659574,
"grad_norm": 0.1884344220161438,
"learning_rate": 0.00012729713525324094,
"loss": 1.5402,
"step": 504
},
{
"epoch": 0.5372340425531915,
"grad_norm": 0.13208945095539093,
"learning_rate": 0.00012692034926474707,
"loss": 1.4391,
"step": 505
},
{
"epoch": 0.5382978723404256,
"grad_norm": 0.1935010552406311,
"learning_rate": 0.0001265436829715026,
"loss": 1.9159,
"step": 506
},
{
"epoch": 0.5393617021276595,
"grad_norm": 0.1699482947587967,
"learning_rate": 0.0001261671406350061,
"loss": 1.7707,
"step": 507
},
{
"epoch": 0.5404255319148936,
"grad_norm": 0.15546508133411407,
"learning_rate": 0.00012579072651535368,
"loss": 1.856,
"step": 508
},
{
"epoch": 0.5414893617021277,
"grad_norm": 0.43234968185424805,
"learning_rate": 0.00012541444487119076,
"loss": 1.696,
"step": 509
},
{
"epoch": 0.5425531914893617,
"grad_norm": 0.15519973635673523,
"learning_rate": 0.00012503829995966413,
"loss": 2.0373,
"step": 510
},
{
"epoch": 0.5436170212765957,
"grad_norm": 0.18885447084903717,
"learning_rate": 0.00012466229603637347,
"loss": 1.7524,
"step": 511
},
{
"epoch": 0.5446808510638298,
"grad_norm": 0.19030006229877472,
"learning_rate": 0.00012428643735532352,
"loss": 1.5325,
"step": 512
},
{
"epoch": 0.5457446808510639,
"grad_norm": 0.15081647038459778,
"learning_rate": 0.0001239107281688756,
"loss": 2.0715,
"step": 513
},
{
"epoch": 0.5468085106382978,
"grad_norm": 0.1772020161151886,
"learning_rate": 0.00012353517272769982,
"loss": 1.7717,
"step": 514
},
{
"epoch": 0.5478723404255319,
"grad_norm": 0.3538069427013397,
"learning_rate": 0.00012315977528072681,
"loss": 2.1194,
"step": 515
},
{
"epoch": 0.548936170212766,
"grad_norm": 0.16277366876602173,
"learning_rate": 0.0001227845400750997,
"loss": 1.6826,
"step": 516
},
{
"epoch": 0.55,
"grad_norm": 0.17881731688976288,
"learning_rate": 0.00012240947135612611,
"loss": 1.9494,
"step": 517
},
{
"epoch": 0.551063829787234,
"grad_norm": 0.1840476393699646,
"learning_rate": 0.00012203457336723006,
"loss": 1.8986,
"step": 518
},
{
"epoch": 0.5521276595744681,
"grad_norm": 0.14919424057006836,
"learning_rate": 0.0001216598503499039,
"loss": 2.0112,
"step": 519
},
{
"epoch": 0.5531914893617021,
"grad_norm": 0.1450183093547821,
"learning_rate": 0.00012128530654366052,
"loss": 1.8921,
"step": 520
},
{
"epoch": 0.5542553191489362,
"grad_norm": 0.16653242707252502,
"learning_rate": 0.00012091094618598516,
"loss": 1.5374,
"step": 521
},
{
"epoch": 0.5553191489361702,
"grad_norm": 0.14227531850337982,
"learning_rate": 0.00012053677351228761,
"loss": 1.8546,
"step": 522
},
{
"epoch": 0.5563829787234043,
"grad_norm": 0.14086997509002686,
"learning_rate": 0.00012016279275585438,
"loss": 1.6852,
"step": 523
},
{
"epoch": 0.5574468085106383,
"grad_norm": 0.20002402365207672,
"learning_rate": 0.00011978900814780039,
"loss": 1.7305,
"step": 524
},
{
"epoch": 0.5585106382978723,
"grad_norm": 0.15960435569286346,
"learning_rate": 0.00011941542391702168,
"loss": 1.3646,
"step": 525
},
{
"epoch": 0.5595744680851064,
"grad_norm": 0.15262611210346222,
"learning_rate": 0.00011904204429014717,
"loss": 1.7815,
"step": 526
},
{
"epoch": 0.5606382978723404,
"grad_norm": 0.155854269862175,
"learning_rate": 0.00011866887349149086,
"loss": 1.9681,
"step": 527
},
{
"epoch": 0.5617021276595745,
"grad_norm": 0.2157667726278305,
"learning_rate": 0.00011829591574300422,
"loss": 1.6775,
"step": 528
},
{
"epoch": 0.5627659574468085,
"grad_norm": 0.1676628291606903,
"learning_rate": 0.00011792317526422839,
"loss": 1.7169,
"step": 529
},
{
"epoch": 0.5638297872340425,
"grad_norm": 0.1837339997291565,
"learning_rate": 0.00011755065627224626,
"loss": 1.7411,
"step": 530
},
{
"epoch": 0.5648936170212766,
"grad_norm": 0.14545366168022156,
"learning_rate": 0.00011717836298163492,
"loss": 1.7205,
"step": 531
},
{
"epoch": 0.5659574468085107,
"grad_norm": 0.1784188151359558,
"learning_rate": 0.00011680629960441794,
"loss": 1.8765,
"step": 532
},
{
"epoch": 0.5670212765957446,
"grad_norm": 0.2114858329296112,
"learning_rate": 0.0001164344703500177,
"loss": 1.9443,
"step": 533
},
{
"epoch": 0.5680851063829787,
"grad_norm": 0.17516778409481049,
"learning_rate": 0.00011606287942520788,
"loss": 1.6652,
"step": 534
},
{
"epoch": 0.5691489361702128,
"grad_norm": 0.15822292864322662,
"learning_rate": 0.00011569153103406557,
"loss": 1.6422,
"step": 535
},
{
"epoch": 0.5702127659574469,
"grad_norm": 0.17169679701328278,
"learning_rate": 0.00011532042937792406,
"loss": 1.5048,
"step": 536
},
{
"epoch": 0.5712765957446808,
"grad_norm": 0.16236886382102966,
"learning_rate": 0.0001149495786553251,
"loss": 1.4069,
"step": 537
},
{
"epoch": 0.5723404255319149,
"grad_norm": 0.16006197035312653,
"learning_rate": 0.00011457898306197145,
"loss": 1.8959,
"step": 538
},
{
"epoch": 0.573404255319149,
"grad_norm": 0.17285117506980896,
"learning_rate": 0.00011420864679067941,
"loss": 1.9631,
"step": 539
},
{
"epoch": 0.574468085106383,
"grad_norm": 0.18171286582946777,
"learning_rate": 0.00011383857403133136,
"loss": 1.73,
"step": 540
},
{
"epoch": 0.575531914893617,
"grad_norm": 0.14552678167819977,
"learning_rate": 0.00011346876897082838,
"loss": 1.4612,
"step": 541
},
{
"epoch": 0.5765957446808511,
"grad_norm": 0.16404442489147186,
"learning_rate": 0.00011309923579304291,
"loss": 2.1242,
"step": 542
},
{
"epoch": 0.5776595744680851,
"grad_norm": 0.1597999781370163,
"learning_rate": 0.00011272997867877135,
"loss": 1.8759,
"step": 543
},
{
"epoch": 0.5787234042553191,
"grad_norm": 0.16365398466587067,
"learning_rate": 0.00011236100180568684,
"loss": 1.8092,
"step": 544
},
{
"epoch": 0.5797872340425532,
"grad_norm": 0.14701658487319946,
"learning_rate": 0.0001119923093482919,
"loss": 2.0718,
"step": 545
},
{
"epoch": 0.5808510638297872,
"grad_norm": 0.15823504328727722,
"learning_rate": 0.00011162390547787127,
"loss": 1.9042,
"step": 546
},
{
"epoch": 0.5819148936170213,
"grad_norm": 0.19485561549663544,
"learning_rate": 0.00011125579436244471,
"loss": 1.7576,
"step": 547
},
{
"epoch": 0.5829787234042553,
"grad_norm": 0.17676511406898499,
"learning_rate": 0.0001108879801667198,
"loss": 1.6758,
"step": 548
},
{
"epoch": 0.5840425531914893,
"grad_norm": 0.15661020576953888,
"learning_rate": 0.00011052046705204486,
"loss": 1.918,
"step": 549
},
{
"epoch": 0.5851063829787234,
"grad_norm": 0.1745099276304245,
"learning_rate": 0.0001101532591763619,
"loss": 1.827,
"step": 550
},
{
"epoch": 0.5861702127659575,
"grad_norm": 0.16982468962669373,
"learning_rate": 0.00010978636069415951,
"loss": 1.9529,
"step": 551
},
{
"epoch": 0.5872340425531914,
"grad_norm": 0.24788892269134521,
"learning_rate": 0.00010941977575642584,
"loss": 1.9002,
"step": 552
},
{
"epoch": 0.5882978723404255,
"grad_norm": 0.17355118691921234,
"learning_rate": 0.00010905350851060183,
"loss": 1.9254,
"step": 553
},
{
"epoch": 0.5893617021276596,
"grad_norm": 0.2144351601600647,
"learning_rate": 0.0001086875631005339,
"loss": 1.7644,
"step": 554
},
{
"epoch": 0.5904255319148937,
"grad_norm": 0.2839835584163666,
"learning_rate": 0.00010832194366642756,
"loss": 1.9132,
"step": 555
},
{
"epoch": 0.5914893617021276,
"grad_norm": 0.17283079028129578,
"learning_rate": 0.00010795665434480018,
"loss": 1.7372,
"step": 556
},
{
"epoch": 0.5925531914893617,
"grad_norm": 0.18223372101783752,
"learning_rate": 0.00010759169926843437,
"loss": 1.882,
"step": 557
},
{
"epoch": 0.5936170212765958,
"grad_norm": 0.15663963556289673,
"learning_rate": 0.00010722708256633115,
"loss": 1.2971,
"step": 558
},
{
"epoch": 0.5946808510638298,
"grad_norm": 0.14517734944820404,
"learning_rate": 0.0001068628083636633,
"loss": 1.748,
"step": 559
},
{
"epoch": 0.5957446808510638,
"grad_norm": 0.16935127973556519,
"learning_rate": 0.00010649888078172865,
"loss": 1.5534,
"step": 560
},
{
"epoch": 0.5968085106382979,
"grad_norm": 0.14080187678337097,
"learning_rate": 0.00010613530393790346,
"loss": 1.8392,
"step": 561
},
{
"epoch": 0.597872340425532,
"grad_norm": 0.1790747493505478,
"learning_rate": 0.00010577208194559582,
"loss": 1.7639,
"step": 562
},
{
"epoch": 0.5989361702127659,
"grad_norm": 0.1672879010438919,
"learning_rate": 0.00010540921891419911,
"loss": 1.851,
"step": 563
},
{
"epoch": 0.6,
"grad_norm": 0.16451695561408997,
"learning_rate": 0.00010504671894904557,
"loss": 1.8741,
"step": 564
},
{
"epoch": 0.601063829787234,
"grad_norm": 0.15079449117183685,
"learning_rate": 0.00010468458615135978,
"loss": 1.7664,
"step": 565
},
{
"epoch": 0.6021276595744681,
"grad_norm": 0.20589475333690643,
"learning_rate": 0.00010432282461821236,
"loss": 1.7941,
"step": 566
},
{
"epoch": 0.6031914893617021,
"grad_norm": 0.17053169012069702,
"learning_rate": 0.00010396143844247341,
"loss": 1.9847,
"step": 567
},
{
"epoch": 0.6031914893617021,
"eval_loss": 2.179853677749634,
"eval_runtime": 228.6364,
"eval_samples_per_second": 0.402,
"eval_steps_per_second": 0.402,
"step": 567
},
{
"epoch": 0.6042553191489362,
"grad_norm": 0.16731402277946472,
"learning_rate": 0.00010360043171276646,
"loss": 2.0162,
"step": 568
},
{
"epoch": 0.6053191489361702,
"grad_norm": 0.19248618185520172,
"learning_rate": 0.00010323980851342199,
"loss": 1.8991,
"step": 569
},
{
"epoch": 0.6063829787234043,
"grad_norm": 0.1472175270318985,
"learning_rate": 0.00010287957292443142,
"loss": 1.754,
"step": 570
},
{
"epoch": 0.6074468085106383,
"grad_norm": 0.1666106879711151,
"learning_rate": 0.00010251972902140081,
"loss": 1.7816,
"step": 571
},
{
"epoch": 0.6085106382978723,
"grad_norm": 0.2124800980091095,
"learning_rate": 0.00010216028087550478,
"loss": 1.9329,
"step": 572
},
{
"epoch": 0.6095744680851064,
"grad_norm": 0.1831895411014557,
"learning_rate": 0.0001018012325534405,
"loss": 1.5186,
"step": 573
},
{
"epoch": 0.6106382978723405,
"grad_norm": 0.17318058013916016,
"learning_rate": 0.00010144258811738161,
"loss": 1.719,
"step": 574
},
{
"epoch": 0.6117021276595744,
"grad_norm": 0.15998592972755432,
"learning_rate": 0.00010108435162493232,
"loss": 1.5606,
"step": 575
},
{
"epoch": 0.6127659574468085,
"grad_norm": 0.16763073205947876,
"learning_rate": 0.00010072652712908143,
"loss": 1.5935,
"step": 576
},
{
"epoch": 0.6138297872340426,
"grad_norm": 0.15910252928733826,
"learning_rate": 0.00010036911867815662,
"loss": 1.7361,
"step": 577
},
{
"epoch": 0.6148936170212767,
"grad_norm": 0.16548095643520355,
"learning_rate": 0.00010001213031577846,
"loss": 1.9521,
"step": 578
},
{
"epoch": 0.6159574468085106,
"grad_norm": 0.15539593994617462,
"learning_rate": 9.965556608081477e-05,
"loss": 1.9327,
"step": 579
},
{
"epoch": 0.6170212765957447,
"grad_norm": 0.14825530350208282,
"learning_rate": 9.929943000733493e-05,
"loss": 1.8114,
"step": 580
},
{
"epoch": 0.6180851063829788,
"grad_norm": 0.19313831627368927,
"learning_rate": 9.894372612456429e-05,
"loss": 1.7668,
"step": 581
},
{
"epoch": 0.6191489361702127,
"grad_norm": 0.16411848366260529,
"learning_rate": 9.858845845683834e-05,
"loss": 1.7161,
"step": 582
},
{
"epoch": 0.6202127659574468,
"grad_norm": 0.14812296628952026,
"learning_rate": 9.823363102355746e-05,
"loss": 1.8038,
"step": 583
},
{
"epoch": 0.6212765957446809,
"grad_norm": 0.184005007147789,
"learning_rate": 9.78792478391414e-05,
"loss": 1.8267,
"step": 584
},
{
"epoch": 0.6223404255319149,
"grad_norm": 0.16978758573532104,
"learning_rate": 9.752531291298366e-05,
"loss": 1.5657,
"step": 585
},
{
"epoch": 0.6234042553191489,
"grad_norm": 0.193622887134552,
"learning_rate": 9.71718302494064e-05,
"loss": 1.7917,
"step": 586
},
{
"epoch": 0.624468085106383,
"grad_norm": 0.19069254398345947,
"learning_rate": 9.68188038476149e-05,
"loss": 1.8504,
"step": 587
},
{
"epoch": 0.625531914893617,
"grad_norm": 0.17065876722335815,
"learning_rate": 9.64662377016525e-05,
"loss": 1.2871,
"step": 588
},
{
"epoch": 0.6265957446808511,
"grad_norm": 0.15738852322101593,
"learning_rate": 9.61141358003553e-05,
"loss": 1.6241,
"step": 589
},
{
"epoch": 0.6276595744680851,
"grad_norm": 0.19283810257911682,
"learning_rate": 9.576250212730706e-05,
"loss": 2.051,
"step": 590
},
{
"epoch": 0.6287234042553191,
"grad_norm": 0.24007220566272736,
"learning_rate": 9.54113406607941e-05,
"loss": 1.6686,
"step": 591
},
{
"epoch": 0.6297872340425532,
"grad_norm": 0.17241282761096954,
"learning_rate": 9.50606553737604e-05,
"loss": 1.9343,
"step": 592
},
{
"epoch": 0.6308510638297873,
"grad_norm": 0.16544866561889648,
"learning_rate": 9.471045023376259e-05,
"loss": 1.8793,
"step": 593
},
{
"epoch": 0.6319148936170212,
"grad_norm": 0.220997154712677,
"learning_rate": 9.436072920292493e-05,
"loss": 1.5908,
"step": 594
},
{
"epoch": 0.6329787234042553,
"grad_norm": 0.1769099086523056,
"learning_rate": 9.401149623789471e-05,
"loss": 1.3959,
"step": 595
},
{
"epoch": 0.6340425531914894,
"grad_norm": 0.16651903092861176,
"learning_rate": 9.366275528979728e-05,
"loss": 1.5989,
"step": 596
},
{
"epoch": 0.6351063829787233,
"grad_norm": 0.2315669059753418,
"learning_rate": 9.331451030419158e-05,
"loss": 1.8365,
"step": 597
},
{
"epoch": 0.6361702127659574,
"grad_norm": 0.31060174107551575,
"learning_rate": 9.296676522102523e-05,
"loss": 1.1766,
"step": 598
},
{
"epoch": 0.6372340425531915,
"grad_norm": 0.20410263538360596,
"learning_rate": 9.261952397459023e-05,
"loss": 1.8241,
"step": 599
},
{
"epoch": 0.6382978723404256,
"grad_norm": 0.21315258741378784,
"learning_rate": 9.22727904934782e-05,
"loss": 1.8114,
"step": 600
},
{
"epoch": 0.6393617021276595,
"grad_norm": 0.17505568265914917,
"learning_rate": 9.192656870053603e-05,
"loss": 1.6593,
"step": 601
},
{
"epoch": 0.6404255319148936,
"grad_norm": 0.1761648803949356,
"learning_rate": 9.158086251282166e-05,
"loss": 1.5484,
"step": 602
},
{
"epoch": 0.6414893617021277,
"grad_norm": 0.2948690354824066,
"learning_rate": 9.123567584155942e-05,
"loss": 1.7679,
"step": 603
},
{
"epoch": 0.6425531914893617,
"grad_norm": 0.15874460339546204,
"learning_rate": 9.089101259209614e-05,
"loss": 1.5578,
"step": 604
},
{
"epoch": 0.6436170212765957,
"grad_norm": 0.14462539553642273,
"learning_rate": 9.054687666385673e-05,
"loss": 1.7078,
"step": 605
},
{
"epoch": 0.6446808510638298,
"grad_norm": 0.16934645175933838,
"learning_rate": 9.020327195030004e-05,
"loss": 1.9469,
"step": 606
},
{
"epoch": 0.6457446808510638,
"grad_norm": 0.23816823959350586,
"learning_rate": 8.98602023388751e-05,
"loss": 1.7715,
"step": 607
},
{
"epoch": 0.6468085106382979,
"grad_norm": 0.17519444227218628,
"learning_rate": 8.951767171097683e-05,
"loss": 1.6478,
"step": 608
},
{
"epoch": 0.6478723404255319,
"grad_norm": 0.18539506196975708,
"learning_rate": 8.917568394190218e-05,
"loss": 1.5463,
"step": 609
},
{
"epoch": 0.648936170212766,
"grad_norm": 0.18039405345916748,
"learning_rate": 8.883424290080644e-05,
"loss": 1.6682,
"step": 610
},
{
"epoch": 0.65,
"grad_norm": 0.165382981300354,
"learning_rate": 8.849335245065936e-05,
"loss": 1.7848,
"step": 611
},
{
"epoch": 0.6510638297872341,
"grad_norm": 0.18024031817913055,
"learning_rate": 8.815301644820148e-05,
"loss": 1.0508,
"step": 612
},
{
"epoch": 0.652127659574468,
"grad_norm": 0.15618403255939484,
"learning_rate": 8.781323874390038e-05,
"loss": 1.757,
"step": 613
},
{
"epoch": 0.6531914893617021,
"grad_norm": 0.15764309465885162,
"learning_rate": 8.747402318190722e-05,
"loss": 1.7571,
"step": 614
},
{
"epoch": 0.6542553191489362,
"grad_norm": 0.19766607880592346,
"learning_rate": 8.713537360001336e-05,
"loss": 1.9208,
"step": 615
},
{
"epoch": 0.6553191489361702,
"grad_norm": 0.19721117615699768,
"learning_rate": 8.679729382960666e-05,
"loss": 1.5942,
"step": 616
},
{
"epoch": 0.6563829787234042,
"grad_norm": 0.1823943555355072,
"learning_rate": 8.645978769562843e-05,
"loss": 1.7942,
"step": 617
},
{
"epoch": 0.6574468085106383,
"grad_norm": 0.17587131261825562,
"learning_rate": 8.612285901652992e-05,
"loss": 1.5356,
"step": 618
},
{
"epoch": 0.6585106382978724,
"grad_norm": 0.16799600422382355,
"learning_rate": 8.578651160422923e-05,
"loss": 1.921,
"step": 619
},
{
"epoch": 0.6595744680851063,
"grad_norm": 0.1665090024471283,
"learning_rate": 8.545074926406819e-05,
"loss": 2.0355,
"step": 620
},
{
"epoch": 0.6606382978723404,
"grad_norm": 0.15373064577579498,
"learning_rate": 8.51155757947693e-05,
"loss": 1.8153,
"step": 621
},
{
"epoch": 0.6617021276595745,
"grad_norm": 0.1743374466896057,
"learning_rate": 8.478099498839273e-05,
"loss": 1.5664,
"step": 622
},
{
"epoch": 0.6627659574468086,
"grad_norm": 0.16877172887325287,
"learning_rate": 8.444701063029336e-05,
"loss": 1.7829,
"step": 623
},
{
"epoch": 0.6638297872340425,
"grad_norm": 0.18060240149497986,
"learning_rate": 8.411362649907808e-05,
"loss": 1.7298,
"step": 624
},
{
"epoch": 0.6648936170212766,
"grad_norm": 0.14296147227287292,
"learning_rate": 8.378084636656303e-05,
"loss": 1.5796,
"step": 625
},
{
"epoch": 0.6659574468085107,
"grad_norm": 0.18668703734874725,
"learning_rate": 8.344867399773072e-05,
"loss": 1.6658,
"step": 626
},
{
"epoch": 0.6670212765957447,
"grad_norm": 0.1682588756084442,
"learning_rate": 8.311711315068771e-05,
"loss": 1.7215,
"step": 627
},
{
"epoch": 0.6680851063829787,
"grad_norm": 0.19542992115020752,
"learning_rate": 8.278616757662199e-05,
"loss": 1.8014,
"step": 628
},
{
"epoch": 0.6691489361702128,
"grad_norm": 0.26618170738220215,
"learning_rate": 8.245584101976042e-05,
"loss": 1.8096,
"step": 629
},
{
"epoch": 0.6702127659574468,
"grad_norm": 0.17876097559928894,
"learning_rate": 8.212613721732655e-05,
"loss": 1.9594,
"step": 630
},
{
"epoch": 0.6712765957446809,
"grad_norm": 0.22028006613254547,
"learning_rate": 8.179705989949823e-05,
"loss": 1.8811,
"step": 631
},
{
"epoch": 0.6723404255319149,
"grad_norm": 0.20614123344421387,
"learning_rate": 8.146861278936545e-05,
"loss": 1.7899,
"step": 632
},
{
"epoch": 0.6734042553191489,
"grad_norm": 0.17731155455112457,
"learning_rate": 8.114079960288819e-05,
"loss": 1.6192,
"step": 633
},
{
"epoch": 0.674468085106383,
"grad_norm": 0.17782393097877502,
"learning_rate": 8.081362404885442e-05,
"loss": 1.8752,
"step": 634
},
{
"epoch": 0.675531914893617,
"grad_norm": 0.15950807929039001,
"learning_rate": 8.048708982883804e-05,
"loss": 1.7752,
"step": 635
},
{
"epoch": 0.676595744680851,
"grad_norm": 0.17458008229732513,
"learning_rate": 8.016120063715716e-05,
"loss": 1.821,
"step": 636
},
{
"epoch": 0.6776595744680851,
"grad_norm": 0.1835324466228485,
"learning_rate": 7.983596016083213e-05,
"loss": 1.652,
"step": 637
},
{
"epoch": 0.6787234042553192,
"grad_norm": 0.19433072209358215,
"learning_rate": 7.9511372079544e-05,
"loss": 1.8008,
"step": 638
},
{
"epoch": 0.6797872340425531,
"grad_norm": 0.1701725274324417,
"learning_rate": 7.918744006559278e-05,
"loss": 1.6133,
"step": 639
},
{
"epoch": 0.6808510638297872,
"grad_norm": 0.20287151634693146,
"learning_rate": 7.886416778385588e-05,
"loss": 1.9726,
"step": 640
},
{
"epoch": 0.6819148936170213,
"grad_norm": 0.21464970707893372,
"learning_rate": 7.85415588917467e-05,
"loss": 1.3563,
"step": 641
},
{
"epoch": 0.6829787234042554,
"grad_norm": 0.1623314917087555,
"learning_rate": 7.821961703917325e-05,
"loss": 1.659,
"step": 642
},
{
"epoch": 0.6840425531914893,
"grad_norm": 0.19890430569648743,
"learning_rate": 7.78983458684968e-05,
"loss": 1.7816,
"step": 643
},
{
"epoch": 0.6851063829787234,
"grad_norm": 0.16741237044334412,
"learning_rate": 7.757774901449075e-05,
"loss": 1.375,
"step": 644
},
{
"epoch": 0.6861702127659575,
"grad_norm": 0.20035420358181,
"learning_rate": 7.725783010429952e-05,
"loss": 1.5914,
"step": 645
},
{
"epoch": 0.6872340425531915,
"grad_norm": 0.19448257982730865,
"learning_rate": 7.693859275739726e-05,
"loss": 1.7364,
"step": 646
},
{
"epoch": 0.6882978723404255,
"grad_norm": 0.17156560719013214,
"learning_rate": 7.66200405855473e-05,
"loss": 1.7058,
"step": 647
},
{
"epoch": 0.6893617021276596,
"grad_norm": 0.17527243494987488,
"learning_rate": 7.630217719276104e-05,
"loss": 1.7223,
"step": 648
},
{
"epoch": 0.6904255319148936,
"grad_norm": 0.1525346040725708,
"learning_rate": 7.598500617525722e-05,
"loss": 1.6461,
"step": 649
},
{
"epoch": 0.6914893617021277,
"grad_norm": 0.14523084461688995,
"learning_rate": 7.566853112142114e-05,
"loss": 1.9347,
"step": 650
},
{
"epoch": 0.6925531914893617,
"grad_norm": 0.14614152908325195,
"learning_rate": 7.53527556117643e-05,
"loss": 1.6018,
"step": 651
},
{
"epoch": 0.6936170212765957,
"grad_norm": 0.16998977959156036,
"learning_rate": 7.503768321888368e-05,
"loss": 1.4355,
"step": 652
},
{
"epoch": 0.6946808510638298,
"grad_norm": 0.17113342881202698,
"learning_rate": 7.472331750742142e-05,
"loss": 1.8263,
"step": 653
},
{
"epoch": 0.6957446808510638,
"grad_norm": 0.17434722185134888,
"learning_rate": 7.440966203402442e-05,
"loss": 1.6064,
"step": 654
},
{
"epoch": 0.6968085106382979,
"grad_norm": 0.17484360933303833,
"learning_rate": 7.409672034730416e-05,
"loss": 1.7635,
"step": 655
},
{
"epoch": 0.6978723404255319,
"grad_norm": 0.1816336214542389,
"learning_rate": 7.378449598779658e-05,
"loss": 1.749,
"step": 656
},
{
"epoch": 0.698936170212766,
"grad_norm": 0.20657788217067719,
"learning_rate": 7.347299248792196e-05,
"loss": 1.5291,
"step": 657
},
{
"epoch": 0.7,
"grad_norm": 0.17004646360874176,
"learning_rate": 7.316221337194484e-05,
"loss": 1.901,
"step": 658
},
{
"epoch": 0.701063829787234,
"grad_norm": 0.1709367334842682,
"learning_rate": 7.285216215593449e-05,
"loss": 1.6908,
"step": 659
},
{
"epoch": 0.7021276595744681,
"grad_norm": 0.1813773363828659,
"learning_rate": 7.254284234772479e-05,
"loss": 1.8814,
"step": 660
},
{
"epoch": 0.7031914893617022,
"grad_norm": 0.21664945781230927,
"learning_rate": 7.22342574468747e-05,
"loss": 1.7577,
"step": 661
},
{
"epoch": 0.7042553191489361,
"grad_norm": 0.2223803550004959,
"learning_rate": 7.192641094462866e-05,
"loss": 1.729,
"step": 662
},
{
"epoch": 0.7053191489361702,
"grad_norm": 0.17629499733448029,
"learning_rate": 7.16193063238771e-05,
"loss": 1.8751,
"step": 663
},
{
"epoch": 0.7063829787234043,
"grad_norm": 0.17627973854541779,
"learning_rate": 7.131294705911689e-05,
"loss": 1.5773,
"step": 664
},
{
"epoch": 0.7074468085106383,
"grad_norm": 0.20003056526184082,
"learning_rate": 7.100733661641225e-05,
"loss": 2.0479,
"step": 665
},
{
"epoch": 0.7085106382978723,
"grad_norm": 0.1355198323726654,
"learning_rate": 7.070247845335545e-05,
"loss": 1.7533,
"step": 666
},
{
"epoch": 0.7095744680851064,
"grad_norm": 0.2013373225927353,
"learning_rate": 7.039837601902757e-05,
"loss": 1.9195,
"step": 667
},
{
"epoch": 0.7106382978723405,
"grad_norm": 0.15443074703216553,
"learning_rate": 7.009503275395975e-05,
"loss": 2.0914,
"step": 668
},
{
"epoch": 0.7117021276595744,
"grad_norm": 0.16669104993343353,
"learning_rate": 6.979245209009389e-05,
"loss": 1.7793,
"step": 669
},
{
"epoch": 0.7127659574468085,
"grad_norm": 0.158504918217659,
"learning_rate": 6.949063745074415e-05,
"loss": 1.7443,
"step": 670
},
{
"epoch": 0.7138297872340426,
"grad_norm": 0.16669638454914093,
"learning_rate": 6.918959225055813e-05,
"loss": 1.3766,
"step": 671
},
{
"epoch": 0.7148936170212766,
"grad_norm": 0.16765549778938293,
"learning_rate": 6.888931989547816e-05,
"loss": 1.7926,
"step": 672
},
{
"epoch": 0.7159574468085106,
"grad_norm": 0.19317655265331268,
"learning_rate": 6.858982378270278e-05,
"loss": 1.7287,
"step": 673
},
{
"epoch": 0.7170212765957447,
"grad_norm": 0.14376698434352875,
"learning_rate": 6.829110730064844e-05,
"loss": 1.4643,
"step": 674
},
{
"epoch": 0.7180851063829787,
"grad_norm": 0.20092540979385376,
"learning_rate": 6.799317382891095e-05,
"loss": 1.8243,
"step": 675
},
{
"epoch": 0.7191489361702128,
"grad_norm": 0.22864454984664917,
"learning_rate": 6.769602673822742e-05,
"loss": 1.2684,
"step": 676
},
{
"epoch": 0.7202127659574468,
"grad_norm": 0.17580509185791016,
"learning_rate": 6.73996693904381e-05,
"loss": 1.7815,
"step": 677
},
{
"epoch": 0.7212765957446808,
"grad_norm": 0.19205164909362793,
"learning_rate": 6.710410513844816e-05,
"loss": 1.9747,
"step": 678
},
{
"epoch": 0.7223404255319149,
"grad_norm": 0.17647920548915863,
"learning_rate": 6.68093373261901e-05,
"loss": 1.5295,
"step": 679
},
{
"epoch": 0.723404255319149,
"grad_norm": 0.15995529294013977,
"learning_rate": 6.65153692885855e-05,
"loss": 1.5583,
"step": 680
},
{
"epoch": 0.7244680851063829,
"grad_norm": 0.18296481668949127,
"learning_rate": 6.622220435150772e-05,
"loss": 1.9198,
"step": 681
},
{
"epoch": 0.725531914893617,
"grad_norm": 0.15332357585430145,
"learning_rate": 6.592984583174394e-05,
"loss": 1.6978,
"step": 682
},
{
"epoch": 0.7265957446808511,
"grad_norm": 0.1648200899362564,
"learning_rate": 6.563829703695781e-05,
"loss": 1.6717,
"step": 683
},
{
"epoch": 0.7276595744680852,
"grad_norm": 0.16624633967876434,
"learning_rate": 6.534756126565194e-05,
"loss": 1.6043,
"step": 684
},
{
"epoch": 0.7287234042553191,
"grad_norm": 0.238071009516716,
"learning_rate": 6.505764180713065e-05,
"loss": 1.8476,
"step": 685
},
{
"epoch": 0.7297872340425532,
"grad_norm": 0.15699748694896698,
"learning_rate": 6.476854194146265e-05,
"loss": 1.8375,
"step": 686
},
{
"epoch": 0.7308510638297873,
"grad_norm": 0.1671568751335144,
"learning_rate": 6.448026493944411e-05,
"loss": 1.7371,
"step": 687
},
{
"epoch": 0.7319148936170212,
"grad_norm": 0.17969168722629547,
"learning_rate": 6.419281406256153e-05,
"loss": 1.8686,
"step": 688
},
{
"epoch": 0.7329787234042553,
"grad_norm": 0.17477388679981232,
"learning_rate": 6.390619256295471e-05,
"loss": 2.0333,
"step": 689
},
{
"epoch": 0.7340425531914894,
"grad_norm": 0.15487664937973022,
"learning_rate": 6.362040368338032e-05,
"loss": 1.8519,
"step": 690
},
{
"epoch": 0.7351063829787234,
"grad_norm": 0.1683463156223297,
"learning_rate": 6.33354506571748e-05,
"loss": 1.844,
"step": 691
},
{
"epoch": 0.7361702127659574,
"grad_norm": 0.19052305817604065,
"learning_rate": 6.305133670821814e-05,
"loss": 1.9067,
"step": 692
},
{
"epoch": 0.7372340425531915,
"grad_norm": 0.15490886569023132,
"learning_rate": 6.276806505089713e-05,
"loss": 2.0279,
"step": 693
},
{
"epoch": 0.7382978723404255,
"grad_norm": 0.1947750747203827,
"learning_rate": 6.248563889006917e-05,
"loss": 1.6706,
"step": 694
},
{
"epoch": 0.7393617021276596,
"grad_norm": 0.1630692034959793,
"learning_rate": 6.220406142102589e-05,
"loss": 2.0676,
"step": 695
},
{
"epoch": 0.7404255319148936,
"grad_norm": 0.15723346173763275,
"learning_rate": 6.192333582945712e-05,
"loss": 1.7353,
"step": 696
},
{
"epoch": 0.7414893617021276,
"grad_norm": 0.23085269331932068,
"learning_rate": 6.164346529141472e-05,
"loss": 1.4591,
"step": 697
},
{
"epoch": 0.7425531914893617,
"grad_norm": 0.19693605601787567,
"learning_rate": 6.136445297327677e-05,
"loss": 1.8418,
"step": 698
},
{
"epoch": 0.7436170212765958,
"grad_norm": 0.15586483478546143,
"learning_rate": 6.108630203171164e-05,
"loss": 2.1864,
"step": 699
},
{
"epoch": 0.7446808510638298,
"grad_norm": 0.21545270085334778,
"learning_rate": 6.080901561364238e-05,
"loss": 1.3905,
"step": 700
},
{
"epoch": 0.7457446808510638,
"grad_norm": 0.1876390278339386,
"learning_rate": 6.0532596856210935e-05,
"loss": 1.7214,
"step": 701
},
{
"epoch": 0.7468085106382979,
"grad_norm": 0.18684272468090057,
"learning_rate": 6.02570488867429e-05,
"loss": 1.9218,
"step": 702
},
{
"epoch": 0.747872340425532,
"grad_norm": 0.1691771149635315,
"learning_rate": 5.9982374822712e-05,
"loss": 1.7906,
"step": 703
},
{
"epoch": 0.7489361702127659,
"grad_norm": 0.1723630279302597,
"learning_rate": 5.970857777170482e-05,
"loss": 1.7991,
"step": 704
},
{
"epoch": 0.75,
"grad_norm": 0.20111820101737976,
"learning_rate": 5.94356608313856e-05,
"loss": 1.9285,
"step": 705
},
{
"epoch": 0.7510638297872341,
"grad_norm": 0.19463780522346497,
"learning_rate": 5.9163627089461314e-05,
"loss": 1.6628,
"step": 706
},
{
"epoch": 0.752127659574468,
"grad_norm": 0.1662537008523941,
"learning_rate": 5.889247962364669e-05,
"loss": 1.6007,
"step": 707
},
{
"epoch": 0.7531914893617021,
"grad_norm": 0.1821126490831375,
"learning_rate": 5.862222150162932e-05,
"loss": 1.7381,
"step": 708
}
],
"logging_steps": 1,
"max_steps": 944,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 236,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.6432043262581146e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}