{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.056627092618254,
  "eval_steps": 42,
  "global_step": 378,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.00014980712332871428,
      "eval_loss": 1.160882592201233,
      "eval_runtime": 311.16,
      "eval_samples_per_second": 36.133,
      "eval_steps_per_second": 4.519,
      "step": 1
    },
    {
      "epoch": 0.00044942136998614283,
      "grad_norm": 0.12129199504852295,
      "learning_rate": 1.5e-05,
      "loss": 1.1193,
      "step": 3
    },
    {
      "epoch": 0.0008988427399722857,
      "grad_norm": 0.11971050500869751,
      "learning_rate": 3e-05,
      "loss": 1.103,
      "step": 6
    },
    {
      "epoch": 0.0013482641099584285,
      "grad_norm": 0.10862895101308823,
      "learning_rate": 4.5e-05,
      "loss": 1.0659,
      "step": 9
    },
    {
      "epoch": 0.0017976854799445713,
      "grad_norm": 0.11977853626012802,
      "learning_rate": 4.9997944716957985e-05,
      "loss": 1.0914,
      "step": 12
    },
    {
      "epoch": 0.0022471068499307144,
      "grad_norm": 0.12554779648780823,
      "learning_rate": 4.99871554050172e-05,
      "loss": 1.1725,
      "step": 15
    },
    {
      "epoch": 0.002696528219916857,
      "grad_norm": 0.13287784159183502,
      "learning_rate": 4.996712222958461e-05,
      "loss": 1.1145,
      "step": 18
    },
    {
      "epoch": 0.003145949589903,
      "grad_norm": 0.1167389452457428,
      "learning_rate": 4.993785260182552e-05,
      "loss": 1.0956,
      "step": 21
    },
    {
      "epoch": 0.0035953709598891426,
      "grad_norm": 0.11548212915658951,
      "learning_rate": 4.989935734988098e-05,
      "loss": 1.12,
      "step": 24
    },
    {
      "epoch": 0.004044792329875285,
      "grad_norm": 0.12251826375722885,
      "learning_rate": 4.9851650714862006e-05,
      "loss": 1.0895,
      "step": 27
    },
    {
      "epoch": 0.004494213699861429,
      "grad_norm": 0.12304849177598953,
      "learning_rate": 4.979475034558115e-05,
      "loss": 1.1302,
      "step": 30
    },
    {
      "epoch": 0.004943635069847571,
      "grad_norm": 0.12594439089298248,
      "learning_rate": 4.9728677292023405e-05,
      "loss": 1.0825,
      "step": 33
    },
    {
      "epoch": 0.005393056439833714,
      "grad_norm": 0.11802031844854355,
      "learning_rate": 4.965345599755887e-05,
      "loss": 1.0949,
      "step": 36
    },
    {
      "epoch": 0.005842477809819857,
      "grad_norm": 0.12052260339260101,
      "learning_rate": 4.95691142899001e-05,
      "loss": 1.2106,
      "step": 39
    },
    {
      "epoch": 0.006291899179806,
      "grad_norm": 0.14151117205619812,
      "learning_rate": 4.9475683370807326e-05,
      "loss": 1.0855,
      "step": 42
    },
    {
      "epoch": 0.006291899179806,
      "eval_loss": 1.1297273635864258,
      "eval_runtime": 313.2801,
      "eval_samples_per_second": 35.888,
      "eval_steps_per_second": 4.488,
      "step": 42
    },
    {
      "epoch": 0.006741320549792143,
      "grad_norm": 0.13759616017341614,
      "learning_rate": 4.937319780454559e-05,
      "loss": 1.1133,
      "step": 45
    },
    {
      "epoch": 0.007190741919778285,
      "grad_norm": 0.14887671172618866,
      "learning_rate": 4.926169550509787e-05,
      "loss": 1.1339,
      "step": 48
    },
    {
      "epoch": 0.007640163289764428,
      "grad_norm": 0.13235358893871307,
      "learning_rate": 4.914121772213898e-05,
      "loss": 1.0057,
      "step": 51
    },
    {
      "epoch": 0.00808958465975057,
      "grad_norm": 0.14951473474502563,
      "learning_rate": 4.9011809025775486e-05,
      "loss": 1.0919,
      "step": 54
    },
    {
      "epoch": 0.008539006029736713,
      "grad_norm": 0.1409119963645935,
      "learning_rate": 4.887351729005726e-05,
      "loss": 1.143,
      "step": 57
    },
    {
      "epoch": 0.008988427399722857,
      "grad_norm": 0.13198496401309967,
      "learning_rate": 4.8726393675266716e-05,
      "loss": 1.0853,
      "step": 60
    },
    {
      "epoch": 0.009437848769709,
      "grad_norm": 0.15319645404815674,
      "learning_rate": 4.8570492608992325e-05,
      "loss": 1.0534,
      "step": 63
    },
    {
      "epoch": 0.009887270139695143,
      "grad_norm": 0.1532219797372818,
      "learning_rate": 4.8405871765993433e-05,
      "loss": 1.1016,
      "step": 66
    },
    {
      "epoch": 0.010336691509681285,
      "grad_norm": 0.14198264479637146,
      "learning_rate": 4.82325920468638e-05,
      "loss": 1.029,
      "step": 69
    },
    {
      "epoch": 0.010786112879667428,
      "grad_norm": 0.12415461987257004,
      "learning_rate": 4.805071755550177e-05,
      "loss": 1.1565,
      "step": 72
    },
    {
      "epoch": 0.01123553424965357,
      "grad_norm": 0.1350441873073578,
      "learning_rate": 4.7860315575395316e-05,
      "loss": 1.1276,
      "step": 75
    },
    {
      "epoch": 0.011684955619639713,
      "grad_norm": 0.12075574696063995,
      "learning_rate": 4.766145654473095e-05,
      "loss": 1.0753,
      "step": 78
    },
    {
      "epoch": 0.012134376989625857,
      "grad_norm": 0.11550460755825043,
      "learning_rate": 4.745421403033548e-05,
      "loss": 1.0078,
      "step": 81
    },
    {
      "epoch": 0.012583798359612,
      "grad_norm": 0.12455905228853226,
      "learning_rate": 4.72386647004603e-05,
      "loss": 1.0373,
      "step": 84
    },
    {
      "epoch": 0.012583798359612,
      "eval_loss": 1.1157194375991821,
      "eval_runtime": 312.8234,
      "eval_samples_per_second": 35.94,
      "eval_steps_per_second": 4.495,
      "step": 84
    },
    {
      "epoch": 0.013033219729598143,
      "grad_norm": 0.12993893027305603,
      "learning_rate": 4.701488829641845e-05,
      "loss": 1.1313,
      "step": 87
    },
    {
      "epoch": 0.013482641099584285,
      "grad_norm": 0.18705669045448303,
      "learning_rate": 4.678296760308474e-05,
      "loss": 1.2752,
      "step": 90
    },
    {
      "epoch": 0.013932062469570428,
      "grad_norm": 0.14135660231113434,
      "learning_rate": 4.6542988418269876e-05,
      "loss": 1.1628,
      "step": 93
    },
    {
      "epoch": 0.01438148383955657,
      "grad_norm": 0.11707276850938797,
      "learning_rate": 4.629503952098011e-05,
      "loss": 1.0801,
      "step": 96
    },
    {
      "epoch": 0.014830905209542713,
      "grad_norm": 0.14167702198028564,
      "learning_rate": 4.6039212638573833e-05,
      "loss": 1.0452,
      "step": 99
    },
    {
      "epoch": 0.015280326579528856,
      "grad_norm": 0.15983764827251434,
      "learning_rate": 4.5775602412827604e-05,
      "loss": 1.1844,
      "step": 102
    },
    {
      "epoch": 0.015729747949515,
      "grad_norm": 0.12750358879566193,
      "learning_rate": 4.55043063649239e-05,
      "loss": 1.0877,
      "step": 105
    },
    {
      "epoch": 0.01617916931950114,
      "grad_norm": 0.1487520933151245,
      "learning_rate": 4.522542485937369e-05,
      "loss": 1.1373,
      "step": 108
    },
    {
      "epoch": 0.016628590689487285,
      "grad_norm": 0.16089919209480286,
      "learning_rate": 4.493906106688712e-05,
      "loss": 1.1806,
      "step": 111
    },
    {
      "epoch": 0.017078012059473426,
      "grad_norm": 0.158865287899971,
      "learning_rate": 4.4645320926206064e-05,
      "loss": 1.0689,
      "step": 114
    },
    {
      "epoch": 0.01752743342945957,
      "grad_norm": 0.16464003920555115,
      "learning_rate": 4.434431310491267e-05,
      "loss": 1.1319,
      "step": 117
    },
    {
      "epoch": 0.017976854799445715,
      "grad_norm": 0.12468679249286652,
      "learning_rate": 4.4036148959228365e-05,
      "loss": 1.0237,
      "step": 120
    },
    {
      "epoch": 0.018426276169431856,
      "grad_norm": 0.14475342631340027,
      "learning_rate": 4.372094249281821e-05,
      "loss": 1.1352,
      "step": 123
    },
    {
      "epoch": 0.018875697539418,
      "grad_norm": 0.1386341154575348,
      "learning_rate": 4.3398810314615876e-05,
      "loss": 1.177,
      "step": 126
    },
    {
      "epoch": 0.018875697539418,
      "eval_loss": 1.1068843603134155,
      "eval_runtime": 312.5635,
      "eval_samples_per_second": 35.97,
      "eval_steps_per_second": 4.498,
      "step": 126
    },
    {
      "epoch": 0.01932511890940414,
      "grad_norm": 0.13838274776935577,
      "learning_rate": 4.306987159568479e-05,
      "loss": 0.9996,
      "step": 129
    },
    {
      "epoch": 0.019774540279390285,
      "grad_norm": 0.14386090636253357,
      "learning_rate": 4.273424802513145e-05,
      "loss": 0.9947,
      "step": 132
    },
    {
      "epoch": 0.020223961649376426,
      "grad_norm": 0.14473505318164825,
      "learning_rate": 4.239206376508717e-05,
      "loss": 1.1779,
      "step": 135
    },
    {
      "epoch": 0.02067338301936257,
      "grad_norm": 0.15975573658943176,
      "learning_rate": 4.204344540477499e-05,
      "loss": 1.2125,
      "step": 138
    },
    {
      "epoch": 0.021122804389348715,
      "grad_norm": 0.12576735019683838,
      "learning_rate": 4.16885219136787e-05,
      "loss": 1.1312,
      "step": 141
    },
    {
      "epoch": 0.021572225759334856,
      "grad_norm": 0.16916967928409576,
      "learning_rate": 4.132742459383122e-05,
      "loss": 1.0823,
      "step": 144
    },
    {
      "epoch": 0.022021647129321,
      "grad_norm": 0.17954471707344055,
      "learning_rate": 4.096028703124014e-05,
      "loss": 1.0728,
      "step": 147
    },
    {
      "epoch": 0.02247106849930714,
      "grad_norm": 0.11972087621688843,
      "learning_rate": 4.058724504646834e-05,
      "loss": 1.0886,
      "step": 150
    },
    {
      "epoch": 0.022920489869293285,
      "grad_norm": 0.14640676975250244,
      "learning_rate": 4.0208436644387834e-05,
      "loss": 1.0472,
      "step": 153
    },
    {
      "epoch": 0.023369911239279426,
      "grad_norm": 0.13775284588336945,
      "learning_rate": 3.982400196312564e-05,
      "loss": 1.0984,
      "step": 156
    },
    {
      "epoch": 0.02381933260926557,
      "grad_norm": 0.14184071123600006,
      "learning_rate": 3.943408322222049e-05,
      "loss": 1.1431,
      "step": 159
    },
    {
      "epoch": 0.024268753979251715,
      "grad_norm": 0.1685701608657837,
      "learning_rate": 3.903882467000937e-05,
      "loss": 1.1531,
      "step": 162
    },
    {
      "epoch": 0.024718175349237856,
      "grad_norm": 0.1468999981880188,
      "learning_rate": 3.8638372530263715e-05,
      "loss": 1.1069,
      "step": 165
    },
    {
      "epoch": 0.025167596719224,
      "grad_norm": 0.13278649747371674,
      "learning_rate": 3.823287494809469e-05,
      "loss": 1.1124,
      "step": 168
    },
    {
      "epoch": 0.025167596719224,
      "eval_loss": 1.1012423038482666,
      "eval_runtime": 312.1008,
      "eval_samples_per_second": 36.024,
      "eval_steps_per_second": 4.505,
      "step": 168
    },
    {
      "epoch": 0.02561701808921014,
      "grad_norm": 0.180728018283844,
      "learning_rate": 3.782248193514766e-05,
      "loss": 1.0867,
      "step": 171
    },
    {
      "epoch": 0.026066439459196285,
      "grad_norm": 0.15069565176963806,
      "learning_rate": 3.740734531410626e-05,
      "loss": 1.0624,
      "step": 174
    },
    {
      "epoch": 0.026515860829182426,
      "grad_norm": 0.13892242312431335,
      "learning_rate": 3.698761866252635e-05,
      "loss": 1.0351,
      "step": 177
    },
    {
      "epoch": 0.02696528219916857,
      "grad_norm": 0.1399199515581131,
      "learning_rate": 3.656345725602089e-05,
      "loss": 1.1609,
      "step": 180
    },
    {
      "epoch": 0.027414703569154715,
      "grad_norm": 0.14930486679077148,
      "learning_rate": 3.6135018010816477e-05,
      "loss": 1.1117,
      "step": 183
    },
    {
      "epoch": 0.027864124939140856,
      "grad_norm": 0.15556196868419647,
      "learning_rate": 3.570245942570315e-05,
      "loss": 1.1169,
      "step": 186
    },
    {
      "epoch": 0.028313546309127,
      "grad_norm": 0.17272590100765228,
      "learning_rate": 3.526594152339845e-05,
      "loss": 1.115,
      "step": 189
    },
    {
      "epoch": 0.02876296767911314,
      "grad_norm": 0.17533355951309204,
      "learning_rate": 3.4825625791348096e-05,
      "loss": 1.1298,
      "step": 192
    },
    {
      "epoch": 0.029212389049099285,
      "grad_norm": 0.14778710901737213,
      "learning_rate": 3.438167512198436e-05,
      "loss": 1.1183,
      "step": 195
    },
    {
      "epoch": 0.029661810419085426,
      "grad_norm": 0.14693984389305115,
      "learning_rate": 3.393425375246503e-05,
      "loss": 1.0647,
      "step": 198
    },
    {
      "epoch": 0.03011123178907157,
      "grad_norm": 0.14994005858898163,
      "learning_rate": 3.348352720391469e-05,
      "loss": 1.0008,
      "step": 201
    },
    {
      "epoch": 0.03056065315905771,
      "grad_norm": 0.1611510068178177,
      "learning_rate": 3.3029662220191144e-05,
      "loss": 1.094,
      "step": 204
    },
    {
      "epoch": 0.031010074529043856,
      "grad_norm": 0.19615799188613892,
      "learning_rate": 3.2572826706199305e-05,
      "loss": 1.051,
      "step": 207
    },
    {
      "epoch": 0.03145949589903,
      "grad_norm": 0.15789468586444855,
      "learning_rate": 3.211318966577581e-05,
      "loss": 1.0302,
      "step": 210
    },
    {
      "epoch": 0.03145949589903,
      "eval_loss": 1.0971506834030151,
      "eval_runtime": 312.684,
      "eval_samples_per_second": 35.956,
      "eval_steps_per_second": 4.497,
      "step": 210
    },
    {
      "epoch": 0.031908917269016145,
      "grad_norm": 0.15568239986896515,
      "learning_rate": 3.165092113916688e-05,
      "loss": 1.12,
      "step": 213
    },
    {
      "epoch": 0.03235833863900228,
      "grad_norm": 0.16250278055667877,
      "learning_rate": 3.118619214012286e-05,
      "loss": 1.0651,
      "step": 216
    },
    {
      "epoch": 0.032807760008988426,
      "grad_norm": 0.16043943166732788,
      "learning_rate": 3.071917459263264e-05,
      "loss": 1.1254,
      "step": 219
    },
    {
      "epoch": 0.03325718137897457,
      "grad_norm": 0.19213679432868958,
      "learning_rate": 3.0250041267321232e-05,
      "loss": 1.1082,
      "step": 222
    },
    {
      "epoch": 0.033706602748960715,
      "grad_norm": 0.17149952054023743,
      "learning_rate": 2.9778965717534313e-05,
      "loss": 1.0345,
      "step": 225
    },
    {
      "epoch": 0.03415602411894685,
      "grad_norm": 0.19161058962345123,
      "learning_rate": 2.9306122215132976e-05,
      "loss": 1.2631,
      "step": 228
    },
    {
      "epoch": 0.034605445488933,
      "grad_norm": 0.12983594834804535,
      "learning_rate": 2.8831685686022897e-05,
      "loss": 1.0125,
      "step": 231
    },
    {
      "epoch": 0.03505486685891914,
      "grad_norm": 0.1923818439245224,
      "learning_rate": 2.8355831645441388e-05,
      "loss": 1.1152,
      "step": 234
    },
    {
      "epoch": 0.035504288228905286,
      "grad_norm": 0.17539535462856293,
      "learning_rate": 2.787873613302649e-05,
      "loss": 1.0698,
      "step": 237
    },
    {
      "epoch": 0.03595370959889143,
      "grad_norm": 0.20096057653427124,
      "learning_rate": 2.7400575647692046e-05,
      "loss": 1.0742,
      "step": 240
    },
    {
      "epoch": 0.03640313096887757,
      "grad_norm": 0.1583949774503708,
      "learning_rate": 2.692152708233292e-05,
      "loss": 1.0255,
      "step": 243
    },
    {
      "epoch": 0.03685255233886371,
      "grad_norm": 0.13673090934753418,
      "learning_rate": 2.6441767658384366e-05,
      "loss": 1.1167,
      "step": 246
    },
    {
      "epoch": 0.037301973708849856,
      "grad_norm": 0.15132947266101837,
      "learning_rate": 2.596147486025996e-05,
      "loss": 1.053,
      "step": 249
    },
    {
      "epoch": 0.037751395078836,
      "grad_norm": 0.1649513989686966,
      "learning_rate": 2.5480826369692178e-05,
      "loss": 1.1041,
      "step": 252
    },
    {
      "epoch": 0.037751395078836,
      "eval_loss": 1.0942788124084473,
      "eval_runtime": 312.947,
      "eval_samples_per_second": 35.926,
      "eval_steps_per_second": 4.493,
      "step": 252
    },
    {
      "epoch": 0.038200816448822145,
      "grad_norm": 0.19113826751708984,
      "learning_rate": 2.5e-05,
      "loss": 1.0117,
      "step": 255
    },
    {
      "epoch": 0.03865023781880828,
      "grad_norm": 0.17212168872356415,
      "learning_rate": 2.4519173630307825e-05,
      "loss": 1.136,
      "step": 258
    },
    {
      "epoch": 0.039099659188794426,
      "grad_norm": 0.16413848102092743,
      "learning_rate": 2.403852513974004e-05,
      "loss": 1.0806,
      "step": 261
    },
    {
      "epoch": 0.03954908055878057,
      "grad_norm": 0.15564818680286407,
      "learning_rate": 2.3558232341615643e-05,
      "loss": 1.0233,
      "step": 264
    },
    {
      "epoch": 0.039998501928766715,
      "grad_norm": 0.15014681220054626,
      "learning_rate": 2.3078472917667092e-05,
      "loss": 1.09,
      "step": 267
    },
    {
      "epoch": 0.04044792329875285,
      "grad_norm": 0.1784486323595047,
      "learning_rate": 2.2599424352307957e-05,
      "loss": 1.0867,
      "step": 270
    },
    {
      "epoch": 0.040897344668739,
      "grad_norm": 0.1629609763622284,
      "learning_rate": 2.212126386697352e-05,
      "loss": 1.0916,
      "step": 273
    },
    {
      "epoch": 0.04134676603872514,
      "grad_norm": 0.1595894992351532,
      "learning_rate": 2.164416835455862e-05,
      "loss": 1.0537,
      "step": 276
    },
    {
      "epoch": 0.041796187408711286,
      "grad_norm": 0.16272102296352386,
      "learning_rate": 2.11683143139771e-05,
      "loss": 1.1907,
      "step": 279
    },
    {
      "epoch": 0.04224560877869743,
      "grad_norm": 0.15896858274936676,
      "learning_rate": 2.069387778486703e-05,
      "loss": 1.0492,
      "step": 282
    },
    {
      "epoch": 0.04269503014868357,
      "grad_norm": 0.18164744973182678,
      "learning_rate": 2.02210342824657e-05,
      "loss": 1.064,
      "step": 285
    },
    {
      "epoch": 0.04314445151866971,
      "grad_norm": 0.17921298742294312,
      "learning_rate": 1.9749958732678767e-05,
      "loss": 1.1456,
      "step": 288
    },
    {
      "epoch": 0.043593872888655856,
      "grad_norm": 0.14521931111812592,
      "learning_rate": 1.928082540736737e-05,
      "loss": 1.1073,
      "step": 291
    },
    {
      "epoch": 0.044043294258642,
      "grad_norm": 0.15362174808979034,
      "learning_rate": 1.8813807859877147e-05,
      "loss": 1.0171,
      "step": 294
    },
    {
      "epoch": 0.044043294258642,
      "eval_loss": 1.0923734903335571,
      "eval_runtime": 312.595,
      "eval_samples_per_second": 35.967,
      "eval_steps_per_second": 4.498,
      "step": 294
    },
    {
      "epoch": 0.044492715628628145,
      "grad_norm": 0.18710507452487946,
      "learning_rate": 1.8349078860833123e-05,
      "loss": 1.0978,
      "step": 297
    },
    {
      "epoch": 0.04494213699861428,
      "grad_norm": 0.1569843292236328,
      "learning_rate": 1.7886810334224192e-05,
      "loss": 1.0783,
      "step": 300
    },
    {
      "epoch": 0.045391558368600426,
      "grad_norm": 0.14208896458148956,
      "learning_rate": 1.74271732938007e-05,
      "loss": 1.0512,
      "step": 303
    },
    {
      "epoch": 0.04584097973858657,
      "grad_norm": 0.16557104885578156,
      "learning_rate": 1.6970337779808862e-05,
      "loss": 1.143,
      "step": 306
    },
    {
      "epoch": 0.046290401108572715,
      "grad_norm": 0.17679089307785034,
      "learning_rate": 1.6516472796085315e-05,
      "loss": 1.0428,
      "step": 309
    },
    {
      "epoch": 0.04673982247855885,
      "grad_norm": 0.21449750661849976,
      "learning_rate": 1.6065746247534984e-05,
      "loss": 1.0541,
      "step": 312
    },
    {
      "epoch": 0.047189243848545,
      "grad_norm": 0.19140197336673737,
      "learning_rate": 1.561832487801565e-05,
      "loss": 1.0943,
      "step": 315
    },
    {
      "epoch": 0.04763866521853114,
      "grad_norm": 0.21146325767040253,
      "learning_rate": 1.5174374208651912e-05,
      "loss": 1.1286,
      "step": 318
    },
    {
      "epoch": 0.048088086588517286,
      "grad_norm": 0.16576792299747467,
      "learning_rate": 1.4734058476601553e-05,
      "loss": 1.1543,
      "step": 321
    },
    {
      "epoch": 0.04853750795850343,
      "grad_norm": 0.17730730772018433,
      "learning_rate": 1.4297540574296869e-05,
      "loss": 1.0809,
      "step": 324
    },
    {
      "epoch": 0.04898692932848957,
      "grad_norm": 0.16303078830242157,
      "learning_rate": 1.386498198918352e-05,
      "loss": 1.0781,
      "step": 327
    },
    {
      "epoch": 0.04943635069847571,
      "grad_norm": 0.15803277492523193,
      "learning_rate": 1.3436542743979125e-05,
      "loss": 1.1138,
      "step": 330
    },
    {
      "epoch": 0.049885772068461856,
      "grad_norm": 0.161447212100029,
      "learning_rate": 1.3012381337473656e-05,
      "loss": 1.0087,
      "step": 333
    },
    {
      "epoch": 0.050335193438448,
      "grad_norm": 0.14916126430034637,
      "learning_rate": 1.2592654685893757e-05,
      "loss": 1.0916,
      "step": 336
    },
    {
      "epoch": 0.050335193438448,
      "eval_loss": 1.0906805992126465,
      "eval_runtime": 312.9768,
      "eval_samples_per_second": 35.923,
      "eval_steps_per_second": 4.492,
      "step": 336
    },
    {
      "epoch": 0.05078461480843414,
      "grad_norm": 0.18624208867549896,
      "learning_rate": 1.217751806485235e-05,
      "loss": 1.1368,
      "step": 339
    },
    {
      "epoch": 0.05123403617842028,
      "grad_norm": 0.18036052584648132,
      "learning_rate": 1.1767125051905315e-05,
      "loss": 1.0682,
      "step": 342
    },
    {
      "epoch": 0.051683457548406427,
      "grad_norm": 0.17344728112220764,
      "learning_rate": 1.1361627469736285e-05,
      "loss": 1.0299,
      "step": 345
    },
    {
      "epoch": 0.05213287891839257,
      "grad_norm": 0.17225950956344604,
      "learning_rate": 1.096117532999063e-05,
      "loss": 1.0217,
      "step": 348
    },
    {
      "epoch": 0.052582300288378715,
      "grad_norm": 0.1770411878824234,
      "learning_rate": 1.0565916777779519e-05,
      "loss": 1.1027,
      "step": 351
    },
    {
      "epoch": 0.05303172165836485,
      "grad_norm": 0.15190331637859344,
      "learning_rate": 1.0175998036874356e-05,
      "loss": 1.0684,
      "step": 354
    },
    {
      "epoch": 0.053481143028351,
      "grad_norm": 0.20575201511383057,
      "learning_rate": 9.791563355612172e-06,
      "loss": 1.0359,
      "step": 357
    },
    {
      "epoch": 0.05393056439833714,
      "grad_norm": 0.2163867950439453,
      "learning_rate": 9.412754953531663e-06,
      "loss": 1.0943,
      "step": 360
    },
    {
      "epoch": 0.054379985768323286,
      "grad_norm": 0.1523335576057434,
      "learning_rate": 9.039712968759864e-06,
      "loss": 1.1357,
      "step": 363
    },
    {
      "epoch": 0.05482940713830943,
      "grad_norm": 0.1748288869857788,
      "learning_rate": 8.672575406168782e-06,
      "loss": 1.0745,
      "step": 366
    },
    {
      "epoch": 0.05527882850829557,
      "grad_norm": 0.16783830523490906,
      "learning_rate": 8.3114780863213e-06,
      "loss": 1.0432,
      "step": 369
    },
    {
      "epoch": 0.05572824987828171,
      "grad_norm": 0.17801252007484436,
      "learning_rate": 7.956554595225016e-06,
      "loss": 1.0649,
      "step": 372
    },
    {
      "epoch": 0.056177671248267856,
      "grad_norm": 0.17944374680519104,
      "learning_rate": 7.607936234912841e-06,
      "loss": 1.0862,
      "step": 375
    },
    {
      "epoch": 0.056627092618254,
      "grad_norm": 0.19096186757087708,
      "learning_rate": 7.265751974868554e-06,
      "loss": 1.0477,
      "step": 378
    },
    {
      "epoch": 0.056627092618254,
      "eval_loss": 1.0896837711334229,
      "eval_runtime": 312.799,
      "eval_samples_per_second": 35.943,
      "eval_steps_per_second": 4.495,
      "step": 378
    }
  ],
  "logging_steps": 3,
  "max_steps": 500,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 42,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 1.2131417357234995e+17,
  "train_batch_size": 8,
  "trial_name": null,
  "trial_params": null
}