{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 14.679611650485437,
  "global_step": 630,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.02,
      "learning_rate": 2.631578947368421e-06,
      "loss": 1.6384,
      "step": 1
    },
    {
      "epoch": 0.05,
      "learning_rate": 5.263157894736842e-06,
      "loss": 1.677,
      "step": 2
    },
    {
      "epoch": 0.07,
      "learning_rate": 7.894736842105263e-06,
      "loss": 1.9569,
      "step": 3
    },
    {
      "epoch": 0.09,
      "learning_rate": 1.0526315789473684e-05,
      "loss": 1.8787,
      "step": 4
    },
    {
      "epoch": 0.12,
      "learning_rate": 1.3157894736842106e-05,
      "loss": 1.9318,
      "step": 5
    },
    {
      "epoch": 0.14,
      "learning_rate": 1.5789473684210526e-05,
      "loss": 2.1207,
      "step": 6
    },
    {
      "epoch": 0.16,
      "learning_rate": 1.8421052631578947e-05,
      "loss": 2.0227,
      "step": 7
    },
    {
      "epoch": 0.19,
      "learning_rate": 2.105263157894737e-05,
      "loss": 2.103,
      "step": 8
    },
    {
      "epoch": 0.21,
      "learning_rate": 2.368421052631579e-05,
      "loss": 2.2184,
      "step": 9
    },
    {
      "epoch": 0.23,
      "learning_rate": 2.6315789473684212e-05,
      "loss": 2.455,
      "step": 10
    },
    {
      "epoch": 0.26,
      "learning_rate": 2.8947368421052634e-05,
      "loss": 1.6635,
      "step": 11
    },
    {
      "epoch": 0.28,
      "learning_rate": 3.157894736842105e-05,
      "loss": 1.6389,
      "step": 12
    },
    {
      "epoch": 0.3,
      "learning_rate": 3.421052631578947e-05,
      "loss": 1.7803,
      "step": 13
    },
    {
      "epoch": 0.33,
      "learning_rate": 3.6842105263157895e-05,
      "loss": 1.8378,
      "step": 14
    },
    {
      "epoch": 0.35,
      "learning_rate": 3.9473684210526316e-05,
      "loss": 1.9676,
      "step": 15
    },
    {
      "epoch": 0.37,
      "learning_rate": 4.210526315789474e-05,
      "loss": 2.0158,
      "step": 16
    },
    {
      "epoch": 0.4,
      "learning_rate": 4.473684210526316e-05,
      "loss": 1.8251,
      "step": 17
    },
    {
      "epoch": 0.42,
      "learning_rate": 4.736842105263158e-05,
      "loss": 1.9671,
      "step": 18
    },
    {
      "epoch": 0.44,
      "learning_rate": 5e-05,
      "loss": 1.9491,
      "step": 19
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.2631578947368424e-05,
      "loss": 1.9589,
      "step": 20
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.526315789473685e-05,
      "loss": 1.6453,
      "step": 21
    },
    {
      "epoch": 0.51,
      "learning_rate": 5.789473684210527e-05,
      "loss": 1.5859,
      "step": 22
    },
    {
      "epoch": 0.54,
      "learning_rate": 6.052631578947369e-05,
      "loss": 1.7574,
      "step": 23
    },
    {
      "epoch": 0.56,
      "learning_rate": 6.31578947368421e-05,
      "loss": 1.8601,
      "step": 24
    },
    {
      "epoch": 0.58,
      "learning_rate": 6.578947368421054e-05,
      "loss": 1.7867,
      "step": 25
    },
    {
      "epoch": 0.61,
      "learning_rate": 6.842105263157895e-05,
      "loss": 1.9736,
      "step": 26
    },
    {
      "epoch": 0.63,
      "learning_rate": 7.105263157894737e-05,
      "loss": 1.96,
      "step": 27
    },
    {
      "epoch": 0.65,
      "learning_rate": 7.368421052631579e-05,
      "loss": 1.7974,
      "step": 28
    },
    {
      "epoch": 0.68,
      "learning_rate": 7.631578947368422e-05,
      "loss": 1.8705,
      "step": 29
    },
    {
      "epoch": 0.7,
      "learning_rate": 7.894736842105263e-05,
      "loss": 1.8634,
      "step": 30
    },
    {
      "epoch": 0.72,
      "learning_rate": 8.157894736842105e-05,
      "loss": 1.6359,
      "step": 31
    },
    {
      "epoch": 0.75,
      "learning_rate": 8.421052631578948e-05,
      "loss": 1.6283,
      "step": 32
    },
    {
      "epoch": 0.77,
      "learning_rate": 8.68421052631579e-05,
      "loss": 1.8596,
      "step": 33
    },
    {
      "epoch": 0.79,
      "learning_rate": 8.947368421052632e-05,
      "loss": 1.7455,
      "step": 34
    },
    {
      "epoch": 0.82,
      "learning_rate": 9.210526315789474e-05,
      "loss": 1.7927,
      "step": 35
    },
    {
      "epoch": 0.84,
      "learning_rate": 9.473684210526316e-05,
      "loss": 1.8419,
      "step": 36
    },
    {
      "epoch": 0.86,
      "learning_rate": 9.736842105263158e-05,
      "loss": 1.8011,
      "step": 37
    },
    {
      "epoch": 0.89,
      "learning_rate": 0.0001,
      "loss": 1.8074,
      "step": 38
    },
    {
      "epoch": 0.91,
      "learning_rate": 9.983108108108109e-05,
      "loss": 2.0587,
      "step": 39
    },
    {
      "epoch": 0.93,
      "learning_rate": 9.966216216216217e-05,
      "loss": 1.9083,
      "step": 40
    },
    {
      "epoch": 0.96,
      "learning_rate": 9.949324324324325e-05,
      "loss": 1.7381,
      "step": 41
    },
    {
      "epoch": 0.98,
      "learning_rate": 9.932432432432433e-05,
      "loss": 1.9415,
      "step": 42
    },
    {
      "epoch": 1.0,
      "learning_rate": 9.915540540540541e-05,
      "loss": 1.6324,
      "step": 43
    },
    {
      "epoch": 1.03,
      "learning_rate": 9.89864864864865e-05,
      "loss": 1.5839,
      "step": 44
    },
    {
      "epoch": 1.05,
      "learning_rate": 9.881756756756757e-05,
      "loss": 1.8506,
      "step": 45
    },
    {
      "epoch": 1.07,
      "learning_rate": 9.864864864864865e-05,
      "loss": 1.8398,
      "step": 46
    },
    {
      "epoch": 1.1,
      "learning_rate": 9.847972972972973e-05,
      "loss": 1.8458,
      "step": 47
    },
    {
      "epoch": 1.12,
      "learning_rate": 9.831081081081081e-05,
      "loss": 1.8309,
      "step": 48
    },
    {
      "epoch": 1.14,
      "learning_rate": 9.81418918918919e-05,
      "loss": 1.8339,
      "step": 49
    },
    {
      "epoch": 1.17,
      "learning_rate": 9.797297297297297e-05,
      "loss": 1.815,
      "step": 50
    },
    {
      "epoch": 1.19,
      "learning_rate": 9.780405405405407e-05,
      "loss": 1.7453,
      "step": 51
    },
    {
      "epoch": 1.21,
      "learning_rate": 9.763513513513513e-05,
      "loss": 1.7974,
      "step": 52
    },
    {
      "epoch": 1.23,
      "learning_rate": 9.746621621621623e-05,
      "loss": 1.7996,
      "step": 53
    },
    {
      "epoch": 1.26,
      "learning_rate": 9.729729729729731e-05,
      "loss": 1.5471,
      "step": 54
    },
    {
      "epoch": 1.28,
      "learning_rate": 9.712837837837838e-05,
      "loss": 1.6582,
      "step": 55
    },
    {
      "epoch": 1.3,
      "learning_rate": 9.695945945945947e-05,
      "loss": 1.7868,
      "step": 56
    },
    {
      "epoch": 1.33,
      "learning_rate": 9.679054054054054e-05,
      "loss": 1.7449,
      "step": 57
    },
    {
      "epoch": 1.35,
      "learning_rate": 9.662162162162163e-05,
      "loss": 1.9309,
      "step": 58
    },
    {
      "epoch": 1.37,
      "learning_rate": 9.64527027027027e-05,
      "loss": 1.6582,
      "step": 59
    },
    {
      "epoch": 1.4,
      "learning_rate": 9.628378378378379e-05,
      "loss": 1.749,
      "step": 60
    },
    {
      "epoch": 1.42,
      "learning_rate": 9.611486486486487e-05,
      "loss": 1.7815,
      "step": 61
    },
    {
      "epoch": 1.44,
      "learning_rate": 9.594594594594595e-05,
      "loss": 1.7758,
      "step": 62
    },
    {
      "epoch": 1.47,
      "learning_rate": 9.577702702702703e-05,
      "loss": 1.6416,
      "step": 63
    },
    {
      "epoch": 1.49,
      "learning_rate": 9.56081081081081e-05,
      "loss": 1.6242,
      "step": 64
    },
    {
      "epoch": 1.51,
      "learning_rate": 9.54391891891892e-05,
      "loss": 1.5596,
      "step": 65
    },
    {
      "epoch": 1.54,
      "learning_rate": 9.527027027027028e-05,
      "loss": 1.7654,
      "step": 66
    },
    {
      "epoch": 1.56,
      "learning_rate": 9.510135135135136e-05,
      "loss": 1.7598,
      "step": 67
    },
    {
      "epoch": 1.58,
      "learning_rate": 9.493243243243244e-05,
      "loss": 1.7758,
      "step": 68
    },
    {
      "epoch": 1.61,
      "learning_rate": 9.476351351351352e-05,
      "loss": 1.9316,
      "step": 69
    },
    {
      "epoch": 1.63,
      "learning_rate": 9.45945945945946e-05,
      "loss": 1.766,
      "step": 70
    },
    {
      "epoch": 1.65,
      "learning_rate": 9.442567567567568e-05,
      "loss": 1.8492,
      "step": 71
    },
    {
      "epoch": 1.68,
      "learning_rate": 9.425675675675676e-05,
      "loss": 1.8198,
      "step": 72
    },
    {
      "epoch": 1.7,
      "learning_rate": 9.408783783783784e-05,
      "loss": 1.445,
      "step": 73
    },
    {
      "epoch": 1.72,
      "learning_rate": 9.391891891891892e-05,
      "loss": 1.5804,
      "step": 74
    },
    {
      "epoch": 1.75,
      "learning_rate": 9.375e-05,
      "loss": 1.5489,
      "step": 75
    },
    {
      "epoch": 1.77,
      "learning_rate": 9.35810810810811e-05,
      "loss": 1.7003,
      "step": 76
    },
    {
      "epoch": 1.79,
      "learning_rate": 9.341216216216216e-05,
      "loss": 1.6788,
      "step": 77
    },
    {
      "epoch": 1.82,
      "learning_rate": 9.324324324324324e-05,
      "loss": 1.8201,
      "step": 78
    },
    {
      "epoch": 1.84,
      "learning_rate": 9.307432432432432e-05,
      "loss": 1.891,
      "step": 79
    },
    {
      "epoch": 1.86,
      "learning_rate": 9.29054054054054e-05,
      "loss": 1.7963,
      "step": 80
    },
    {
      "epoch": 1.89,
      "learning_rate": 9.27364864864865e-05,
      "loss": 1.8285,
      "step": 81
    },
    {
      "epoch": 1.91,
      "learning_rate": 9.256756756756757e-05,
      "loss": 1.9649,
      "step": 82
    },
    {
      "epoch": 1.93,
      "learning_rate": 9.239864864864866e-05,
      "loss": 1.471,
      "step": 83
    },
    {
      "epoch": 1.96,
      "learning_rate": 9.222972972972973e-05,
      "loss": 1.6209,
      "step": 84
    },
    {
      "epoch": 1.98,
      "learning_rate": 9.206081081081082e-05,
      "loss": 1.8244,
      "step": 85
    },
    {
      "epoch": 2.0,
      "learning_rate": 9.18918918918919e-05,
      "loss": 1.8515,
      "step": 86
    },
    {
      "epoch": 2.03,
      "learning_rate": 9.172297297297297e-05,
      "loss": 1.6504,
      "step": 87
    },
    {
      "epoch": 2.05,
      "learning_rate": 9.155405405405406e-05,
      "loss": 1.5222,
      "step": 88
    },
    {
      "epoch": 2.07,
      "learning_rate": 9.138513513513513e-05,
      "loss": 1.7524,
      "step": 89
    },
    {
      "epoch": 2.1,
      "learning_rate": 9.121621621621623e-05,
      "loss": 1.8002,
      "step": 90
    },
    {
      "epoch": 2.12,
      "learning_rate": 9.10472972972973e-05,
      "loss": 1.8035,
      "step": 91
    },
    {
      "epoch": 2.14,
      "learning_rate": 9.087837837837839e-05,
      "loss": 1.8115,
      "step": 92
    },
    {
      "epoch": 2.17,
      "learning_rate": 9.070945945945947e-05,
      "loss": 1.7192,
      "step": 93
    },
    {
      "epoch": 2.19,
      "learning_rate": 9.054054054054055e-05,
      "loss": 1.7066,
      "step": 94
    },
    {
      "epoch": 2.21,
      "learning_rate": 9.037162162162163e-05,
      "loss": 1.6312,
      "step": 95
    },
    {
      "epoch": 2.24,
      "learning_rate": 9.02027027027027e-05,
      "loss": 1.2849,
      "step": 96
    },
    {
      "epoch": 2.26,
      "learning_rate": 9.003378378378379e-05,
      "loss": 1.5149,
      "step": 97
    },
    {
      "epoch": 2.28,
      "learning_rate": 8.986486486486487e-05,
      "loss": 1.4894,
      "step": 98
    },
    {
      "epoch": 2.31,
      "learning_rate": 8.969594594594595e-05,
      "loss": 1.7987,
      "step": 99
    },
    {
      "epoch": 2.33,
      "learning_rate": 8.952702702702703e-05,
      "loss": 1.6454,
      "step": 100
    },
    {
      "epoch": 2.35,
      "learning_rate": 8.935810810810811e-05,
      "loss": 1.8256,
      "step": 101
    },
    {
      "epoch": 2.38,
      "learning_rate": 8.918918918918919e-05,
      "loss": 1.7403,
      "step": 102
    },
    {
      "epoch": 2.4,
      "learning_rate": 8.902027027027027e-05,
      "loss": 1.6801,
      "step": 103
    },
    {
      "epoch": 2.42,
      "learning_rate": 8.885135135135135e-05,
      "loss": 1.7089,
      "step": 104
    },
    {
      "epoch": 2.45,
      "learning_rate": 8.868243243243243e-05,
      "loss": 1.7009,
      "step": 105
    },
    {
      "epoch": 2.47,
      "learning_rate": 8.851351351351352e-05,
      "loss": 1.3085,
      "step": 106
    },
    {
      "epoch": 2.49,
      "learning_rate": 8.83445945945946e-05,
      "loss": 1.5615,
      "step": 107
    },
    {
      "epoch": 2.52,
      "learning_rate": 8.817567567567569e-05,
      "loss": 1.6415,
      "step": 108
    },
    {
      "epoch": 2.54,
      "learning_rate": 8.800675675675676e-05,
      "loss": 1.5999,
      "step": 109
    },
    {
      "epoch": 2.56,
      "learning_rate": 8.783783783783784e-05,
      "loss": 1.6671,
      "step": 110
    },
    {
      "epoch": 2.59,
      "learning_rate": 8.766891891891892e-05,
      "loss": 1.8566,
      "step": 111
    },
    {
      "epoch": 2.61,
      "learning_rate": 8.75e-05,
      "loss": 1.9677,
      "step": 112
    },
    {
      "epoch": 2.63,
      "learning_rate": 8.73310810810811e-05,
      "loss": 1.6473,
      "step": 113
    },
    {
      "epoch": 2.66,
      "learning_rate": 8.716216216216216e-05,
      "loss": 1.6044,
      "step": 114
    },
    {
      "epoch": 2.68,
      "learning_rate": 8.699324324324325e-05,
      "loss": 1.7437,
      "step": 115
    },
    {
      "epoch": 2.7,
      "learning_rate": 8.682432432432432e-05,
      "loss": 1.2056,
      "step": 116
    },
    {
      "epoch": 2.73,
      "learning_rate": 8.665540540540542e-05,
      "loss": 1.5716,
      "step": 117
    },
    {
      "epoch": 2.75,
      "learning_rate": 8.64864864864865e-05,
      "loss": 1.6996,
      "step": 118
    },
    {
      "epoch": 2.77,
      "learning_rate": 8.631756756756756e-05,
      "loss": 1.7896,
      "step": 119
    },
    {
      "epoch": 2.8,
      "learning_rate": 8.614864864864866e-05,
      "loss": 1.7133,
      "step": 120
    },
    {
      "epoch": 2.82,
      "learning_rate": 8.597972972972972e-05,
      "loss": 1.6978,
      "step": 121
    },
    {
      "epoch": 2.84,
      "learning_rate": 8.581081081081082e-05,
      "loss": 1.7591,
      "step": 122
    },
    {
      "epoch": 2.87,
      "learning_rate": 8.56418918918919e-05,
      "loss": 1.6284,
      "step": 123
    },
    {
      "epoch": 2.89,
      "learning_rate": 8.547297297297298e-05,
      "loss": 1.8141,
      "step": 124
    },
    {
      "epoch": 2.91,
      "learning_rate": 8.530405405405406e-05,
      "loss": 1.6254,
      "step": 125
    },
    {
      "epoch": 2.94,
      "learning_rate": 8.513513513513514e-05,
      "loss": 1.5776,
      "step": 126
    },
    {
      "epoch": 2.96,
      "learning_rate": 8.496621621621622e-05,
      "loss": 1.7454,
      "step": 127
    },
    {
      "epoch": 2.98,
      "learning_rate": 8.47972972972973e-05,
      "loss": 1.8205,
      "step": 128
    },
    {
      "epoch": 3.01,
      "learning_rate": 8.462837837837838e-05,
      "loss": 1.623,
      "step": 129
    },
    {
      "epoch": 3.03,
      "learning_rate": 8.445945945945946e-05,
      "loss": 1.6958,
      "step": 130
    },
    {
      "epoch": 3.05,
      "learning_rate": 8.429054054054054e-05,
      "loss": 1.6087,
      "step": 131
    },
    {
      "epoch": 3.08,
      "learning_rate": 8.412162162162163e-05,
      "loss": 1.7991,
      "step": 132
    },
    {
      "epoch": 3.1,
      "learning_rate": 8.39527027027027e-05,
      "loss": 1.7432,
      "step": 133
    },
    {
      "epoch": 3.12,
      "learning_rate": 8.378378378378379e-05,
      "loss": 1.786,
      "step": 134
    },
    {
      "epoch": 3.15,
      "learning_rate": 8.361486486486487e-05,
      "loss": 1.641,
      "step": 135
    },
    {
      "epoch": 3.17,
      "learning_rate": 8.344594594594595e-05,
      "loss": 1.7251,
      "step": 136
    },
    {
      "epoch": 3.19,
      "learning_rate": 8.327702702702703e-05,
      "loss": 1.514,
      "step": 137
    },
    {
      "epoch": 3.22,
      "learning_rate": 8.310810810810811e-05,
      "loss": 1.5448,
      "step": 138
    },
    {
      "epoch": 3.24,
      "learning_rate": 8.293918918918919e-05,
      "loss": 1.3041,
      "step": 139
    },
    {
      "epoch": 3.26,
      "learning_rate": 8.277027027027028e-05,
      "loss": 1.5238,
      "step": 140
    },
    {
      "epoch": 3.29,
      "learning_rate": 8.260135135135135e-05,
      "loss": 1.6456,
      "step": 141
    },
    {
      "epoch": 3.31,
      "learning_rate": 8.243243243243243e-05,
      "loss": 1.6673,
      "step": 142
    },
    {
      "epoch": 3.33,
      "learning_rate": 8.226351351351351e-05,
      "loss": 1.6674,
      "step": 143
    },
    {
      "epoch": 3.36,
      "learning_rate": 8.209459459459459e-05,
      "loss": 1.7751,
      "step": 144
    },
    {
      "epoch": 3.38,
      "learning_rate": 8.192567567567569e-05,
      "loss": 1.6804,
      "step": 145
    },
    {
      "epoch": 3.4,
      "learning_rate": 8.175675675675675e-05,
      "loss": 1.7012,
      "step": 146
    },
    {
      "epoch": 3.43,
      "learning_rate": 8.158783783783785e-05,
      "loss": 1.6802,
      "step": 147
    },
    {
      "epoch": 3.45,
      "learning_rate": 8.141891891891892e-05,
      "loss": 1.4938,
      "step": 148
    },
    {
      "epoch": 3.47,
      "learning_rate": 8.125000000000001e-05,
      "loss": 1.2082,
      "step": 149
    },
    {
      "epoch": 3.5,
      "learning_rate": 8.108108108108109e-05,
      "loss": 1.5124,
      "step": 150
    },
    {
      "epoch": 3.52,
      "learning_rate": 8.091216216216216e-05,
      "loss": 1.4804,
      "step": 151
    },
    {
      "epoch": 3.54,
      "learning_rate": 8.074324324324325e-05,
      "loss": 1.7145,
      "step": 152
    },
    {
      "epoch": 3.57,
      "learning_rate": 8.057432432432432e-05,
      "loss": 1.6206,
      "step": 153
    },
    {
      "epoch": 3.59,
      "learning_rate": 8.040540540540541e-05,
      "loss": 1.6627,
      "step": 154
    },
    {
      "epoch": 3.61,
      "learning_rate": 8.02364864864865e-05,
      "loss": 1.7419,
      "step": 155
    },
    {
      "epoch": 3.63,
      "learning_rate": 8.006756756756757e-05,
      "loss": 1.6048,
      "step": 156
    },
    {
      "epoch": 3.66,
      "learning_rate": 7.989864864864865e-05,
      "loss": 1.3656,
      "step": 157
    },
    {
      "epoch": 3.68,
      "learning_rate": 7.972972972972974e-05,
      "loss": 1.4486,
      "step": 158
    },
    {
      "epoch": 3.7,
      "learning_rate": 7.956081081081082e-05,
      "loss": 0.9608,
      "step": 159
    },
    {
      "epoch": 3.73,
      "learning_rate": 7.93918918918919e-05,
      "loss": 1.6131,
      "step": 160
    },
    {
      "epoch": 3.75,
      "learning_rate": 7.922297297297298e-05,
      "loss": 1.5565,
      "step": 161
    },
    {
      "epoch": 3.77,
      "learning_rate": 7.905405405405406e-05,
      "loss": 1.6731,
      "step": 162
    },
    {
      "epoch": 3.8,
      "learning_rate": 7.888513513513514e-05,
      "loss": 1.6897,
      "step": 163
    },
    {
      "epoch": 3.82,
      "learning_rate": 7.871621621621622e-05,
      "loss": 1.6846,
      "step": 164
    },
    {
      "epoch": 3.84,
      "learning_rate": 7.85472972972973e-05,
      "loss": 1.7189,
      "step": 165
    },
    {
      "epoch": 3.87,
      "learning_rate": 7.837837837837838e-05,
      "loss": 1.6709,
      "step": 166
    },
    {
      "epoch": 3.89,
      "learning_rate": 7.820945945945946e-05,
      "loss": 1.7089,
      "step": 167
    },
    {
      "epoch": 3.91,
      "learning_rate": 7.804054054054054e-05,
      "loss": 1.4097,
      "step": 168
    },
    {
      "epoch": 3.94,
      "learning_rate": 7.787162162162162e-05,
      "loss": 1.0461,
      "step": 169
    },
    {
      "epoch": 3.96,
      "learning_rate": 7.77027027027027e-05,
      "loss": 1.6682,
      "step": 170
    },
    {
      "epoch": 3.98,
      "learning_rate": 7.753378378378378e-05,
      "loss": 1.3962,
      "step": 171
    },
    {
      "epoch": 4.01,
      "learning_rate": 7.736486486486488e-05,
      "loss": 1.296,
      "step": 172
    },
    {
      "epoch": 4.03,
      "learning_rate": 7.719594594594595e-05,
      "loss": 1.5885,
      "step": 173
    },
    {
      "epoch": 4.05,
      "learning_rate": 7.702702702702703e-05,
      "loss": 1.6864,
      "step": 174
    },
    {
      "epoch": 4.08,
      "learning_rate": 7.68581081081081e-05,
      "loss": 1.7391,
      "step": 175
    },
    {
      "epoch": 4.1,
      "learning_rate": 7.668918918918919e-05,
      "loss": 1.6422,
      "step": 176
    },
    {
      "epoch": 4.12,
      "learning_rate": 7.652027027027028e-05,
      "loss": 1.7226,
      "step": 177
    },
    {
      "epoch": 4.15,
      "learning_rate": 7.635135135135135e-05,
      "loss": 1.5965,
      "step": 178
    },
    {
      "epoch": 4.17,
      "learning_rate": 7.618243243243244e-05,
      "loss": 1.4925,
      "step": 179
    },
    {
      "epoch": 4.19,
      "learning_rate": 7.601351351351351e-05,
      "loss": 1.3451,
      "step": 180
    },
    {
      "epoch": 4.22,
      "learning_rate": 7.58445945945946e-05,
      "loss": 1.2172,
      "step": 181
    },
    {
      "epoch": 4.24,
      "learning_rate": 7.567567567567568e-05,
      "loss": 0.8116,
      "step": 182
    },
    {
      "epoch": 4.26,
      "learning_rate": 7.550675675675675e-05,
      "loss": 1.4927,
      "step": 183
    },
    {
      "epoch": 4.29,
      "learning_rate": 7.533783783783785e-05,
      "loss": 1.522,
      "step": 184
    },
    {
      "epoch": 4.31,
      "learning_rate": 7.516891891891891e-05,
      "loss": 1.6993,
      "step": 185
    },
    {
      "epoch": 4.33,
      "learning_rate": 7.500000000000001e-05,
      "loss": 1.5105,
      "step": 186
    },
    {
      "epoch": 4.36,
      "learning_rate": 7.483108108108109e-05,
      "loss": 1.6603,
      "step": 187
    },
    {
      "epoch": 4.38,
      "learning_rate": 7.466216216216217e-05,
      "loss": 1.7891,
      "step": 188
    },
    {
      "epoch": 4.4,
      "learning_rate": 7.449324324324325e-05,
      "loss": 1.6565,
      "step": 189
    },
    {
      "epoch": 4.43,
      "learning_rate": 7.432432432432433e-05,
      "loss": 1.4568,
      "step": 190
    },
    {
      "epoch": 4.45,
      "learning_rate": 7.415540540540541e-05,
      "loss": 1.0829,
      "step": 191
    },
    {
      "epoch": 4.47,
      "learning_rate": 7.398648648648649e-05,
      "loss": 0.9379,
      "step": 192
    },
    {
      "epoch": 4.5,
      "learning_rate": 7.381756756756757e-05,
      "loss": 1.5581,
      "step": 193
    },
    {
      "epoch": 4.52,
      "learning_rate": 7.364864864864865e-05,
      "loss": 1.4996,
      "step": 194
    },
    {
      "epoch": 4.54,
      "learning_rate": 7.347972972972973e-05,
      "loss": 1.6636,
      "step": 195
    },
    {
      "epoch": 4.57,
      "learning_rate": 7.331081081081081e-05,
      "loss": 1.5504,
      "step": 196
    },
    {
      "epoch": 4.59,
      "learning_rate": 7.31418918918919e-05,
      "loss": 1.6647,
      "step": 197
    },
    {
      "epoch": 4.61,
      "learning_rate": 7.297297297297297e-05,
      "loss": 1.401,
      "step": 198
    },
    {
      "epoch": 4.64,
      "learning_rate": 7.280405405405406e-05,
      "loss": 1.4887,
      "step": 199
    },
    {
      "epoch": 4.66,
      "learning_rate": 7.263513513513514e-05,
      "loss": 1.4007,
      "step": 200
    },
    {
      "epoch": 4.68,
      "learning_rate": 7.246621621621622e-05,
      "loss": 1.2743,
      "step": 201
    },
    {
      "epoch": 4.71,
      "learning_rate": 7.229729729729731e-05,
      "loss": 0.9989,
      "step": 202
    },
    {
      "epoch": 4.73,
      "learning_rate": 7.212837837837838e-05,
      "loss": 1.5474,
      "step": 203
    },
    {
      "epoch": 4.75,
      "learning_rate": 7.195945945945947e-05,
      "loss": 1.5369,
      "step": 204
    },
    {
      "epoch": 4.78,
      "learning_rate": 7.179054054054054e-05,
      "loss": 1.672,
      "step": 205
    },
    {
      "epoch": 4.8,
      "learning_rate": 7.162162162162162e-05,
      "loss": 1.5355,
      "step": 206
    },
    {
      "epoch": 4.82,
      "learning_rate": 7.14527027027027e-05,
      "loss": 1.5252,
      "step": 207
    },
    {
      "epoch": 4.85,
      "learning_rate": 7.128378378378378e-05,
      "loss": 1.488,
      "step": 208
    },
    {
      "epoch": 4.87,
      "learning_rate": 7.111486486486488e-05,
      "loss": 1.6075,
      "step": 209
    },
    {
      "epoch": 4.89,
      "learning_rate": 7.094594594594594e-05,
      "loss": 1.2268,
      "step": 210
    },
    {
      "epoch": 4.92,
      "learning_rate": 7.077702702702704e-05,
      "loss": 1.21,
      "step": 211
    },
    {
      "epoch": 4.94,
      "learning_rate": 7.06081081081081e-05,
      "loss": 0.8415,
      "step": 212
    },
    {
      "epoch": 4.96,
      "learning_rate": 7.04391891891892e-05,
      "loss": 1.4969,
      "step": 213
    },
    {
      "epoch": 4.99,
      "learning_rate": 7.027027027027028e-05,
      "loss": 1.4499,
      "step": 214
    },
    {
      "epoch": 5.01,
      "learning_rate": 7.010135135135135e-05,
      "loss": 1.2599,
      "step": 215
    },
    {
      "epoch": 5.03,
      "learning_rate": 6.993243243243244e-05,
      "loss": 1.5025,
      "step": 216
    },
    {
      "epoch": 5.06,
      "learning_rate": 6.97635135135135e-05,
      "loss": 1.4723,
      "step": 217
    },
    {
      "epoch": 5.08,
      "learning_rate": 6.95945945945946e-05,
      "loss": 1.615,
      "step": 218
    },
    {
      "epoch": 5.1,
      "learning_rate": 6.942567567567568e-05,
      "loss": 1.4314,
      "step": 219
    },
    {
      "epoch": 5.13,
      "learning_rate": 6.925675675675676e-05,
      "loss": 1.5237,
      "step": 220
    },
    {
      "epoch": 5.15,
      "learning_rate": 6.908783783783784e-05,
      "loss": 1.4065,
      "step": 221
    },
    {
      "epoch": 5.17,
      "learning_rate": 6.891891891891892e-05,
      "loss": 1.3899,
      "step": 222
    },
    {
      "epoch": 5.2,
      "learning_rate": 6.875e-05,
      "loss": 1.1835,
      "step": 223
    },
    {
      "epoch": 5.22,
      "learning_rate": 6.858108108108108e-05,
      "loss": 1.0849,
      "step": 224
    },
    {
      "epoch": 5.24,
      "learning_rate": 6.841216216216217e-05,
      "loss": 0.9062,
      "step": 225
    },
    {
      "epoch": 5.27,
      "learning_rate": 6.824324324324325e-05,
      "loss": 1.5028,
      "step": 226
    },
    {
      "epoch": 5.29,
      "learning_rate": 6.807432432432433e-05,
      "loss": 1.6093,
      "step": 227
    },
    {
      "epoch": 5.31,
      "learning_rate": 6.790540540540541e-05,
      "loss": 1.5773,
      "step": 228
    },
    {
      "epoch": 5.34,
      "learning_rate": 6.773648648648649e-05,
      "loss": 1.4287,
      "step": 229
    },
    {
      "epoch": 5.36,
      "learning_rate": 6.756756756756757e-05,
      "loss": 1.3902,
      "step": 230
    },
    {
      "epoch": 5.38,
      "learning_rate": 6.739864864864865e-05,
      "loss": 1.4446,
      "step": 231
    },
    {
      "epoch": 5.41,
      "learning_rate": 6.722972972972973e-05,
      "loss": 1.3632,
      "step": 232
    },
    {
      "epoch": 5.43,
      "learning_rate": 6.706081081081081e-05,
      "loss": 1.017,
      "step": 233
    },
    {
      "epoch": 5.45,
      "learning_rate": 6.68918918918919e-05,
      "loss": 0.5359,
      "step": 234
    },
    {
      "epoch": 5.48,
      "learning_rate": 6.672297297297297e-05,
      "loss": 0.8123,
      "step": 235
    },
    {
      "epoch": 5.5,
      "learning_rate": 6.655405405405407e-05,
      "loss": 1.5714,
      "step": 236
    },
    {
      "epoch": 5.52,
      "learning_rate": 6.638513513513513e-05,
      "loss": 1.535,
      "step": 237
    },
    {
      "epoch": 5.55,
      "learning_rate": 6.621621621621621e-05,
      "loss": 1.7028,
      "step": 238
    },
    {
      "epoch": 5.57,
      "learning_rate": 6.604729729729731e-05,
      "loss": 1.4892,
      "step": 239
    },
    {
      "epoch": 5.59,
      "learning_rate": 6.587837837837837e-05,
      "loss": 1.6333,
      "step": 240
    },
    {
      "epoch": 5.62,
      "learning_rate": 6.570945945945947e-05,
      "loss": 1.5218,
      "step": 241
    },
    {
      "epoch": 5.64,
      "learning_rate": 6.554054054054054e-05,
      "loss": 1.281,
      "step": 242
    },
    {
      "epoch": 5.66,
      "learning_rate": 6.537162162162163e-05,
      "loss": 1.1913,
      "step": 243
    },
    {
      "epoch": 5.69,
      "learning_rate": 6.52027027027027e-05,
      "loss": 0.9462,
      "step": 244
    },
    {
      "epoch": 5.71,
      "learning_rate": 6.503378378378379e-05,
      "loss": 0.8686,
      "step": 245
    },
    {
      "epoch": 5.73,
      "learning_rate": 6.486486486486487e-05,
      "loss": 1.4999,
      "step": 246
    },
    {
      "epoch": 5.76,
      "learning_rate": 6.469594594594594e-05,
      "loss": 1.5577,
      "step": 247
    },
    {
      "epoch": 5.78,
      "learning_rate": 6.452702702702703e-05,
      "loss": 1.5312,
      "step": 248
    },
    {
      "epoch": 5.8,
      "learning_rate": 6.43581081081081e-05,
      "loss": 1.4374,
      "step": 249
    },
    {
      "epoch": 5.83,
      "learning_rate": 6.41891891891892e-05,
      "loss": 1.5982,
      "step": 250
    },
    {
      "epoch": 5.85,
      "learning_rate": 6.402027027027028e-05,
      "loss": 1.5016,
      "step": 251
    },
    {
      "epoch": 5.87,
      "learning_rate": 6.385135135135136e-05,
      "loss": 1.2885,
      "step": 252
    },
    {
      "epoch": 5.9,
      "learning_rate": 6.368243243243244e-05,
      "loss": 1.0857,
      "step": 253
    },
    {
      "epoch": 5.92,
      "learning_rate": 6.351351351351352e-05,
      "loss": 0.8698,
      "step": 254
    },
    {
      "epoch": 5.94,
      "learning_rate": 6.33445945945946e-05,
      "loss": 0.843,
      "step": 255
    },
    {
      "epoch": 5.97,
      "learning_rate": 6.317567567567568e-05,
      "loss": 1.6899,
      "step": 256
    },
    {
      "epoch": 5.99,
      "learning_rate": 6.300675675675676e-05,
      "loss": 1.4388,
      "step": 257
    },
    {
      "epoch": 6.01,
      "learning_rate": 6.283783783783784e-05,
      "loss": 1.2304,
      "step": 258
    },
    {
      "epoch": 6.03,
      "learning_rate": 6.266891891891892e-05,
      "loss": 1.4798,
      "step": 259
    },
    {
      "epoch": 6.06,
      "learning_rate": 6.25e-05,
      "loss": 1.4962,
      "step": 260
    },
    {
      "epoch": 6.08,
      "learning_rate": 6.233108108108108e-05,
      "loss": 1.5346,
      "step": 261
    },
    {
      "epoch": 6.1,
      "learning_rate": 6.216216216216216e-05,
      "loss": 1.5259,
      "step": 262
    },
    {
      "epoch": 6.13,
      "learning_rate": 6.199324324324324e-05,
      "loss": 1.3478,
      "step": 263
    },
    {
      "epoch": 6.15,
      "learning_rate": 6.182432432432432e-05,
      "loss": 1.2349,
      "step": 264
    },
    {
      "epoch": 6.17,
      "learning_rate": 6.16554054054054e-05,
      "loss": 1.3345,
      "step": 265
    },
    {
      "epoch": 6.2,
      "learning_rate": 6.14864864864865e-05,
      "loss": 1.0341,
      "step": 266
    },
    {
      "epoch": 6.22,
      "learning_rate": 6.131756756756757e-05,
      "loss": 0.7703,
      "step": 267
    },
    {
      "epoch": 6.24,
      "learning_rate": 6.114864864864866e-05,
      "loss": 0.8846,
      "step": 268
    },
    {
      "epoch": 6.27,
      "learning_rate": 6.097972972972973e-05,
      "loss": 1.4626,
      "step": 269
    },
    {
      "epoch": 6.29,
      "learning_rate": 6.0810810810810814e-05,
      "loss": 1.4799,
      "step": 270
    },
    {
      "epoch": 6.31,
      "learning_rate": 6.06418918918919e-05,
      "loss": 1.4394,
      "step": 271
    },
    {
      "epoch": 6.34,
      "learning_rate": 6.0472972972972976e-05,
      "loss": 1.2576,
      "step": 272
    },
    {
      "epoch": 6.36,
      "learning_rate": 6.030405405405406e-05,
      "loss": 1.5549,
      "step": 273
    },
    {
      "epoch": 6.38,
      "learning_rate": 6.013513513513514e-05,
      "loss": 1.2343,
      "step": 274
    },
    {
      "epoch": 6.41,
      "learning_rate": 5.996621621621622e-05,
      "loss": 1.1688,
      "step": 275
    },
    {
      "epoch": 6.43,
      "learning_rate": 5.9797297297297305e-05,
      "loss": 0.8035,
      "step": 276
    },
    {
      "epoch": 6.45,
      "learning_rate": 5.962837837837838e-05,
      "loss": 0.4664,
      "step": 277
    },
    {
      "epoch": 6.48,
      "learning_rate": 5.9459459459459466e-05,
      "loss": 0.9317,
      "step": 278
    },
    {
      "epoch": 6.5,
      "learning_rate": 5.929054054054054e-05,
      "loss": 1.5164,
      "step": 279
    },
    {
      "epoch": 6.52,
      "learning_rate": 5.912162162162163e-05,
      "loss": 1.5362,
      "step": 280
    },
    {
      "epoch": 6.55,
      "learning_rate": 5.89527027027027e-05,
      "loss": 1.6044,
      "step": 281
    },
    {
      "epoch": 6.57,
      "learning_rate": 5.878378378378379e-05,
      "loss": 1.2894,
      "step": 282
    },
    {
      "epoch": 6.59,
      "learning_rate": 5.861486486486487e-05,
      "loss": 1.5595,
      "step": 283
    },
    {
      "epoch": 6.62,
      "learning_rate": 5.8445945945945943e-05,
      "loss": 1.3092,
      "step": 284
    },
    {
      "epoch": 6.64,
      "learning_rate": 5.827702702702703e-05,
      "loss": 1.1681,
      "step": 285
    },
    {
      "epoch": 6.66,
      "learning_rate": 5.8108108108108105e-05,
      "loss": 0.9013,
      "step": 286
    },
    {
      "epoch": 6.69,
      "learning_rate": 5.793918918918919e-05,
      "loss": 0.4539,
      "step": 287
    },
    {
      "epoch": 6.71,
      "learning_rate": 5.777027027027028e-05,
      "loss": 0.8379,
      "step": 288
    },
    {
      "epoch": 6.73,
      "learning_rate": 5.760135135135135e-05,
      "loss": 1.4747,
      "step": 289
    },
    {
      "epoch": 6.76,
      "learning_rate": 5.7432432432432434e-05,
      "loss": 1.6548,
      "step": 290
    },
    {
      "epoch": 6.78,
      "learning_rate": 5.7263513513513515e-05,
      "loss": 1.5068,
      "step": 291
    },
    {
      "epoch": 6.8,
      "learning_rate": 5.7094594594594595e-05,
      "loss": 1.4375,
      "step": 292
    },
    {
      "epoch": 6.83,
      "learning_rate": 5.692567567567568e-05,
      "loss": 1.3555,
      "step": 293
    },
    {
      "epoch": 6.85,
      "learning_rate": 5.6756756756756757e-05,
      "loss": 1.3434,
      "step": 294
    },
    {
      "epoch": 6.87,
      "learning_rate": 5.6587837837837844e-05,
      "loss": 1.1171,
      "step": 295
    },
    {
      "epoch": 6.9,
      "learning_rate": 5.641891891891892e-05,
      "loss": 1.0005,
      "step": 296
    },
    {
      "epoch": 6.92,
      "learning_rate": 5.6250000000000005e-05,
      "loss": 0.5054,
      "step": 297
    },
    {
      "epoch": 6.94,
      "learning_rate": 5.6081081081081086e-05,
      "loss": 0.8416,
      "step": 298
    },
    {
      "epoch": 6.97,
      "learning_rate": 5.591216216216216e-05,
      "loss": 1.5673,
      "step": 299
    },
    {
      "epoch": 6.99,
      "learning_rate": 5.574324324324325e-05,
      "loss": 1.0909,
      "step": 300
    },
    {
      "epoch": 7.01,
      "learning_rate": 5.557432432432432e-05,
      "loss": 1.0109,
      "step": 301
    },
    {
      "epoch": 7.04,
      "learning_rate": 5.540540540540541e-05,
      "loss": 1.5546,
      "step": 302
    },
    {
      "epoch": 7.06,
      "learning_rate": 5.5236486486486496e-05,
      "loss": 1.3989,
      "step": 303
    },
    {
      "epoch": 7.08,
      "learning_rate": 5.506756756756757e-05,
      "loss": 1.4302,
      "step": 304
    },
    {
      "epoch": 7.11,
      "learning_rate": 5.489864864864866e-05,
      "loss": 1.1426,
      "step": 305
    },
    {
      "epoch": 7.13,
      "learning_rate": 5.472972972972973e-05,
      "loss": 1.2958,
      "step": 306
    },
    {
      "epoch": 7.15,
      "learning_rate": 5.456081081081081e-05,
      "loss": 1.1989,
      "step": 307
    },
    {
      "epoch": 7.18,
      "learning_rate": 5.43918918918919e-05,
      "loss": 0.9586,
      "step": 308
    },
    {
      "epoch": 7.2,
      "learning_rate": 5.422297297297297e-05,
      "loss": 0.7062,
      "step": 309
    },
    {
      "epoch": 7.22,
      "learning_rate": 5.405405405405406e-05,
      "loss": 0.3819,
      "step": 310
    },
    {
      "epoch": 7.25,
      "learning_rate": 5.3885135135135134e-05,
      "loss": 0.9815,
      "step": 311
    },
    {
      "epoch": 7.27,
      "learning_rate": 5.371621621621622e-05,
      "loss": 1.5157,
      "step": 312
    },
    {
      "epoch": 7.29,
      "learning_rate": 5.35472972972973e-05,
      "loss": 1.518,
      "step": 313
    },
    {
      "epoch": 7.32,
      "learning_rate": 5.337837837837838e-05,
      "loss": 1.2894,
      "step": 314
    },
    {
      "epoch": 7.34,
      "learning_rate": 5.3209459459459463e-05,
      "loss": 1.254,
      "step": 315
    },
    {
      "epoch": 7.36,
      "learning_rate": 5.304054054054054e-05,
      "loss": 1.3321,
      "step": 316
    },
    {
      "epoch": 7.39,
      "learning_rate": 5.2871621621621625e-05,
      "loss": 1.0895,
      "step": 317
    },
    {
      "epoch": 7.41,
      "learning_rate": 5.27027027027027e-05,
      "loss": 0.9004,
      "step": 318
    },
    {
      "epoch": 7.43,
      "learning_rate": 5.2533783783783786e-05,
      "loss": 0.4367,
      "step": 319
    },
    {
      "epoch": 7.46,
      "learning_rate": 5.2364864864864873e-05,
      "loss": 0.2074,
      "step": 320
    },
    {
      "epoch": 7.48,
      "learning_rate": 5.219594594594595e-05,
      "loss": 0.9496,
      "step": 321
    },
    {
      "epoch": 7.5,
      "learning_rate": 5.202702702702703e-05,
      "loss": 1.4987,
      "step": 322
    },
    {
      "epoch": 7.53,
      "learning_rate": 5.185810810810811e-05,
      "loss": 1.3814,
      "step": 323
    },
    {
      "epoch": 7.55,
      "learning_rate": 5.168918918918919e-05,
      "loss": 1.5098,
      "step": 324
    },
    {
      "epoch": 7.57,
      "learning_rate": 5.152027027027028e-05,
      "loss": 1.3092,
      "step": 325
    },
    {
      "epoch": 7.6,
      "learning_rate": 5.135135135135135e-05,
      "loss": 1.2461,
      "step": 326
    },
    {
      "epoch": 7.62,
      "learning_rate": 5.118243243243244e-05,
      "loss": 1.2363,
      "step": 327
    },
    {
      "epoch": 7.64,
      "learning_rate": 5.101351351351351e-05,
      "loss": 1.0741,
      "step": 328
    },
    {
      "epoch": 7.67,
      "learning_rate": 5.08445945945946e-05,
      "loss": 0.9025,
      "step": 329
    },
    {
      "epoch": 7.69,
      "learning_rate": 5.067567567567568e-05,
      "loss": 0.5086,
      "step": 330
    },
    {
      "epoch": 7.71,
      "learning_rate": 5.0506756756756754e-05,
      "loss": 0.936,
      "step": 331
    },
    {
      "epoch": 7.74,
      "learning_rate": 5.033783783783784e-05,
      "loss": 1.4979,
      "step": 332
    },
    {
      "epoch": 7.76,
      "learning_rate": 5.0168918918918915e-05,
      "loss": 1.5263,
      "step": 333
    },
    {
      "epoch": 7.78,
      "learning_rate": 5e-05,
      "loss": 1.4706,
      "step": 334
    },
    {
      "epoch": 7.81,
      "learning_rate": 4.983108108108108e-05,
      "loss": 1.4829,
      "step": 335
    },
    {
      "epoch": 7.83,
      "learning_rate": 4.9662162162162164e-05,
      "loss": 1.4617,
      "step": 336
    },
    {
      "epoch": 7.85,
      "learning_rate": 4.949324324324325e-05,
      "loss": 1.0703,
      "step": 337
    },
    {
      "epoch": 7.88,
      "learning_rate": 4.9324324324324325e-05,
      "loss": 0.9947,
      "step": 338
    },
    {
      "epoch": 7.9,
      "learning_rate": 4.9155405405405406e-05,
      "loss": 0.6647,
      "step": 339
    },
    {
      "epoch": 7.92,
      "learning_rate": 4.8986486486486486e-05,
      "loss": 0.261,
      "step": 340
    },
    {
      "epoch": 7.95,
      "learning_rate": 4.881756756756757e-05,
      "loss": 0.9715,
      "step": 341
    },
    {
      "epoch": 7.97,
      "learning_rate": 4.8648648648648654e-05,
      "loss": 1.3903,
      "step": 342
    },
    {
      "epoch": 7.99,
      "learning_rate": 4.8479729729729735e-05,
      "loss": 1.0558,
      "step": 343
    },
    {
      "epoch": 8.02,
      "learning_rate": 4.8310810810810816e-05,
      "loss": 1.1678,
      "step": 344
    },
    {
      "epoch": 8.04,
      "learning_rate": 4.8141891891891896e-05,
      "loss": 1.394,
      "step": 345
    },
    {
      "epoch": 8.06,
      "learning_rate": 4.797297297297298e-05,
      "loss": 1.4896,
      "step": 346
    },
    {
      "epoch": 8.09,
      "learning_rate": 4.780405405405405e-05,
      "loss": 1.4282,
      "step": 347
    },
    {
      "epoch": 8.11,
      "learning_rate": 4.763513513513514e-05,
      "loss": 1.086,
      "step": 348
    },
    {
      "epoch": 8.13,
      "learning_rate": 4.746621621621622e-05,
      "loss": 1.1409,
      "step": 349
    },
    {
      "epoch": 8.16,
      "learning_rate": 4.72972972972973e-05,
      "loss": 0.8166,
      "step": 350
    },
    {
      "epoch": 8.18,
      "learning_rate": 4.712837837837838e-05,
      "loss": 0.7514,
      "step": 351
    },
    {
      "epoch": 8.2,
      "learning_rate": 4.695945945945946e-05,
      "loss": 0.3686,
      "step": 352
    },
    {
      "epoch": 8.23,
      "learning_rate": 4.679054054054055e-05,
      "loss": 0.1424,
      "step": 353
    },
    {
      "epoch": 8.25,
      "learning_rate": 4.662162162162162e-05,
      "loss": 1.0332,
      "step": 354
    },
    {
      "epoch": 8.27,
      "learning_rate": 4.64527027027027e-05,
      "loss": 1.5302,
      "step": 355
    },
    {
      "epoch": 8.3,
      "learning_rate": 4.628378378378378e-05,
      "loss": 1.5217,
      "step": 356
    },
    {
      "epoch": 8.32,
      "learning_rate": 4.6114864864864864e-05,
      "loss": 1.351,
      "step": 357
    },
    {
      "epoch": 8.34,
      "learning_rate": 4.594594594594595e-05,
      "loss": 1.188,
      "step": 358
    },
    {
      "epoch": 8.37,
      "learning_rate": 4.577702702702703e-05,
      "loss": 1.1897,
      "step": 359
    },
    {
      "epoch": 8.39,
      "learning_rate": 4.560810810810811e-05,
      "loss": 1.0521,
      "step": 360
    },
    {
      "epoch": 8.41,
      "learning_rate": 4.543918918918919e-05,
      "loss": 0.7367,
      "step": 361
    },
    {
      "epoch": 8.43,
      "learning_rate": 4.5270270270270274e-05,
      "loss": 0.4731,
      "step": 362
    },
    {
      "epoch": 8.46,
      "learning_rate": 4.510135135135135e-05,
      "loss": 0.1531,
      "step": 363
    },
    {
      "epoch": 8.48,
      "learning_rate": 4.4932432432432435e-05,
      "loss": 1.1014,
      "step": 364
    },
    {
      "epoch": 8.5,
      "learning_rate": 4.4763513513513516e-05,
      "loss": 1.4258,
      "step": 365
    },
    {
      "epoch": 8.53,
      "learning_rate": 4.4594594594594596e-05,
      "loss": 1.5873,
      "step": 366
    },
    {
      "epoch": 8.55,
      "learning_rate": 4.442567567567568e-05,
      "loss": 1.3525,
      "step": 367
    },
    {
      "epoch": 8.57,
      "learning_rate": 4.425675675675676e-05,
      "loss": 1.1992,
      "step": 368
    },
    {
      "epoch": 8.6,
      "learning_rate": 4.4087837837837845e-05,
      "loss": 1.1764,
      "step": 369
    },
    {
      "epoch": 8.62,
      "learning_rate": 4.391891891891892e-05,
      "loss": 1.05,
      "step": 370
    },
    {
      "epoch": 8.64,
      "learning_rate": 4.375e-05,
      "loss": 0.9974,
      "step": 371
    },
    {
      "epoch": 8.67,
      "learning_rate": 4.358108108108108e-05,
      "loss": 0.6386,
      "step": 372
    },
    {
      "epoch": 8.69,
      "learning_rate": 4.341216216216216e-05,
      "loss": 0.2399,
      "step": 373
    },
    {
      "epoch": 8.71,
      "learning_rate": 4.324324324324325e-05,
      "loss": 1.1005,
      "step": 374
    },
    {
      "epoch": 8.74,
      "learning_rate": 4.307432432432433e-05,
      "loss": 1.4695,
      "step": 375
    },
    {
      "epoch": 8.76,
      "learning_rate": 4.290540540540541e-05,
      "loss": 1.3389,
      "step": 376
    },
    {
      "epoch": 8.78,
      "learning_rate": 4.273648648648649e-05,
      "loss": 1.2377,
      "step": 377
    },
    {
      "epoch": 8.81,
      "learning_rate": 4.256756756756757e-05,
      "loss": 1.252,
      "step": 378
    },
    {
      "epoch": 8.83,
      "learning_rate": 4.239864864864865e-05,
      "loss": 1.2136,
      "step": 379
    },
    {
      "epoch": 8.85,
      "learning_rate": 4.222972972972973e-05,
      "loss": 1.1641,
      "step": 380
    },
    {
      "epoch": 8.88,
      "learning_rate": 4.206081081081081e-05,
      "loss": 0.8238,
      "step": 381
    },
    {
      "epoch": 8.9,
      "learning_rate": 4.189189189189189e-05,
      "loss": 0.542,
      "step": 382
    },
    {
      "epoch": 8.92,
      "learning_rate": 4.1722972972972974e-05,
      "loss": 0.1831,
      "step": 383
    },
    {
      "epoch": 8.95,
      "learning_rate": 4.1554054054054055e-05,
      "loss": 0.9515,
      "step": 384
    },
    {
      "epoch": 8.97,
      "learning_rate": 4.138513513513514e-05,
      "loss": 1.1611,
      "step": 385
    },
    {
      "epoch": 8.99,
      "learning_rate": 4.1216216216216216e-05,
      "loss": 0.8458,
      "step": 386
    },
    {
      "epoch": 9.02,
      "learning_rate": 4.1047297297297297e-05,
      "loss": 1.1778,
      "step": 387
    },
    {
      "epoch": 9.04,
      "learning_rate": 4.087837837837838e-05,
      "loss": 1.4086,
      "step": 388
    },
    {
      "epoch": 9.06,
      "learning_rate": 4.070945945945946e-05,
      "loss": 1.5189,
      "step": 389
    },
    {
      "epoch": 9.09,
      "learning_rate": 4.0540540540540545e-05,
      "loss": 1.2386,
      "step": 390
    },
    {
      "epoch": 9.11,
      "learning_rate": 4.0371621621621626e-05,
      "loss": 1.0867,
      "step": 391
    },
    {
      "epoch": 9.13,
      "learning_rate": 4.0202702702702707e-05,
      "loss": 1.1069,
      "step": 392
    },
    {
      "epoch": 9.16,
      "learning_rate": 4.003378378378379e-05,
      "loss": 0.9551,
      "step": 393
    },
    {
      "epoch": 9.18,
      "learning_rate": 3.986486486486487e-05,
      "loss": 0.6643,
      "step": 394
    },
    {
      "epoch": 9.2,
      "learning_rate": 3.969594594594595e-05,
      "loss": 0.3289,
      "step": 395
    },
    {
      "epoch": 9.23,
      "learning_rate": 3.952702702702703e-05,
      "loss": 0.1099,
      "step": 396
    },
    {
      "epoch": 9.25,
      "learning_rate": 3.935810810810811e-05,
      "loss": 1.2255,
      "step": 397
    },
    {
      "epoch": 9.27,
      "learning_rate": 3.918918918918919e-05,
      "loss": 1.3092,
      "step": 398
    },
    {
      "epoch": 9.3,
      "learning_rate": 3.902027027027027e-05,
      "loss": 1.4971,
      "step": 399
    },
    {
      "epoch": 9.32,
      "learning_rate": 3.885135135135135e-05,
      "loss": 1.1879,
      "step": 400
    },
    {
      "epoch": 9.34,
      "learning_rate": 3.868243243243244e-05,
      "loss": 1.1136,
      "step": 401
    },
    {
      "epoch": 9.37,
      "learning_rate": 3.851351351351351e-05,
      "loss": 0.9842,
      "step": 402
    },
    {
      "epoch": 9.39,
      "learning_rate": 3.8344594594594594e-05,
      "loss": 0.9181,
      "step": 403
    },
    {
      "epoch": 9.41,
      "learning_rate": 3.8175675675675674e-05,
      "loss": 0.5751,
      "step": 404
    },
    {
      "epoch": 9.44,
      "learning_rate": 3.8006756756756755e-05,
      "loss": 0.2316,
      "step": 405
    },
    {
      "epoch": 9.46,
      "learning_rate": 3.783783783783784e-05,
      "loss": 0.1078,
      "step": 406
    },
    {
      "epoch": 9.48,
      "learning_rate": 3.766891891891892e-05,
      "loss": 1.1299,
      "step": 407
    },
    {
      "epoch": 9.51,
      "learning_rate": 3.7500000000000003e-05,
      "loss": 1.4829,
      "step": 408
    },
    {
      "epoch": 9.53,
      "learning_rate": 3.7331081081081084e-05,
      "loss": 1.2994,
      "step": 409
    },
    {
      "epoch": 9.55,
      "learning_rate": 3.7162162162162165e-05,
      "loss": 1.1856,
      "step": 410
    },
    {
      "epoch": 9.58,
      "learning_rate": 3.6993243243243245e-05,
      "loss": 1.1424,
      "step": 411
    },
    {
      "epoch": 9.6,
      "learning_rate": 3.6824324324324326e-05,
      "loss": 0.8504,
      "step": 412
    },
    {
      "epoch": 9.62,
      "learning_rate": 3.665540540540541e-05,
      "loss": 0.6786,
      "step": 413
    },
    {
      "epoch": 9.65,
      "learning_rate": 3.648648648648649e-05,
      "loss": 0.5237,
      "step": 414
    },
    {
      "epoch": 9.67,
      "learning_rate": 3.631756756756757e-05,
      "loss": 0.3276,
      "step": 415
    },
    {
      "epoch": 9.69,
      "learning_rate": 3.6148648648648655e-05,
      "loss": 0.1227,
      "step": 416
    },
    {
      "epoch": 9.72,
      "learning_rate": 3.5979729729729736e-05,
      "loss": 1.2094,
      "step": 417
    },
    {
      "epoch": 9.74,
      "learning_rate": 3.581081081081081e-05,
      "loss": 1.4648,
      "step": 418
    },
    {
      "epoch": 9.76,
      "learning_rate": 3.564189189189189e-05,
      "loss": 1.4279,
      "step": 419
    },
    {
      "epoch": 9.79,
      "learning_rate": 3.547297297297297e-05,
      "loss": 1.3182,
      "step": 420
    },
    {
      "epoch": 9.81,
      "learning_rate": 3.530405405405405e-05,
      "loss": 1.1322,
      "step": 421
    },
    {
      "epoch": 9.83,
      "learning_rate": 3.513513513513514e-05,
      "loss": 1.0189,
      "step": 422
    },
    {
      "epoch": 9.86,
      "learning_rate": 3.496621621621622e-05,
      "loss": 0.9552,
      "step": 423
    },
    {
      "epoch": 9.88,
      "learning_rate": 3.47972972972973e-05,
      "loss": 0.8533,
      "step": 424
    },
    {
      "epoch": 9.9,
      "learning_rate": 3.462837837837838e-05,
      "loss": 0.4739,
      "step": 425
    },
    {
      "epoch": 9.93,
      "learning_rate": 3.445945945945946e-05,
      "loss": 0.1936,
      "step": 426
    },
    {
      "epoch": 9.95,
      "learning_rate": 3.429054054054054e-05,
      "loss": 1.1758,
      "step": 427
    },
    {
      "epoch": 9.97,
      "learning_rate": 3.412162162162162e-05,
      "loss": 1.26,
      "step": 428
    },
    {
      "epoch": 10.0,
      "learning_rate": 3.3952702702702704e-05,
      "loss": 0.6075,
      "step": 429
    },
    {
      "epoch": 10.02,
      "learning_rate": 3.3783783783783784e-05,
      "loss": 1.3471,
      "step": 430
    },
    {
      "epoch": 10.04,
      "learning_rate": 3.3614864864864865e-05,
      "loss": 1.4233,
      "step": 431
    },
    {
      "epoch": 10.07,
      "learning_rate": 3.344594594594595e-05,
      "loss": 1.4049,
      "step": 432
    },
    {
      "epoch": 10.09,
      "learning_rate": 3.327702702702703e-05,
      "loss": 0.9861,
      "step": 433
    },
    {
      "epoch": 10.11,
      "learning_rate": 3.310810810810811e-05,
      "loss": 0.9746,
      "step": 434
    },
    {
      "epoch": 10.14,
      "learning_rate": 3.293918918918919e-05,
      "loss": 0.8351,
      "step": 435
    },
    {
      "epoch": 10.16,
      "learning_rate": 3.277027027027027e-05,
      "loss": 0.7374,
      "step": 436
    },
    {
      "epoch": 10.18,
      "learning_rate": 3.260135135135135e-05,
      "loss": 0.5089,
      "step": 437
    },
    {
      "epoch": 10.21,
      "learning_rate": 3.2432432432432436e-05,
      "loss": 0.2,
      "step": 438
    },
    {
      "epoch": 10.23,
      "learning_rate": 3.226351351351352e-05,
      "loss": 0.0868,
      "step": 439
    },
    {
      "epoch": 10.25,
      "learning_rate": 3.20945945945946e-05,
      "loss": 1.2262,
      "step": 440
    },
    {
      "epoch": 10.28,
      "learning_rate": 3.192567567567568e-05,
      "loss": 1.3318,
      "step": 441
    },
    {
      "epoch": 10.3,
      "learning_rate": 3.175675675675676e-05,
      "loss": 1.1406,
      "step": 442
    },
    {
      "epoch": 10.32,
      "learning_rate": 3.158783783783784e-05,
      "loss": 1.1061,
      "step": 443
    },
    {
      "epoch": 10.35,
      "learning_rate": 3.141891891891892e-05,
      "loss": 1.138,
      "step": 444
    },
    {
      "epoch": 10.37,
      "learning_rate": 3.125e-05,
      "loss": 0.9824,
      "step": 445
    },
    {
      "epoch": 10.39,
      "learning_rate": 3.108108108108108e-05,
      "loss": 0.765,
      "step": 446
    },
    {
      "epoch": 10.42,
      "learning_rate": 3.091216216216216e-05,
      "loss": 0.6426,
      "step": 447
    },
    {
      "epoch": 10.44,
      "learning_rate": 3.074324324324325e-05,
      "loss": 0.2964,
      "step": 448
    },
    {
      "epoch": 10.46,
      "learning_rate": 3.057432432432433e-05,
      "loss": 0.0943,
      "step": 449
    },
    {
      "epoch": 10.49,
      "learning_rate": 3.0405405405405407e-05,
      "loss": 1.363,
      "step": 450
    },
    {
      "epoch": 10.51,
      "learning_rate": 3.0236486486486488e-05,
      "loss": 1.3897,
      "step": 451
    },
    {
      "epoch": 10.53,
      "learning_rate": 3.006756756756757e-05,
      "loss": 1.4514,
      "step": 452
    },
    {
      "epoch": 10.56,
      "learning_rate": 2.9898648648648653e-05,
      "loss": 1.2872,
      "step": 453
    },
    {
      "epoch": 10.58,
      "learning_rate": 2.9729729729729733e-05,
      "loss": 0.9454,
      "step": 454
    },
    {
      "epoch": 10.6,
      "learning_rate": 2.9560810810810814e-05,
      "loss": 0.9129,
      "step": 455
    },
    {
      "epoch": 10.63,
      "learning_rate": 2.9391891891891894e-05,
      "loss": 0.7195,
      "step": 456
    },
    {
      "epoch": 10.65,
      "learning_rate": 2.9222972972972972e-05,
      "loss": 0.4835,
      "step": 457
    },
    {
      "epoch": 10.67,
      "learning_rate": 2.9054054054054052e-05,
      "loss": 0.2545,
      "step": 458
    },
    {
      "epoch": 10.7,
      "learning_rate": 2.888513513513514e-05,
      "loss": 0.1068,
      "step": 459
    },
    {
      "epoch": 10.72,
      "learning_rate": 2.8716216216216217e-05,
      "loss": 1.3296,
      "step": 460
    },
    {
      "epoch": 10.74,
      "learning_rate": 2.8547297297297298e-05,
      "loss": 1.4269,
      "step": 461
    },
    {
      "epoch": 10.77,
      "learning_rate": 2.8378378378378378e-05,
      "loss": 1.4975,
      "step": 462
    },
    {
      "epoch": 10.79,
      "learning_rate": 2.820945945945946e-05,
      "loss": 1.3133,
      "step": 463
    },
    {
      "epoch": 10.81,
      "learning_rate": 2.8040540540540543e-05,
      "loss": 1.086,
      "step": 464
    },
    {
      "epoch": 10.83,
      "learning_rate": 2.7871621621621624e-05,
      "loss": 1.0501,
      "step": 465
    },
    {
      "epoch": 10.86,
      "learning_rate": 2.7702702702702704e-05,
      "loss": 0.6794,
      "step": 466
    },
    {
      "epoch": 10.88,
      "learning_rate": 2.7533783783783785e-05,
      "loss": 0.3583,
      "step": 467
    },
    {
      "epoch": 10.9,
      "learning_rate": 2.7364864864864865e-05,
      "loss": 0.1665,
      "step": 468
    },
    {
      "epoch": 10.93,
      "learning_rate": 2.719594594594595e-05,
      "loss": 0.1007,
      "step": 469
    },
    {
      "epoch": 10.95,
      "learning_rate": 2.702702702702703e-05,
      "loss": 1.1639,
      "step": 470
    },
    {
      "epoch": 10.97,
      "learning_rate": 2.685810810810811e-05,
      "loss": 0.996,
      "step": 471
    },
    {
      "epoch": 11.0,
      "learning_rate": 2.668918918918919e-05,
      "loss": 0.3746,
      "step": 472
    },
    {
      "epoch": 11.02,
      "learning_rate": 2.652027027027027e-05,
      "loss": 1.473,
      "step": 473
    },
    {
      "epoch": 11.04,
      "learning_rate": 2.635135135135135e-05,
      "loss": 1.4788,
      "step": 474
    },
    {
      "epoch": 11.07,
      "learning_rate": 2.6182432432432437e-05,
      "loss": 1.2421,
      "step": 475
    },
    {
      "epoch": 11.09,
      "learning_rate": 2.6013513513513514e-05,
      "loss": 0.9503,
      "step": 476
    },
    {
      "epoch": 11.11,
      "learning_rate": 2.5844594594594595e-05,
      "loss": 0.912,
      "step": 477
    },
    {
      "epoch": 11.14,
      "learning_rate": 2.5675675675675675e-05,
      "loss": 0.8364,
      "step": 478
    },
    {
      "epoch": 11.16,
      "learning_rate": 2.5506756756756756e-05,
      "loss": 0.5568,
      "step": 479
    },
    {
      "epoch": 11.18,
      "learning_rate": 2.533783783783784e-05,
      "loss": 0.3318,
      "step": 480
    },
    {
      "epoch": 11.21,
      "learning_rate": 2.516891891891892e-05,
      "loss": 0.1312,
      "step": 481
    },
    {
      "epoch": 11.23,
      "learning_rate": 2.5e-05,
      "loss": 0.075,
      "step": 482
    },
    {
      "epoch": 11.25,
      "learning_rate": 2.4831081081081082e-05,
      "loss": 1.4243,
      "step": 483
    },
    {
      "epoch": 11.28,
      "learning_rate": 2.4662162162162162e-05,
      "loss": 1.3901,
      "step": 484
    },
    {
      "epoch": 11.3,
      "learning_rate": 2.4493243243243243e-05,
      "loss": 1.2246,
      "step": 485
    },
    {
      "epoch": 11.32,
      "learning_rate": 2.4324324324324327e-05,
      "loss": 1.1818,
      "step": 486
    },
    {
      "epoch": 11.35,
      "learning_rate": 2.4155405405405408e-05,
      "loss": 1.0109,
      "step": 487
    },
    {
      "epoch": 11.37,
      "learning_rate": 2.398648648648649e-05,
      "loss": 0.8281,
      "step": 488
    },
    {
      "epoch": 11.39,
      "learning_rate": 2.381756756756757e-05,
      "loss": 0.6501,
      "step": 489
    },
    {
      "epoch": 11.42,
      "learning_rate": 2.364864864864865e-05,
      "loss": 0.4087,
      "step": 490
    },
    {
      "epoch": 11.44,
      "learning_rate": 2.347972972972973e-05,
      "loss": 0.1322,
      "step": 491
    },
    {
      "epoch": 11.46,
      "learning_rate": 2.331081081081081e-05,
      "loss": 0.0803,
      "step": 492
    },
    {
      "epoch": 11.49,
      "learning_rate": 2.314189189189189e-05,
      "loss": 1.3958,
      "step": 493
    },
    {
      "epoch": 11.51,
      "learning_rate": 2.2972972972972976e-05,
      "loss": 1.3419,
      "step": 494
    },
    {
      "epoch": 11.53,
      "learning_rate": 2.2804054054054056e-05,
      "loss": 1.3391,
      "step": 495
    },
    {
      "epoch": 11.56,
      "learning_rate": 2.2635135135135137e-05,
      "loss": 1.0946,
      "step": 496
    },
    {
      "epoch": 11.58,
      "learning_rate": 2.2466216216216218e-05,
      "loss": 0.9677,
      "step": 497
    },
    {
      "epoch": 11.6,
      "learning_rate": 2.2297297297297298e-05,
      "loss": 0.8767,
      "step": 498
    },
    {
      "epoch": 11.63,
      "learning_rate": 2.212837837837838e-05,
      "loss": 0.6602,
      "step": 499
    },
    {
      "epoch": 11.65,
      "learning_rate": 2.195945945945946e-05,
      "loss": 0.5327,
      "step": 500
    },
    {
      "epoch": 11.67,
      "learning_rate": 2.179054054054054e-05,
      "loss": 0.1465,
      "step": 501
    },
    {
      "epoch": 11.7,
      "learning_rate": 2.1621621621621624e-05,
      "loss": 0.0838,
      "step": 502
    },
    {
      "epoch": 11.72,
      "learning_rate": 2.1452702702702705e-05,
      "loss": 1.3506,
      "step": 503
    },
    {
      "epoch": 11.74,
      "learning_rate": 2.1283783783783785e-05,
      "loss": 1.4057,
      "step": 504
    },
    {
      "epoch": 11.77,
      "learning_rate": 2.1114864864864866e-05,
      "loss": 1.3849,
      "step": 505
    },
    {
      "epoch": 11.79,
      "learning_rate": 2.0945945945945947e-05,
      "loss": 1.1261,
      "step": 506
    },
    {
      "epoch": 11.81,
      "learning_rate": 2.0777027027027027e-05,
      "loss": 0.9476,
      "step": 507
    },
    {
      "epoch": 11.84,
      "learning_rate": 2.0608108108108108e-05,
      "loss": 0.7997,
      "step": 508
    },
    {
      "epoch": 11.86,
      "learning_rate": 2.043918918918919e-05,
      "loss": 0.5517,
      "step": 509
    },
    {
      "epoch": 11.88,
      "learning_rate": 2.0270270270270273e-05,
      "loss": 0.3328,
      "step": 510
    },
    {
      "epoch": 11.91,
      "learning_rate": 2.0101351351351353e-05,
      "loss": 0.1092,
      "step": 511
    },
    {
      "epoch": 11.93,
      "learning_rate": 1.9932432432432434e-05,
      "loss": 0.0769,
      "step": 512
    },
    {
      "epoch": 11.95,
      "learning_rate": 1.9763513513513515e-05,
      "loss": 1.2126,
      "step": 513
    },
    {
      "epoch": 11.98,
      "learning_rate": 1.9594594594594595e-05,
      "loss": 0.7629,
      "step": 514
    },
    {
      "epoch": 12.0,
      "learning_rate": 1.9425675675675676e-05,
      "loss": 0.2915,
      "step": 515
    },
    {
      "epoch": 12.02,
      "learning_rate": 1.9256756756756756e-05,
      "loss": 1.5412,
      "step": 516
    },
    {
      "epoch": 12.05,
      "learning_rate": 1.9087837837837837e-05,
      "loss": 1.3511,
      "step": 517
    },
    {
      "epoch": 12.07,
      "learning_rate": 1.891891891891892e-05,
      "loss": 1.2045,
      "step": 518
    },
    {
      "epoch": 12.09,
      "learning_rate": 1.8750000000000002e-05,
      "loss": 1.1388,
      "step": 519
    },
    {
      "epoch": 12.12,
      "learning_rate": 1.8581081081081082e-05,
      "loss": 0.9554,
      "step": 520
    },
    {
      "epoch": 12.14,
      "learning_rate": 1.8412162162162163e-05,
      "loss": 0.818,
      "step": 521
    },
    {
      "epoch": 12.16,
      "learning_rate": 1.8243243243243244e-05,
      "loss": 0.6028,
      "step": 522
    },
    {
      "epoch": 12.19,
      "learning_rate": 1.8074324324324328e-05,
      "loss": 0.3043,
      "step": 523
    },
    {
      "epoch": 12.21,
      "learning_rate": 1.7905405405405405e-05,
      "loss": 0.1182,
      "step": 524
    },
    {
      "epoch": 12.23,
      "learning_rate": 1.7736486486486486e-05,
      "loss": 0.0753,
      "step": 525
    },
    {
      "epoch": 12.26,
      "learning_rate": 1.756756756756757e-05,
      "loss": 1.5145,
      "step": 526
    },
    {
      "epoch": 12.28,
      "learning_rate": 1.739864864864865e-05,
      "loss": 1.4023,
      "step": 527
    },
    {
      "epoch": 12.3,
      "learning_rate": 1.722972972972973e-05,
      "loss": 1.2662,
      "step": 528
    },
    {
      "epoch": 12.33,
      "learning_rate": 1.706081081081081e-05,
      "loss": 0.8407,
      "step": 529
    },
    {
      "epoch": 12.35,
      "learning_rate": 1.6891891891891892e-05,
      "loss": 0.8093,
      "step": 530
    },
    {
      "epoch": 12.37,
      "learning_rate": 1.6722972972972976e-05,
      "loss": 0.6592,
      "step": 531
    },
    {
      "epoch": 12.4,
      "learning_rate": 1.6554054054054053e-05,
      "loss": 0.4537,
      "step": 532
    },
    {
      "epoch": 12.42,
      "learning_rate": 1.6385135135135134e-05,
      "loss": 0.1556,
      "step": 533
    },
    {
      "epoch": 12.44,
      "learning_rate": 1.6216216216216218e-05,
      "loss": 0.0614,
      "step": 534
    },
    {
      "epoch": 12.47,
      "learning_rate": 1.60472972972973e-05,
      "loss": 0.0796,
      "step": 535
    },
    {
      "epoch": 12.49,
      "learning_rate": 1.587837837837838e-05,
      "loss": 1.5708,
      "step": 536
    },
    {
      "epoch": 12.51,
      "learning_rate": 1.570945945945946e-05,
      "loss": 1.4246,
      "step": 537
    },
    {
      "epoch": 12.54,
      "learning_rate": 1.554054054054054e-05,
      "loss": 1.2018,
      "step": 538
    },
    {
      "epoch": 12.56,
      "learning_rate": 1.5371621621621625e-05,
      "loss": 0.9333,
      "step": 539
    },
    {
      "epoch": 12.58,
      "learning_rate": 1.5202702702702704e-05,
      "loss": 0.7467,
      "step": 540
    },
    {
      "epoch": 12.61,
      "learning_rate": 1.5033783783783784e-05,
      "loss": 0.7427,
      "step": 541
    },
    {
      "epoch": 12.63,
      "learning_rate": 1.4864864864864867e-05,
      "loss": 0.4841,
      "step": 542
    },
    {
      "epoch": 12.65,
      "learning_rate": 1.4695945945945947e-05,
      "loss": 0.2673,
      "step": 543
    },
    {
      "epoch": 12.68,
      "learning_rate": 1.4527027027027026e-05,
      "loss": 0.0821,
      "step": 544
    },
    {
      "epoch": 12.7,
      "learning_rate": 1.4358108108108108e-05,
      "loss": 0.0826,
      "step": 545
    },
    {
      "epoch": 12.72,
      "learning_rate": 1.4189189189189189e-05,
      "loss": 1.4956,
      "step": 546
    },
    {
      "epoch": 12.75,
      "learning_rate": 1.4020270270270271e-05,
      "loss": 1.3049,
      "step": 547
    },
    {
      "epoch": 12.77,
      "learning_rate": 1.3851351351351352e-05,
      "loss": 1.3169,
      "step": 548
    },
    {
      "epoch": 12.79,
      "learning_rate": 1.3682432432432433e-05,
      "loss": 1.1652,
      "step": 549
    },
    {
      "epoch": 12.82,
      "learning_rate": 1.3513513513513515e-05,
      "loss": 0.8853,
      "step": 550
    },
    {
      "epoch": 12.84,
      "learning_rate": 1.3344594594594596e-05,
      "loss": 0.7856,
      "step": 551
    },
    {
      "epoch": 12.86,
      "learning_rate": 1.3175675675675675e-05,
      "loss": 0.582,
      "step": 552
    },
    {
      "epoch": 12.89,
      "learning_rate": 1.3006756756756757e-05,
      "loss": 0.3285,
      "step": 553
    },
    {
      "epoch": 12.91,
      "learning_rate": 1.2837837837837838e-05,
      "loss": 0.1846,
      "step": 554
    },
    {
      "epoch": 12.93,
      "learning_rate": 1.266891891891892e-05,
      "loss": 0.0767,
      "step": 555
    },
    {
      "epoch": 12.96,
      "learning_rate": 1.25e-05,
      "loss": 1.1816,
      "step": 556
    },
    {
      "epoch": 12.98,
      "learning_rate": 1.2331081081081081e-05,
      "loss": 0.6397,
      "step": 557
    },
    {
      "epoch": 13.0,
      "learning_rate": 1.2162162162162164e-05,
      "loss": 0.3475,
      "step": 558
    },
    {
      "epoch": 13.03,
      "learning_rate": 1.1993243243243244e-05,
      "loss": 1.494,
      "step": 559
    },
    {
      "epoch": 13.05,
      "learning_rate": 1.1824324324324325e-05,
      "loss": 1.2451,
      "step": 560
    },
    {
      "epoch": 13.07,
      "learning_rate": 1.1655405405405405e-05,
      "loss": 1.1281,
      "step": 561
    },
    {
      "epoch": 13.1,
      "learning_rate": 1.1486486486486488e-05,
      "loss": 0.8359,
      "step": 562
    },
    {
      "epoch": 13.12,
      "learning_rate": 1.1317567567567568e-05,
      "loss": 0.8191,
      "step": 563
    },
    {
      "epoch": 13.14,
      "learning_rate": 1.1148648648648649e-05,
      "loss": 0.671,
      "step": 564
    },
    {
      "epoch": 13.17,
      "learning_rate": 1.097972972972973e-05,
      "loss": 0.3535,
      "step": 565
    },
    {
      "epoch": 13.19,
      "learning_rate": 1.0810810810810812e-05,
      "loss": 0.1521,
      "step": 566
    },
    {
      "epoch": 13.21,
      "learning_rate": 1.0641891891891893e-05,
      "loss": 0.0681,
      "step": 567
    },
    {
      "epoch": 13.23,
      "learning_rate": 1.0472972972972973e-05,
      "loss": 0.2038,
      "step": 568
    },
    {
      "epoch": 13.26,
      "learning_rate": 1.0304054054054054e-05,
      "loss": 1.4731,
      "step": 569
    },
    {
      "epoch": 13.28,
      "learning_rate": 1.0135135135135136e-05,
      "loss": 1.3288,
      "step": 570
    },
    {
      "epoch": 13.3,
      "learning_rate": 9.966216216216217e-06,
      "loss": 1.0323,
      "step": 571
    },
    {
      "epoch": 13.33,
      "learning_rate": 9.797297297297298e-06,
      "loss": 0.7761,
      "step": 572
    },
    {
      "epoch": 13.35,
      "learning_rate": 9.628378378378378e-06,
      "loss": 0.727,
      "step": 573
    },
    {
      "epoch": 13.37,
      "learning_rate": 9.45945945945946e-06,
      "loss": 0.6381,
      "step": 574
    },
    {
      "epoch": 13.4,
      "learning_rate": 9.290540540540541e-06,
      "loss": 0.4303,
      "step": 575
    },
    {
      "epoch": 13.42,
      "learning_rate": 9.121621621621622e-06,
      "loss": 0.229,
      "step": 576
    },
    {
      "epoch": 13.44,
      "learning_rate": 8.952702702702702e-06,
      "loss": 0.0785,
      "step": 577
    },
    {
      "epoch": 13.47,
      "learning_rate": 8.783783783783785e-06,
      "loss": 0.2095,
      "step": 578
    },
    {
      "epoch": 13.49,
      "learning_rate": 8.614864864864865e-06,
      "loss": 1.4983,
      "step": 579
    },
    {
      "epoch": 13.51,
      "learning_rate": 8.445945945945946e-06,
      "loss": 1.3521,
      "step": 580
    },
    {
      "epoch": 13.54,
      "learning_rate": 8.277027027027027e-06,
      "loss": 1.3146,
      "step": 581
    },
    {
      "epoch": 13.56,
      "learning_rate": 8.108108108108109e-06,
      "loss": 0.9123,
      "step": 582
    },
    {
      "epoch": 13.58,
      "learning_rate": 7.93918918918919e-06,
      "loss": 0.7487,
      "step": 583
    },
    {
      "epoch": 13.61,
      "learning_rate": 7.77027027027027e-06,
      "loss": 0.4495,
      "step": 584
    },
    {
      "epoch": 13.63,
      "learning_rate": 7.601351351351352e-06,
      "loss": 0.3396,
      "step": 585
    },
    {
      "epoch": 13.65,
      "learning_rate": 7.432432432432433e-06,
      "loss": 0.1205,
      "step": 586
    },
    {
      "epoch": 13.68,
      "learning_rate": 7.263513513513513e-06,
      "loss": 0.0633,
      "step": 587
    },
    {
      "epoch": 13.7,
      "learning_rate": 7.0945945945945946e-06,
      "loss": 0.1904,
      "step": 588
    },
    {
      "epoch": 13.72,
      "learning_rate": 6.925675675675676e-06,
      "loss": 1.4863,
      "step": 589
    },
    {
      "epoch": 13.75,
      "learning_rate": 6.7567567567567575e-06,
      "loss": 1.4163,
      "step": 590
    },
    {
      "epoch": 13.77,
      "learning_rate": 6.587837837837837e-06,
      "loss": 1.3777,
      "step": 591
    },
    {
      "epoch": 13.79,
      "learning_rate": 6.418918918918919e-06,
      "loss": 1.0749,
      "step": 592
    },
    {
      "epoch": 13.82,
      "learning_rate": 6.25e-06,
      "loss": 0.9292,
      "step": 593
    },
    {
      "epoch": 13.84,
      "learning_rate": 6.081081081081082e-06,
      "loss": 0.7355,
      "step": 594
    },
    {
      "epoch": 13.86,
      "learning_rate": 5.912162162162162e-06,
      "loss": 0.6468,
      "step": 595
    },
    {
      "epoch": 13.89,
      "learning_rate": 5.743243243243244e-06,
      "loss": 0.3443,
      "step": 596
    },
    {
      "epoch": 13.91,
      "learning_rate": 5.5743243243243245e-06,
      "loss": 0.1483,
      "step": 597
    },
    {
      "epoch": 13.93,
      "learning_rate": 5.405405405405406e-06,
      "loss": 0.1853,
      "step": 598
    },
    {
      "epoch": 13.96,
      "learning_rate": 5.236486486486487e-06,
      "loss": 1.3186,
      "step": 599
    },
    {
      "epoch": 13.98,
      "learning_rate": 5.067567567567568e-06,
      "loss": 0.6906,
      "step": 600
    },
    {
      "epoch": 14.0,
      "learning_rate": 4.898648648648649e-06,
      "loss": 0.4546,
      "step": 601
    },
    {
      "epoch": 14.03,
      "learning_rate": 4.72972972972973e-06,
      "loss": 1.535,
      "step": 602
    },
    {
      "epoch": 14.05,
      "learning_rate": 4.560810810810811e-06,
      "loss": 1.4158,
      "step": 603
    },
    {
      "epoch": 14.07,
      "learning_rate": 4.391891891891892e-06,
      "loss": 0.9301,
      "step": 604
    },
    {
      "epoch": 14.1,
      "learning_rate": 4.222972972972973e-06,
      "loss": 0.9221,
      "step": 605
    },
    {
      "epoch": 14.12,
      "learning_rate": 4.0540540540540545e-06,
      "loss": 0.6948,
      "step": 606
    },
    {
      "epoch": 14.14,
      "learning_rate": 3.885135135135135e-06,
      "loss": 0.6297,
      "step": 607
    },
    {
      "epoch": 14.17,
      "learning_rate": 3.7162162162162166e-06,
      "loss": 0.3427,
      "step": 608
    },
    {
      "epoch": 14.19,
      "learning_rate": 3.5472972972972973e-06,
      "loss": 0.1821,
      "step": 609
    },
    {
      "epoch": 14.21,
      "learning_rate": 3.3783783783783788e-06,
      "loss": 0.0664,
      "step": 610
    },
    {
      "epoch": 14.24,
      "learning_rate": 3.2094594594594594e-06,
      "loss": 0.2936,
      "step": 611
    },
    {
      "epoch": 14.26,
      "learning_rate": 3.040540540540541e-06,
      "loss": 1.463,
      "step": 612
    },
    {
      "epoch": 14.28,
      "learning_rate": 2.871621621621622e-06,
      "loss": 1.3162,
      "step": 613
    },
    {
      "epoch": 14.31,
      "learning_rate": 2.702702702702703e-06,
      "loss": 1.1814,
      "step": 614
    },
    {
      "epoch": 14.33,
      "learning_rate": 2.533783783783784e-06,
      "loss": 1.0283,
      "step": 615
    },
    {
      "epoch": 14.35,
      "learning_rate": 2.364864864864865e-06,
      "loss": 0.7701,
      "step": 616
    },
    {
      "epoch": 14.38,
      "learning_rate": 2.195945945945946e-06,
      "loss": 0.6445,
      "step": 617
    },
    {
      "epoch": 14.4,
      "learning_rate": 2.0270270270270273e-06,
      "loss": 0.5021,
      "step": 618
    },
    {
      "epoch": 14.42,
      "learning_rate": 1.8581081081081083e-06,
      "loss": 0.2472,
      "step": 619
    },
    {
      "epoch": 14.45,
      "learning_rate": 1.6891891891891894e-06,
      "loss": 0.0763,
      "step": 620
    },
    {
      "epoch": 14.47,
      "learning_rate": 1.5202702702702704e-06,
      "loss": 0.2919,
      "step": 621
    },
    {
      "epoch": 14.49,
      "learning_rate": 1.3513513513513515e-06,
      "loss": 1.4695,
      "step": 622
    },
    {
      "epoch": 14.52,
      "learning_rate": 1.1824324324324326e-06,
      "loss": 1.2762,
      "step": 623
    },
    {
      "epoch": 14.54,
      "learning_rate": 1.0135135135135136e-06,
      "loss": 1.1309,
      "step": 624
    },
    {
      "epoch": 14.56,
      "learning_rate": 8.445945945945947e-07,
      "loss": 0.7696,
      "step": 625
    },
    {
      "epoch": 14.59,
      "learning_rate": 6.756756756756758e-07,
      "loss": 0.6515,
      "step": 626
    },
    {
      "epoch": 14.61,
      "learning_rate": 5.067567567567568e-07,
      "loss": 0.5216,
      "step": 627
    },
    {
      "epoch": 14.63,
      "learning_rate": 3.378378378378379e-07,
      "loss": 0.3135,
      "step": 628
    },
    {
      "epoch": 14.66,
      "learning_rate": 1.6891891891891894e-07,
      "loss": 0.1084,
      "step": 629
    },
    {
      "epoch": 14.68,
      "learning_rate": 0.0,
      "loss": 0.0578,
      "step": 630
    },
    {
      "epoch": 14.68,
      "step": 630,
      "total_flos": 2.281331051123835e+17,
      "train_loss": 1.2019227477174903,
      "train_runtime": 20058.3189,
      "train_samples_per_second": 0.77,
      "train_steps_per_second": 0.031
    }
  ],
  "max_steps": 630,
  "num_train_epochs": 15,
  "total_flos": 2.281331051123835e+17,
  "trial_name": null,
  "trial_params": null
}