ViPer / VPE2 /trainer_state.json
miaw1419's picture
Upload 8 files
4e8a3d9 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.03918367346938775,
"eval_steps": 50,
"global_step": 1200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 3.265306122448979e-05,
"grad_norm": 12.553070068359375,
"learning_rate": 2.0000000000000002e-07,
"loss": 2.3591,
"step": 1
},
{
"epoch": 6.530612244897959e-05,
"grad_norm": 12.028986930847168,
"learning_rate": 4.0000000000000003e-07,
"loss": 2.3144,
"step": 2
},
{
"epoch": 9.79591836734694e-05,
"grad_norm": 12.335774421691895,
"learning_rate": 6.000000000000001e-07,
"loss": 2.3323,
"step": 3
},
{
"epoch": 0.00013061224489795917,
"grad_norm": 12.695511817932129,
"learning_rate": 8.000000000000001e-07,
"loss": 2.3651,
"step": 4
},
{
"epoch": 0.00016326530612244898,
"grad_norm": 12.702936172485352,
"learning_rate": 1.0000000000000002e-06,
"loss": 2.3563,
"step": 5
},
{
"epoch": 0.0001959183673469388,
"grad_norm": 11.689526557922363,
"learning_rate": 1.2000000000000002e-06,
"loss": 2.3263,
"step": 6
},
{
"epoch": 0.00022857142857142857,
"grad_norm": 11.332144737243652,
"learning_rate": 1.4000000000000001e-06,
"loss": 2.2861,
"step": 7
},
{
"epoch": 0.00026122448979591835,
"grad_norm": 11.530237197875977,
"learning_rate": 1.6000000000000001e-06,
"loss": 2.2867,
"step": 8
},
{
"epoch": 0.0002938775510204082,
"grad_norm": 10.193593978881836,
"learning_rate": 1.8000000000000001e-06,
"loss": 2.2521,
"step": 9
},
{
"epoch": 0.00032653061224489796,
"grad_norm": 9.98508071899414,
"learning_rate": 2.0000000000000003e-06,
"loss": 2.249,
"step": 10
},
{
"epoch": 0.00035918367346938774,
"grad_norm": 9.584476470947266,
"learning_rate": 2.2e-06,
"loss": 2.2116,
"step": 11
},
{
"epoch": 0.0003918367346938776,
"grad_norm": 9.574864387512207,
"learning_rate": 2.4000000000000003e-06,
"loss": 2.2343,
"step": 12
},
{
"epoch": 0.00042448979591836735,
"grad_norm": 8.929769515991211,
"learning_rate": 2.6e-06,
"loss": 2.1599,
"step": 13
},
{
"epoch": 0.00045714285714285713,
"grad_norm": 8.262470245361328,
"learning_rate": 2.8000000000000003e-06,
"loss": 2.1152,
"step": 14
},
{
"epoch": 0.0004897959183673469,
"grad_norm": 8.28957748413086,
"learning_rate": 3e-06,
"loss": 2.1048,
"step": 15
},
{
"epoch": 0.0005224489795918367,
"grad_norm": 8.160731315612793,
"learning_rate": 3.2000000000000003e-06,
"loss": 2.054,
"step": 16
},
{
"epoch": 0.0005551020408163266,
"grad_norm": 7.5022077560424805,
"learning_rate": 3.4000000000000005e-06,
"loss": 2.0559,
"step": 17
},
{
"epoch": 0.0005877551020408164,
"grad_norm": 7.718921661376953,
"learning_rate": 3.6000000000000003e-06,
"loss": 2.003,
"step": 18
},
{
"epoch": 0.0006204081632653061,
"grad_norm": 7.6420392990112305,
"learning_rate": 3.8000000000000005e-06,
"loss": 1.9646,
"step": 19
},
{
"epoch": 0.0006530612244897959,
"grad_norm": 7.363818645477295,
"learning_rate": 4.000000000000001e-06,
"loss": 1.918,
"step": 20
},
{
"epoch": 0.0006857142857142857,
"grad_norm": 7.367091178894043,
"learning_rate": 4.2000000000000004e-06,
"loss": 1.9065,
"step": 21
},
{
"epoch": 0.0007183673469387755,
"grad_norm": 7.201758861541748,
"learning_rate": 4.4e-06,
"loss": 1.8989,
"step": 22
},
{
"epoch": 0.0007510204081632653,
"grad_norm": 7.093530654907227,
"learning_rate": 4.600000000000001e-06,
"loss": 1.8381,
"step": 23
},
{
"epoch": 0.0007836734693877551,
"grad_norm": 7.22179651260376,
"learning_rate": 4.800000000000001e-06,
"loss": 1.7959,
"step": 24
},
{
"epoch": 0.0008163265306122449,
"grad_norm": 7.264500141143799,
"learning_rate": 5e-06,
"loss": 1.7654,
"step": 25
},
{
"epoch": 0.0008489795918367347,
"grad_norm": 6.931406021118164,
"learning_rate": 5.2e-06,
"loss": 1.6971,
"step": 26
},
{
"epoch": 0.0008816326530612245,
"grad_norm": 6.8040547370910645,
"learning_rate": 5.400000000000001e-06,
"loss": 1.6738,
"step": 27
},
{
"epoch": 0.0009142857142857143,
"grad_norm": 7.414416313171387,
"learning_rate": 5.600000000000001e-06,
"loss": 1.6131,
"step": 28
},
{
"epoch": 0.000946938775510204,
"grad_norm": 8.00181770324707,
"learning_rate": 5.8e-06,
"loss": 1.5743,
"step": 29
},
{
"epoch": 0.0009795918367346938,
"grad_norm": 8.280529022216797,
"learning_rate": 6e-06,
"loss": 1.5139,
"step": 30
},
{
"epoch": 0.0010122448979591836,
"grad_norm": 24.16852378845215,
"learning_rate": 6.200000000000001e-06,
"loss": 1.4998,
"step": 31
},
{
"epoch": 0.0010448979591836734,
"grad_norm": 25.735715866088867,
"learning_rate": 6.4000000000000006e-06,
"loss": 1.4774,
"step": 32
},
{
"epoch": 0.0010775510204081632,
"grad_norm": 18.589033126831055,
"learning_rate": 6.600000000000001e-06,
"loss": 1.3795,
"step": 33
},
{
"epoch": 0.0011102040816326532,
"grad_norm": 7.6697096824646,
"learning_rate": 6.800000000000001e-06,
"loss": 1.3513,
"step": 34
},
{
"epoch": 0.001142857142857143,
"grad_norm": 12.835890769958496,
"learning_rate": 7e-06,
"loss": 1.2908,
"step": 35
},
{
"epoch": 0.0011755102040816327,
"grad_norm": 12.289154052734375,
"learning_rate": 7.2000000000000005e-06,
"loss": 1.2354,
"step": 36
},
{
"epoch": 0.0012081632653061225,
"grad_norm": 5.951991558074951,
"learning_rate": 7.4e-06,
"loss": 1.2242,
"step": 37
},
{
"epoch": 0.0012408163265306123,
"grad_norm": 5.6563825607299805,
"learning_rate": 7.600000000000001e-06,
"loss": 1.2448,
"step": 38
},
{
"epoch": 0.001273469387755102,
"grad_norm": 5.884244441986084,
"learning_rate": 7.800000000000002e-06,
"loss": 1.2011,
"step": 39
},
{
"epoch": 0.0013061224489795918,
"grad_norm": 4.279225826263428,
"learning_rate": 8.000000000000001e-06,
"loss": 1.1616,
"step": 40
},
{
"epoch": 0.0013387755102040816,
"grad_norm": 4.273413181304932,
"learning_rate": 8.2e-06,
"loss": 1.1361,
"step": 41
},
{
"epoch": 0.0013714285714285714,
"grad_norm": 4.514435768127441,
"learning_rate": 8.400000000000001e-06,
"loss": 1.1194,
"step": 42
},
{
"epoch": 0.0014040816326530612,
"grad_norm": 3.1108927726745605,
"learning_rate": 8.6e-06,
"loss": 1.1359,
"step": 43
},
{
"epoch": 0.001436734693877551,
"grad_norm": 3.187659740447998,
"learning_rate": 8.8e-06,
"loss": 1.1088,
"step": 44
},
{
"epoch": 0.0014693877551020407,
"grad_norm": 3.0588572025299072,
"learning_rate": 9e-06,
"loss": 1.0367,
"step": 45
},
{
"epoch": 0.0015020408163265305,
"grad_norm": 3.0041353702545166,
"learning_rate": 9.200000000000002e-06,
"loss": 1.065,
"step": 46
},
{
"epoch": 0.0015346938775510203,
"grad_norm": 3.0067152976989746,
"learning_rate": 9.4e-06,
"loss": 1.106,
"step": 47
},
{
"epoch": 0.0015673469387755103,
"grad_norm": 2.8196375370025635,
"learning_rate": 9.600000000000001e-06,
"loss": 1.0463,
"step": 48
},
{
"epoch": 0.0016,
"grad_norm": 2.7588930130004883,
"learning_rate": 9.800000000000001e-06,
"loss": 1.0707,
"step": 49
},
{
"epoch": 0.0016326530612244899,
"grad_norm": 2.776578903198242,
"learning_rate": 1e-05,
"loss": 1.057,
"step": 50
},
{
"epoch": 0.0016326530612244899,
"eval_loss": 1.0555493831634521,
"eval_runtime": 80.9919,
"eval_samples_per_second": 1.235,
"eval_steps_per_second": 1.235,
"step": 50
},
{
"epoch": 0.0016653061224489796,
"grad_norm": 3.0007307529449463,
"learning_rate": 9.999967341606794e-06,
"loss": 1.0359,
"step": 51
},
{
"epoch": 0.0016979591836734694,
"grad_norm": 2.820908308029175,
"learning_rate": 9.999934683213587e-06,
"loss": 1.0659,
"step": 52
},
{
"epoch": 0.0017306122448979592,
"grad_norm": 3.070078134536743,
"learning_rate": 9.999902024820379e-06,
"loss": 1.0676,
"step": 53
},
{
"epoch": 0.001763265306122449,
"grad_norm": 2.826664686203003,
"learning_rate": 9.999869366427172e-06,
"loss": 1.0396,
"step": 54
},
{
"epoch": 0.0017959183673469388,
"grad_norm": 2.8928892612457275,
"learning_rate": 9.999836708033965e-06,
"loss": 1.0043,
"step": 55
},
{
"epoch": 0.0018285714285714285,
"grad_norm": 2.9964358806610107,
"learning_rate": 9.999804049640759e-06,
"loss": 1.0335,
"step": 56
},
{
"epoch": 0.0018612244897959183,
"grad_norm": 2.796006202697754,
"learning_rate": 9.999771391247552e-06,
"loss": 0.9906,
"step": 57
},
{
"epoch": 0.001893877551020408,
"grad_norm": 2.638707399368286,
"learning_rate": 9.999738732854345e-06,
"loss": 1.0347,
"step": 58
},
{
"epoch": 0.0019265306122448979,
"grad_norm": 2.6315219402313232,
"learning_rate": 9.999706074461137e-06,
"loss": 1.0213,
"step": 59
},
{
"epoch": 0.0019591836734693877,
"grad_norm": 2.963063955307007,
"learning_rate": 9.99967341606793e-06,
"loss": 0.9669,
"step": 60
},
{
"epoch": 0.0019918367346938777,
"grad_norm": 2.913827657699585,
"learning_rate": 9.999640757674723e-06,
"loss": 1.013,
"step": 61
},
{
"epoch": 0.002024489795918367,
"grad_norm": 2.830906391143799,
"learning_rate": 9.999608099281516e-06,
"loss": 1.0036,
"step": 62
},
{
"epoch": 0.002057142857142857,
"grad_norm": 2.856300115585327,
"learning_rate": 9.99957544088831e-06,
"loss": 0.9806,
"step": 63
},
{
"epoch": 0.0020897959183673468,
"grad_norm": 2.8459157943725586,
"learning_rate": 9.999542782495103e-06,
"loss": 1.0253,
"step": 64
},
{
"epoch": 0.0021224489795918368,
"grad_norm": 2.748532772064209,
"learning_rate": 9.999510124101896e-06,
"loss": 1.0017,
"step": 65
},
{
"epoch": 0.0021551020408163263,
"grad_norm": 2.910276412963867,
"learning_rate": 9.999477465708688e-06,
"loss": 0.978,
"step": 66
},
{
"epoch": 0.0021877551020408163,
"grad_norm": 2.8665144443511963,
"learning_rate": 9.99944480731548e-06,
"loss": 0.9914,
"step": 67
},
{
"epoch": 0.0022204081632653063,
"grad_norm": 2.6549315452575684,
"learning_rate": 9.999412148922274e-06,
"loss": 0.9735,
"step": 68
},
{
"epoch": 0.002253061224489796,
"grad_norm": 2.7277936935424805,
"learning_rate": 9.999379490529067e-06,
"loss": 0.9498,
"step": 69
},
{
"epoch": 0.002285714285714286,
"grad_norm": 2.849747657775879,
"learning_rate": 9.99934683213586e-06,
"loss": 0.9976,
"step": 70
},
{
"epoch": 0.0023183673469387754,
"grad_norm": 2.8592312335968018,
"learning_rate": 9.999314173742652e-06,
"loss": 0.9523,
"step": 71
},
{
"epoch": 0.0023510204081632654,
"grad_norm": 2.7041122913360596,
"learning_rate": 9.999281515349445e-06,
"loss": 0.9692,
"step": 72
},
{
"epoch": 0.002383673469387755,
"grad_norm": 2.717122793197632,
"learning_rate": 9.999248856956238e-06,
"loss": 0.9491,
"step": 73
},
{
"epoch": 0.002416326530612245,
"grad_norm": 2.516847610473633,
"learning_rate": 9.999216198563032e-06,
"loss": 0.9632,
"step": 74
},
{
"epoch": 0.0024489795918367346,
"grad_norm": 2.766266107559204,
"learning_rate": 9.999183540169825e-06,
"loss": 0.932,
"step": 75
},
{
"epoch": 0.0024816326530612246,
"grad_norm": 2.8596270084381104,
"learning_rate": 9.999150881776616e-06,
"loss": 0.9479,
"step": 76
},
{
"epoch": 0.002514285714285714,
"grad_norm": 2.6034138202667236,
"learning_rate": 9.99911822338341e-06,
"loss": 0.9712,
"step": 77
},
{
"epoch": 0.002546938775510204,
"grad_norm": 2.662513256072998,
"learning_rate": 9.999085564990203e-06,
"loss": 0.9882,
"step": 78
},
{
"epoch": 0.0025795918367346937,
"grad_norm": 2.900061845779419,
"learning_rate": 9.999052906596996e-06,
"loss": 0.9232,
"step": 79
},
{
"epoch": 0.0026122448979591837,
"grad_norm": 2.7503371238708496,
"learning_rate": 9.99902024820379e-06,
"loss": 0.9483,
"step": 80
},
{
"epoch": 0.0026448979591836737,
"grad_norm": 2.61838960647583,
"learning_rate": 9.998987589810583e-06,
"loss": 0.9104,
"step": 81
},
{
"epoch": 0.0026775510204081632,
"grad_norm": 2.775689125061035,
"learning_rate": 9.998954931417376e-06,
"loss": 0.9145,
"step": 82
},
{
"epoch": 0.0027102040816326532,
"grad_norm": 2.695878744125366,
"learning_rate": 9.998922273024167e-06,
"loss": 0.9075,
"step": 83
},
{
"epoch": 0.002742857142857143,
"grad_norm": 2.748655080795288,
"learning_rate": 9.99888961463096e-06,
"loss": 0.966,
"step": 84
},
{
"epoch": 0.002775510204081633,
"grad_norm": 2.5372986793518066,
"learning_rate": 9.998856956237754e-06,
"loss": 0.9541,
"step": 85
},
{
"epoch": 0.0028081632653061224,
"grad_norm": 2.6666107177734375,
"learning_rate": 9.998824297844547e-06,
"loss": 0.9678,
"step": 86
},
{
"epoch": 0.0028408163265306124,
"grad_norm": 2.450654983520508,
"learning_rate": 9.99879163945134e-06,
"loss": 0.9404,
"step": 87
},
{
"epoch": 0.002873469387755102,
"grad_norm": 2.6407573223114014,
"learning_rate": 9.998758981058134e-06,
"loss": 0.9661,
"step": 88
},
{
"epoch": 0.002906122448979592,
"grad_norm": 2.589488983154297,
"learning_rate": 9.998726322664925e-06,
"loss": 0.9528,
"step": 89
},
{
"epoch": 0.0029387755102040815,
"grad_norm": 3.260467767715454,
"learning_rate": 9.998693664271718e-06,
"loss": 0.9625,
"step": 90
},
{
"epoch": 0.0029714285714285715,
"grad_norm": 2.8341681957244873,
"learning_rate": 9.998661005878512e-06,
"loss": 0.9088,
"step": 91
},
{
"epoch": 0.003004081632653061,
"grad_norm": 3.00934100151062,
"learning_rate": 9.998628347485305e-06,
"loss": 0.9148,
"step": 92
},
{
"epoch": 0.003036734693877551,
"grad_norm": 2.8187415599823,
"learning_rate": 9.998595689092098e-06,
"loss": 0.9028,
"step": 93
},
{
"epoch": 0.0030693877551020406,
"grad_norm": 2.617946147918701,
"learning_rate": 9.99856303069889e-06,
"loss": 0.8886,
"step": 94
},
{
"epoch": 0.0031020408163265306,
"grad_norm": 2.7720468044281006,
"learning_rate": 9.998530372305683e-06,
"loss": 0.9417,
"step": 95
},
{
"epoch": 0.0031346938775510206,
"grad_norm": 2.8178319931030273,
"learning_rate": 9.998497713912476e-06,
"loss": 0.9012,
"step": 96
},
{
"epoch": 0.00316734693877551,
"grad_norm": 2.9068961143493652,
"learning_rate": 9.99846505551927e-06,
"loss": 0.9318,
"step": 97
},
{
"epoch": 0.0032,
"grad_norm": 3.227482557296753,
"learning_rate": 9.998432397126063e-06,
"loss": 0.8956,
"step": 98
},
{
"epoch": 0.0032326530612244897,
"grad_norm": 2.8666388988494873,
"learning_rate": 9.998399738732854e-06,
"loss": 0.9288,
"step": 99
},
{
"epoch": 0.0032653061224489797,
"grad_norm": 2.861967086791992,
"learning_rate": 9.998367080339647e-06,
"loss": 0.8923,
"step": 100
},
{
"epoch": 0.0032653061224489797,
"eval_loss": 0.9108777046203613,
"eval_runtime": 73.7729,
"eval_samples_per_second": 1.356,
"eval_steps_per_second": 1.356,
"step": 100
},
{
"epoch": 0.0032979591836734693,
"grad_norm": 3.026766777038574,
"learning_rate": 9.99833442194644e-06,
"loss": 0.9085,
"step": 101
},
{
"epoch": 0.0033306122448979593,
"grad_norm": 3.0358951091766357,
"learning_rate": 9.998301763553234e-06,
"loss": 0.9019,
"step": 102
},
{
"epoch": 0.003363265306122449,
"grad_norm": 2.7646968364715576,
"learning_rate": 9.998269105160027e-06,
"loss": 0.8954,
"step": 103
},
{
"epoch": 0.003395918367346939,
"grad_norm": 2.880887746810913,
"learning_rate": 9.99823644676682e-06,
"loss": 0.9257,
"step": 104
},
{
"epoch": 0.0034285714285714284,
"grad_norm": 3.1517140865325928,
"learning_rate": 9.998203788373614e-06,
"loss": 0.8951,
"step": 105
},
{
"epoch": 0.0034612244897959184,
"grad_norm": 2.7021565437316895,
"learning_rate": 9.998171129980407e-06,
"loss": 0.9521,
"step": 106
},
{
"epoch": 0.003493877551020408,
"grad_norm": 2.860952854156494,
"learning_rate": 9.998138471587198e-06,
"loss": 0.902,
"step": 107
},
{
"epoch": 0.003526530612244898,
"grad_norm": 3.0253970623016357,
"learning_rate": 9.998105813193992e-06,
"loss": 0.8935,
"step": 108
},
{
"epoch": 0.003559183673469388,
"grad_norm": 2.776489734649658,
"learning_rate": 9.998073154800785e-06,
"loss": 0.9323,
"step": 109
},
{
"epoch": 0.0035918367346938775,
"grad_norm": 2.3988196849823,
"learning_rate": 9.998040496407578e-06,
"loss": 0.8951,
"step": 110
},
{
"epoch": 0.0036244897959183675,
"grad_norm": 2.6600584983825684,
"learning_rate": 9.998007838014371e-06,
"loss": 0.8913,
"step": 111
},
{
"epoch": 0.003657142857142857,
"grad_norm": 2.4915781021118164,
"learning_rate": 9.997975179621163e-06,
"loss": 0.8814,
"step": 112
},
{
"epoch": 0.003689795918367347,
"grad_norm": 2.7426276206970215,
"learning_rate": 9.997942521227956e-06,
"loss": 0.9027,
"step": 113
},
{
"epoch": 0.0037224489795918366,
"grad_norm": 2.80908465385437,
"learning_rate": 9.99790986283475e-06,
"loss": 0.9033,
"step": 114
},
{
"epoch": 0.0037551020408163266,
"grad_norm": 2.5127768516540527,
"learning_rate": 9.997877204441543e-06,
"loss": 0.8886,
"step": 115
},
{
"epoch": 0.003787755102040816,
"grad_norm": 2.7205052375793457,
"learning_rate": 9.997844546048336e-06,
"loss": 0.9199,
"step": 116
},
{
"epoch": 0.003820408163265306,
"grad_norm": 2.6644845008850098,
"learning_rate": 9.997811887655127e-06,
"loss": 0.8722,
"step": 117
},
{
"epoch": 0.0038530612244897957,
"grad_norm": 3.05825138092041,
"learning_rate": 9.99777922926192e-06,
"loss": 0.8926,
"step": 118
},
{
"epoch": 0.0038857142857142857,
"grad_norm": 2.8610692024230957,
"learning_rate": 9.997746570868714e-06,
"loss": 0.886,
"step": 119
},
{
"epoch": 0.003918367346938775,
"grad_norm": 2.8283677101135254,
"learning_rate": 9.997713912475507e-06,
"loss": 0.8838,
"step": 120
},
{
"epoch": 0.003951020408163266,
"grad_norm": 3.36906361579895,
"learning_rate": 9.9976812540823e-06,
"loss": 0.8614,
"step": 121
},
{
"epoch": 0.003983673469387755,
"grad_norm": 2.949343681335449,
"learning_rate": 9.997648595689093e-06,
"loss": 0.8915,
"step": 122
},
{
"epoch": 0.004016326530612245,
"grad_norm": 2.986492156982422,
"learning_rate": 9.997615937295885e-06,
"loss": 0.9053,
"step": 123
},
{
"epoch": 0.004048979591836734,
"grad_norm": 2.7720727920532227,
"learning_rate": 9.997583278902678e-06,
"loss": 0.8915,
"step": 124
},
{
"epoch": 0.004081632653061225,
"grad_norm": 2.5769472122192383,
"learning_rate": 9.997550620509471e-06,
"loss": 0.8845,
"step": 125
},
{
"epoch": 0.004114285714285714,
"grad_norm": 2.9634106159210205,
"learning_rate": 9.997517962116265e-06,
"loss": 0.9057,
"step": 126
},
{
"epoch": 0.004146938775510204,
"grad_norm": 2.6103193759918213,
"learning_rate": 9.997485303723058e-06,
"loss": 0.8843,
"step": 127
},
{
"epoch": 0.0041795918367346935,
"grad_norm": 2.892089366912842,
"learning_rate": 9.997452645329851e-06,
"loss": 0.8875,
"step": 128
},
{
"epoch": 0.004212244897959184,
"grad_norm": 3.0502076148986816,
"learning_rate": 9.997419986936644e-06,
"loss": 0.8654,
"step": 129
},
{
"epoch": 0.0042448979591836735,
"grad_norm": 2.968538522720337,
"learning_rate": 9.997387328543436e-06,
"loss": 0.9211,
"step": 130
},
{
"epoch": 0.004277551020408163,
"grad_norm": 2.7077767848968506,
"learning_rate": 9.99735467015023e-06,
"loss": 0.849,
"step": 131
},
{
"epoch": 0.004310204081632653,
"grad_norm": 2.8962769508361816,
"learning_rate": 9.997322011757022e-06,
"loss": 0.8844,
"step": 132
},
{
"epoch": 0.004342857142857143,
"grad_norm": 2.5692780017852783,
"learning_rate": 9.997289353363816e-06,
"loss": 0.8494,
"step": 133
},
{
"epoch": 0.004375510204081633,
"grad_norm": 2.603320837020874,
"learning_rate": 9.997256694970609e-06,
"loss": 0.8994,
"step": 134
},
{
"epoch": 0.004408163265306122,
"grad_norm": 2.762920618057251,
"learning_rate": 9.9972240365774e-06,
"loss": 0.8894,
"step": 135
},
{
"epoch": 0.004440816326530613,
"grad_norm": 2.7908272743225098,
"learning_rate": 9.997191378184194e-06,
"loss": 0.888,
"step": 136
},
{
"epoch": 0.004473469387755102,
"grad_norm": 2.405191421508789,
"learning_rate": 9.997158719790987e-06,
"loss": 0.8953,
"step": 137
},
{
"epoch": 0.004506122448979592,
"grad_norm": 3.086392879486084,
"learning_rate": 9.99712606139778e-06,
"loss": 0.9024,
"step": 138
},
{
"epoch": 0.004538775510204081,
"grad_norm": 3.10508394241333,
"learning_rate": 9.997093403004573e-06,
"loss": 0.8636,
"step": 139
},
{
"epoch": 0.004571428571428572,
"grad_norm": 2.7642314434051514,
"learning_rate": 9.997060744611365e-06,
"loss": 0.8589,
"step": 140
},
{
"epoch": 0.004604081632653061,
"grad_norm": 2.9966800212860107,
"learning_rate": 9.997028086218158e-06,
"loss": 0.8564,
"step": 141
},
{
"epoch": 0.004636734693877551,
"grad_norm": 2.998603582382202,
"learning_rate": 9.996995427824951e-06,
"loss": 0.8596,
"step": 142
},
{
"epoch": 0.0046693877551020405,
"grad_norm": 2.815833330154419,
"learning_rate": 9.996962769431745e-06,
"loss": 0.9006,
"step": 143
},
{
"epoch": 0.004702040816326531,
"grad_norm": 2.654766798019409,
"learning_rate": 9.996930111038538e-06,
"loss": 0.8683,
"step": 144
},
{
"epoch": 0.0047346938775510205,
"grad_norm": 2.9405479431152344,
"learning_rate": 9.996897452645331e-06,
"loss": 0.8756,
"step": 145
},
{
"epoch": 0.00476734693877551,
"grad_norm": 3.789085626602173,
"learning_rate": 9.996864794252124e-06,
"loss": 0.8991,
"step": 146
},
{
"epoch": 0.0048,
"grad_norm": 3.062678813934326,
"learning_rate": 9.996832135858918e-06,
"loss": 0.8461,
"step": 147
},
{
"epoch": 0.00483265306122449,
"grad_norm": 2.656879425048828,
"learning_rate": 9.996799477465709e-06,
"loss": 0.8755,
"step": 148
},
{
"epoch": 0.00486530612244898,
"grad_norm": 2.66681170463562,
"learning_rate": 9.996766819072502e-06,
"loss": 0.8882,
"step": 149
},
{
"epoch": 0.004897959183673469,
"grad_norm": 2.69744873046875,
"learning_rate": 9.996734160679296e-06,
"loss": 0.8904,
"step": 150
},
{
"epoch": 0.004897959183673469,
"eval_loss": 0.8796091675758362,
"eval_runtime": 74.0605,
"eval_samples_per_second": 1.35,
"eval_steps_per_second": 1.35,
"step": 150
},
{
"epoch": 0.0049306122448979596,
"grad_norm": 2.9235658645629883,
"learning_rate": 9.996701502286089e-06,
"loss": 0.819,
"step": 151
},
{
"epoch": 0.004963265306122449,
"grad_norm": 3.2201597690582275,
"learning_rate": 9.996668843892882e-06,
"loss": 0.8964,
"step": 152
},
{
"epoch": 0.004995918367346939,
"grad_norm": 2.7854557037353516,
"learning_rate": 9.996636185499674e-06,
"loss": 0.8886,
"step": 153
},
{
"epoch": 0.005028571428571428,
"grad_norm": 2.4900546073913574,
"learning_rate": 9.996603527106467e-06,
"loss": 0.8496,
"step": 154
},
{
"epoch": 0.005061224489795919,
"grad_norm": 2.7506489753723145,
"learning_rate": 9.99657086871326e-06,
"loss": 0.9044,
"step": 155
},
{
"epoch": 0.005093877551020408,
"grad_norm": 2.8616607189178467,
"learning_rate": 9.996538210320053e-06,
"loss": 0.8685,
"step": 156
},
{
"epoch": 0.005126530612244898,
"grad_norm": 2.814704656600952,
"learning_rate": 9.996505551926847e-06,
"loss": 0.8771,
"step": 157
},
{
"epoch": 0.005159183673469387,
"grad_norm": 3.006065845489502,
"learning_rate": 9.996472893533638e-06,
"loss": 0.8571,
"step": 158
},
{
"epoch": 0.005191836734693878,
"grad_norm": 2.821923017501831,
"learning_rate": 9.996440235140431e-06,
"loss": 0.8629,
"step": 159
},
{
"epoch": 0.005224489795918367,
"grad_norm": 3.278881072998047,
"learning_rate": 9.996407576747225e-06,
"loss": 0.893,
"step": 160
},
{
"epoch": 0.005257142857142857,
"grad_norm": 2.744616985321045,
"learning_rate": 9.996374918354018e-06,
"loss": 0.8682,
"step": 161
},
{
"epoch": 0.005289795918367347,
"grad_norm": 2.697544574737549,
"learning_rate": 9.996342259960811e-06,
"loss": 0.8974,
"step": 162
},
{
"epoch": 0.005322448979591837,
"grad_norm": 2.9768247604370117,
"learning_rate": 9.996309601567604e-06,
"loss": 0.8931,
"step": 163
},
{
"epoch": 0.0053551020408163265,
"grad_norm": 3.027183771133423,
"learning_rate": 9.996276943174396e-06,
"loss": 0.8897,
"step": 164
},
{
"epoch": 0.005387755102040816,
"grad_norm": 2.7222235202789307,
"learning_rate": 9.996244284781189e-06,
"loss": 0.8787,
"step": 165
},
{
"epoch": 0.0054204081632653065,
"grad_norm": 2.7184181213378906,
"learning_rate": 9.996211626387982e-06,
"loss": 0.8832,
"step": 166
},
{
"epoch": 0.005453061224489796,
"grad_norm": 2.8517744541168213,
"learning_rate": 9.996178967994776e-06,
"loss": 0.8463,
"step": 167
},
{
"epoch": 0.005485714285714286,
"grad_norm": 3.104189395904541,
"learning_rate": 9.996146309601569e-06,
"loss": 0.8724,
"step": 168
},
{
"epoch": 0.005518367346938775,
"grad_norm": 2.9883522987365723,
"learning_rate": 9.996113651208362e-06,
"loss": 0.8654,
"step": 169
},
{
"epoch": 0.005551020408163266,
"grad_norm": 3.018421173095703,
"learning_rate": 9.996080992815155e-06,
"loss": 0.8836,
"step": 170
},
{
"epoch": 0.005583673469387755,
"grad_norm": 2.795041561126709,
"learning_rate": 9.996048334421947e-06,
"loss": 0.8843,
"step": 171
},
{
"epoch": 0.005616326530612245,
"grad_norm": 2.594553232192993,
"learning_rate": 9.99601567602874e-06,
"loss": 0.8874,
"step": 172
},
{
"epoch": 0.005648979591836734,
"grad_norm": 2.946117877960205,
"learning_rate": 9.995983017635533e-06,
"loss": 0.8265,
"step": 173
},
{
"epoch": 0.005681632653061225,
"grad_norm": 3.060215950012207,
"learning_rate": 9.995950359242327e-06,
"loss": 0.8872,
"step": 174
},
{
"epoch": 0.005714285714285714,
"grad_norm": 2.5649561882019043,
"learning_rate": 9.99591770084912e-06,
"loss": 0.898,
"step": 175
},
{
"epoch": 0.005746938775510204,
"grad_norm": 2.7028987407684326,
"learning_rate": 9.995885042455911e-06,
"loss": 0.8168,
"step": 176
},
{
"epoch": 0.005779591836734694,
"grad_norm": 2.918105125427246,
"learning_rate": 9.995852384062705e-06,
"loss": 0.9006,
"step": 177
},
{
"epoch": 0.005812244897959184,
"grad_norm": 3.0559136867523193,
"learning_rate": 9.995819725669498e-06,
"loss": 0.9106,
"step": 178
},
{
"epoch": 0.005844897959183673,
"grad_norm": 2.7586793899536133,
"learning_rate": 9.995787067276291e-06,
"loss": 0.8643,
"step": 179
},
{
"epoch": 0.005877551020408163,
"grad_norm": 2.6476991176605225,
"learning_rate": 9.995754408883084e-06,
"loss": 0.8952,
"step": 180
},
{
"epoch": 0.005910204081632653,
"grad_norm": 2.624241352081299,
"learning_rate": 9.995721750489876e-06,
"loss": 0.8836,
"step": 181
},
{
"epoch": 0.005942857142857143,
"grad_norm": 3.0197315216064453,
"learning_rate": 9.995689092096669e-06,
"loss": 0.87,
"step": 182
},
{
"epoch": 0.0059755102040816325,
"grad_norm": 2.9282002449035645,
"learning_rate": 9.995656433703462e-06,
"loss": 0.8904,
"step": 183
},
{
"epoch": 0.006008163265306122,
"grad_norm": 2.6132211685180664,
"learning_rate": 9.995623775310255e-06,
"loss": 0.8565,
"step": 184
},
{
"epoch": 0.0060408163265306125,
"grad_norm": 2.6105284690856934,
"learning_rate": 9.995591116917049e-06,
"loss": 0.8842,
"step": 185
},
{
"epoch": 0.006073469387755102,
"grad_norm": 2.868211030960083,
"learning_rate": 9.995558458523842e-06,
"loss": 0.8575,
"step": 186
},
{
"epoch": 0.006106122448979592,
"grad_norm": 2.7017822265625,
"learning_rate": 9.995525800130633e-06,
"loss": 0.8429,
"step": 187
},
{
"epoch": 0.006138775510204081,
"grad_norm": 2.7130422592163086,
"learning_rate": 9.995493141737427e-06,
"loss": 0.8451,
"step": 188
},
{
"epoch": 0.006171428571428572,
"grad_norm": 2.681940793991089,
"learning_rate": 9.99546048334422e-06,
"loss": 0.866,
"step": 189
},
{
"epoch": 0.006204081632653061,
"grad_norm": 2.6012067794799805,
"learning_rate": 9.995427824951013e-06,
"loss": 0.8575,
"step": 190
},
{
"epoch": 0.006236734693877551,
"grad_norm": 2.7982876300811768,
"learning_rate": 9.995395166557806e-06,
"loss": 0.8818,
"step": 191
},
{
"epoch": 0.006269387755102041,
"grad_norm": 2.7027719020843506,
"learning_rate": 9.9953625081646e-06,
"loss": 0.8638,
"step": 192
},
{
"epoch": 0.006302040816326531,
"grad_norm": 2.6796092987060547,
"learning_rate": 9.995329849771393e-06,
"loss": 0.8473,
"step": 193
},
{
"epoch": 0.00633469387755102,
"grad_norm": 2.882390022277832,
"learning_rate": 9.995297191378184e-06,
"loss": 0.8854,
"step": 194
},
{
"epoch": 0.00636734693877551,
"grad_norm": 2.970914125442505,
"learning_rate": 9.995264532984978e-06,
"loss": 0.8499,
"step": 195
},
{
"epoch": 0.0064,
"grad_norm": 2.7777442932128906,
"learning_rate": 9.995231874591771e-06,
"loss": 0.8761,
"step": 196
},
{
"epoch": 0.00643265306122449,
"grad_norm": 2.6736974716186523,
"learning_rate": 9.995199216198564e-06,
"loss": 0.8645,
"step": 197
},
{
"epoch": 0.006465306122448979,
"grad_norm": 2.9803805351257324,
"learning_rate": 9.995166557805357e-06,
"loss": 0.8257,
"step": 198
},
{
"epoch": 0.006497959183673469,
"grad_norm": 2.7522194385528564,
"learning_rate": 9.995133899412149e-06,
"loss": 0.8467,
"step": 199
},
{
"epoch": 0.006530612244897959,
"grad_norm": 2.6975879669189453,
"learning_rate": 9.995101241018942e-06,
"loss": 0.872,
"step": 200
},
{
"epoch": 0.006530612244897959,
"eval_loss": 0.8632190823554993,
"eval_runtime": 74.7709,
"eval_samples_per_second": 1.337,
"eval_steps_per_second": 1.337,
"step": 200
},
{
"epoch": 0.006563265306122449,
"grad_norm": 2.8385777473449707,
"learning_rate": 9.995068582625735e-06,
"loss": 0.8592,
"step": 201
},
{
"epoch": 0.0065959183673469386,
"grad_norm": 2.8501343727111816,
"learning_rate": 9.995035924232529e-06,
"loss": 0.8516,
"step": 202
},
{
"epoch": 0.006628571428571429,
"grad_norm": 2.841095447540283,
"learning_rate": 9.995003265839322e-06,
"loss": 0.8589,
"step": 203
},
{
"epoch": 0.0066612244897959185,
"grad_norm": 2.7659661769866943,
"learning_rate": 9.994970607446113e-06,
"loss": 0.8434,
"step": 204
},
{
"epoch": 0.006693877551020408,
"grad_norm": 2.892019748687744,
"learning_rate": 9.994937949052907e-06,
"loss": 0.8896,
"step": 205
},
{
"epoch": 0.006726530612244898,
"grad_norm": 2.7645156383514404,
"learning_rate": 9.9949052906597e-06,
"loss": 0.8654,
"step": 206
},
{
"epoch": 0.006759183673469388,
"grad_norm": 2.7840843200683594,
"learning_rate": 9.994872632266493e-06,
"loss": 0.8558,
"step": 207
},
{
"epoch": 0.006791836734693878,
"grad_norm": 2.843766212463379,
"learning_rate": 9.994839973873286e-06,
"loss": 0.8482,
"step": 208
},
{
"epoch": 0.006824489795918367,
"grad_norm": 2.7724156379699707,
"learning_rate": 9.99480731548008e-06,
"loss": 0.8664,
"step": 209
},
{
"epoch": 0.006857142857142857,
"grad_norm": 2.8699095249176025,
"learning_rate": 9.994774657086873e-06,
"loss": 0.8506,
"step": 210
},
{
"epoch": 0.006889795918367347,
"grad_norm": 2.988041877746582,
"learning_rate": 9.994741998693666e-06,
"loss": 0.8781,
"step": 211
},
{
"epoch": 0.006922448979591837,
"grad_norm": 2.597806930541992,
"learning_rate": 9.994709340300458e-06,
"loss": 0.8586,
"step": 212
},
{
"epoch": 0.006955102040816326,
"grad_norm": 2.7888684272766113,
"learning_rate": 9.994676681907251e-06,
"loss": 0.8431,
"step": 213
},
{
"epoch": 0.006987755102040816,
"grad_norm": 2.80710768699646,
"learning_rate": 9.994644023514044e-06,
"loss": 0.8911,
"step": 214
},
{
"epoch": 0.007020408163265306,
"grad_norm": 2.6957030296325684,
"learning_rate": 9.994611365120837e-06,
"loss": 0.8626,
"step": 215
},
{
"epoch": 0.007053061224489796,
"grad_norm": 2.8306329250335693,
"learning_rate": 9.99457870672763e-06,
"loss": 0.8619,
"step": 216
},
{
"epoch": 0.0070857142857142855,
"grad_norm": 2.735477924346924,
"learning_rate": 9.994546048334422e-06,
"loss": 0.8661,
"step": 217
},
{
"epoch": 0.007118367346938776,
"grad_norm": 2.737856388092041,
"learning_rate": 9.994513389941215e-06,
"loss": 0.8718,
"step": 218
},
{
"epoch": 0.0071510204081632655,
"grad_norm": 2.743589401245117,
"learning_rate": 9.994480731548009e-06,
"loss": 0.8781,
"step": 219
},
{
"epoch": 0.007183673469387755,
"grad_norm": 2.8569047451019287,
"learning_rate": 9.994448073154802e-06,
"loss": 0.8512,
"step": 220
},
{
"epoch": 0.007216326530612245,
"grad_norm": 3.129603385925293,
"learning_rate": 9.994415414761595e-06,
"loss": 0.8839,
"step": 221
},
{
"epoch": 0.007248979591836735,
"grad_norm": 2.8460073471069336,
"learning_rate": 9.994382756368387e-06,
"loss": 0.8568,
"step": 222
},
{
"epoch": 0.007281632653061225,
"grad_norm": 3.024534225463867,
"learning_rate": 9.99435009797518e-06,
"loss": 0.8672,
"step": 223
},
{
"epoch": 0.007314285714285714,
"grad_norm": 3.0231759548187256,
"learning_rate": 9.994317439581973e-06,
"loss": 0.8899,
"step": 224
},
{
"epoch": 0.007346938775510204,
"grad_norm": 2.8645284175872803,
"learning_rate": 9.994284781188766e-06,
"loss": 0.8319,
"step": 225
},
{
"epoch": 0.007379591836734694,
"grad_norm": 2.741297483444214,
"learning_rate": 9.99425212279556e-06,
"loss": 0.8726,
"step": 226
},
{
"epoch": 0.007412244897959184,
"grad_norm": 2.8565313816070557,
"learning_rate": 9.994219464402353e-06,
"loss": 0.8807,
"step": 227
},
{
"epoch": 0.007444897959183673,
"grad_norm": 2.9202141761779785,
"learning_rate": 9.994186806009144e-06,
"loss": 0.8098,
"step": 228
},
{
"epoch": 0.007477551020408164,
"grad_norm": 2.955820083618164,
"learning_rate": 9.994154147615938e-06,
"loss": 0.8638,
"step": 229
},
{
"epoch": 0.007510204081632653,
"grad_norm": 2.7195394039154053,
"learning_rate": 9.99412148922273e-06,
"loss": 0.8591,
"step": 230
},
{
"epoch": 0.007542857142857143,
"grad_norm": 2.91768479347229,
"learning_rate": 9.994088830829524e-06,
"loss": 0.8774,
"step": 231
},
{
"epoch": 0.007575510204081632,
"grad_norm": 2.7152891159057617,
"learning_rate": 9.994056172436317e-06,
"loss": 0.8488,
"step": 232
},
{
"epoch": 0.007608163265306123,
"grad_norm": 2.5756800174713135,
"learning_rate": 9.99402351404311e-06,
"loss": 0.8598,
"step": 233
},
{
"epoch": 0.007640816326530612,
"grad_norm": 2.603003978729248,
"learning_rate": 9.993990855649904e-06,
"loss": 0.8473,
"step": 234
},
{
"epoch": 0.007673469387755102,
"grad_norm": 2.758294105529785,
"learning_rate": 9.993958197256695e-06,
"loss": 0.8624,
"step": 235
},
{
"epoch": 0.0077061224489795915,
"grad_norm": 2.8660902976989746,
"learning_rate": 9.993925538863489e-06,
"loss": 0.8564,
"step": 236
},
{
"epoch": 0.007738775510204082,
"grad_norm": 2.8767125606536865,
"learning_rate": 9.993892880470282e-06,
"loss": 0.8578,
"step": 237
},
{
"epoch": 0.0077714285714285715,
"grad_norm": 2.906562089920044,
"learning_rate": 9.993860222077075e-06,
"loss": 0.8383,
"step": 238
},
{
"epoch": 0.007804081632653061,
"grad_norm": 2.74647855758667,
"learning_rate": 9.993827563683868e-06,
"loss": 0.8656,
"step": 239
},
{
"epoch": 0.00783673469387755,
"grad_norm": 3.078082323074341,
"learning_rate": 9.99379490529066e-06,
"loss": 0.8244,
"step": 240
},
{
"epoch": 0.00786938775510204,
"grad_norm": 2.7712790966033936,
"learning_rate": 9.993762246897453e-06,
"loss": 0.8536,
"step": 241
},
{
"epoch": 0.007902040816326531,
"grad_norm": 2.852389335632324,
"learning_rate": 9.993729588504246e-06,
"loss": 0.8496,
"step": 242
},
{
"epoch": 0.007934693877551021,
"grad_norm": 2.7907660007476807,
"learning_rate": 9.99369693011104e-06,
"loss": 0.8993,
"step": 243
},
{
"epoch": 0.00796734693877551,
"grad_norm": 2.8219473361968994,
"learning_rate": 9.993664271717833e-06,
"loss": 0.8652,
"step": 244
},
{
"epoch": 0.008,
"grad_norm": 2.6678013801574707,
"learning_rate": 9.993631613324624e-06,
"loss": 0.8636,
"step": 245
},
{
"epoch": 0.00803265306122449,
"grad_norm": 2.724881887435913,
"learning_rate": 9.993598954931417e-06,
"loss": 0.8559,
"step": 246
},
{
"epoch": 0.00806530612244898,
"grad_norm": 2.8275516033172607,
"learning_rate": 9.99356629653821e-06,
"loss": 0.8646,
"step": 247
},
{
"epoch": 0.008097959183673469,
"grad_norm": 2.8110673427581787,
"learning_rate": 9.993533638145004e-06,
"loss": 0.8713,
"step": 248
},
{
"epoch": 0.008130612244897958,
"grad_norm": 2.652221202850342,
"learning_rate": 9.993500979751797e-06,
"loss": 0.8373,
"step": 249
},
{
"epoch": 0.00816326530612245,
"grad_norm": 2.6357979774475098,
"learning_rate": 9.99346832135859e-06,
"loss": 0.8537,
"step": 250
},
{
"epoch": 0.00816326530612245,
"eval_loss": 0.8545005917549133,
"eval_runtime": 74.0963,
"eval_samples_per_second": 1.35,
"eval_steps_per_second": 1.35,
"step": 250
},
{
"epoch": 0.00819591836734694,
"grad_norm": 2.6132781505584717,
"learning_rate": 9.993435662965384e-06,
"loss": 0.8592,
"step": 251
},
{
"epoch": 0.008228571428571429,
"grad_norm": 2.942228078842163,
"learning_rate": 9.993403004572177e-06,
"loss": 0.7732,
"step": 252
},
{
"epoch": 0.008261224489795918,
"grad_norm": 2.5452983379364014,
"learning_rate": 9.993370346178968e-06,
"loss": 0.8666,
"step": 253
},
{
"epoch": 0.008293877551020408,
"grad_norm": 2.863976240158081,
"learning_rate": 9.993337687785762e-06,
"loss": 0.8232,
"step": 254
},
{
"epoch": 0.008326530612244898,
"grad_norm": 2.640972137451172,
"learning_rate": 9.993305029392555e-06,
"loss": 0.8703,
"step": 255
},
{
"epoch": 0.008359183673469387,
"grad_norm": 2.98362135887146,
"learning_rate": 9.993272370999348e-06,
"loss": 0.8165,
"step": 256
},
{
"epoch": 0.008391836734693878,
"grad_norm": 2.848294258117676,
"learning_rate": 9.993239712606141e-06,
"loss": 0.9104,
"step": 257
},
{
"epoch": 0.008424489795918368,
"grad_norm": 2.7992308139801025,
"learning_rate": 9.993207054212933e-06,
"loss": 0.8146,
"step": 258
},
{
"epoch": 0.008457142857142858,
"grad_norm": 2.905052900314331,
"learning_rate": 9.993174395819726e-06,
"loss": 0.8606,
"step": 259
},
{
"epoch": 0.008489795918367347,
"grad_norm": 2.97420334815979,
"learning_rate": 9.99314173742652e-06,
"loss": 0.7909,
"step": 260
},
{
"epoch": 0.008522448979591837,
"grad_norm": 2.7139410972595215,
"learning_rate": 9.993109079033313e-06,
"loss": 0.8593,
"step": 261
},
{
"epoch": 0.008555102040816326,
"grad_norm": 2.7178266048431396,
"learning_rate": 9.993076420640106e-06,
"loss": 0.8785,
"step": 262
},
{
"epoch": 0.008587755102040816,
"grad_norm": 2.6858327388763428,
"learning_rate": 9.993043762246897e-06,
"loss": 0.846,
"step": 263
},
{
"epoch": 0.008620408163265305,
"grad_norm": 2.813204288482666,
"learning_rate": 9.99301110385369e-06,
"loss": 0.8386,
"step": 264
},
{
"epoch": 0.008653061224489797,
"grad_norm": 2.795591115951538,
"learning_rate": 9.992978445460484e-06,
"loss": 0.8417,
"step": 265
},
{
"epoch": 0.008685714285714286,
"grad_norm": 2.8094005584716797,
"learning_rate": 9.992945787067277e-06,
"loss": 0.8396,
"step": 266
},
{
"epoch": 0.008718367346938776,
"grad_norm": 2.723392963409424,
"learning_rate": 9.99291312867407e-06,
"loss": 0.8357,
"step": 267
},
{
"epoch": 0.008751020408163265,
"grad_norm": 2.9623281955718994,
"learning_rate": 9.992880470280864e-06,
"loss": 0.8874,
"step": 268
},
{
"epoch": 0.008783673469387755,
"grad_norm": 2.8696374893188477,
"learning_rate": 9.992847811887655e-06,
"loss": 0.8595,
"step": 269
},
{
"epoch": 0.008816326530612244,
"grad_norm": 2.906726837158203,
"learning_rate": 9.992815153494448e-06,
"loss": 0.8736,
"step": 270
},
{
"epoch": 0.008848979591836734,
"grad_norm": 2.814854383468628,
"learning_rate": 9.992782495101242e-06,
"loss": 0.8603,
"step": 271
},
{
"epoch": 0.008881632653061225,
"grad_norm": 2.799457550048828,
"learning_rate": 9.992749836708035e-06,
"loss": 0.8469,
"step": 272
},
{
"epoch": 0.008914285714285715,
"grad_norm": 2.755631923675537,
"learning_rate": 9.992717178314828e-06,
"loss": 0.8702,
"step": 273
},
{
"epoch": 0.008946938775510204,
"grad_norm": 2.9200055599212646,
"learning_rate": 9.992684519921621e-06,
"loss": 0.8407,
"step": 274
},
{
"epoch": 0.008979591836734694,
"grad_norm": 3.2251176834106445,
"learning_rate": 9.992651861528415e-06,
"loss": 0.8572,
"step": 275
},
{
"epoch": 0.009012244897959184,
"grad_norm": 3.8850901126861572,
"learning_rate": 9.992619203135206e-06,
"loss": 0.8458,
"step": 276
},
{
"epoch": 0.009044897959183673,
"grad_norm": 2.8082637786865234,
"learning_rate": 9.992586544742e-06,
"loss": 0.8415,
"step": 277
},
{
"epoch": 0.009077551020408163,
"grad_norm": 2.8923559188842773,
"learning_rate": 9.992553886348793e-06,
"loss": 0.8116,
"step": 278
},
{
"epoch": 0.009110204081632652,
"grad_norm": 2.849003791809082,
"learning_rate": 9.992521227955586e-06,
"loss": 0.834,
"step": 279
},
{
"epoch": 0.009142857142857144,
"grad_norm": 2.7201125621795654,
"learning_rate": 9.992488569562379e-06,
"loss": 0.8518,
"step": 280
},
{
"epoch": 0.009175510204081633,
"grad_norm": 2.9798922538757324,
"learning_rate": 9.99245591116917e-06,
"loss": 0.8691,
"step": 281
},
{
"epoch": 0.009208163265306123,
"grad_norm": 2.805250406265259,
"learning_rate": 9.992423252775964e-06,
"loss": 0.8537,
"step": 282
},
{
"epoch": 0.009240816326530612,
"grad_norm": 2.6364924907684326,
"learning_rate": 9.992390594382757e-06,
"loss": 0.8368,
"step": 283
},
{
"epoch": 0.009273469387755102,
"grad_norm": 2.692296028137207,
"learning_rate": 9.99235793598955e-06,
"loss": 0.8373,
"step": 284
},
{
"epoch": 0.009306122448979591,
"grad_norm": 2.876549005508423,
"learning_rate": 9.992325277596344e-06,
"loss": 0.8348,
"step": 285
},
{
"epoch": 0.009338775510204081,
"grad_norm": 2.77553129196167,
"learning_rate": 9.992292619203135e-06,
"loss": 0.8377,
"step": 286
},
{
"epoch": 0.009371428571428572,
"grad_norm": 2.900949478149414,
"learning_rate": 9.992259960809928e-06,
"loss": 0.8469,
"step": 287
},
{
"epoch": 0.009404081632653062,
"grad_norm": 2.99700927734375,
"learning_rate": 9.992227302416722e-06,
"loss": 0.8926,
"step": 288
},
{
"epoch": 0.009436734693877551,
"grad_norm": 2.738802194595337,
"learning_rate": 9.992194644023515e-06,
"loss": 0.8511,
"step": 289
},
{
"epoch": 0.009469387755102041,
"grad_norm": 2.7471158504486084,
"learning_rate": 9.992161985630308e-06,
"loss": 0.8449,
"step": 290
},
{
"epoch": 0.00950204081632653,
"grad_norm": 2.7274839878082275,
"learning_rate": 9.992129327237101e-06,
"loss": 0.8308,
"step": 291
},
{
"epoch": 0.00953469387755102,
"grad_norm": 2.554326057434082,
"learning_rate": 9.992096668843893e-06,
"loss": 0.868,
"step": 292
},
{
"epoch": 0.00956734693877551,
"grad_norm": 2.578237295150757,
"learning_rate": 9.992064010450688e-06,
"loss": 0.8557,
"step": 293
},
{
"epoch": 0.0096,
"grad_norm": 3.0283281803131104,
"learning_rate": 9.99203135205748e-06,
"loss": 0.8289,
"step": 294
},
{
"epoch": 0.00963265306122449,
"grad_norm": 2.785170316696167,
"learning_rate": 9.991998693664272e-06,
"loss": 0.878,
"step": 295
},
{
"epoch": 0.00966530612244898,
"grad_norm": 2.8442277908325195,
"learning_rate": 9.991966035271066e-06,
"loss": 0.8282,
"step": 296
},
{
"epoch": 0.00969795918367347,
"grad_norm": 2.847158908843994,
"learning_rate": 9.991933376877859e-06,
"loss": 0.844,
"step": 297
},
{
"epoch": 0.00973061224489796,
"grad_norm": 2.6960904598236084,
"learning_rate": 9.991900718484652e-06,
"loss": 0.863,
"step": 298
},
{
"epoch": 0.009763265306122449,
"grad_norm": 2.97739839553833,
"learning_rate": 9.991868060091444e-06,
"loss": 0.8448,
"step": 299
},
{
"epoch": 0.009795918367346938,
"grad_norm": 2.9316720962524414,
"learning_rate": 9.991835401698237e-06,
"loss": 0.8512,
"step": 300
},
{
"epoch": 0.009795918367346938,
"eval_loss": 0.8451775312423706,
"eval_runtime": 73.1659,
"eval_samples_per_second": 1.367,
"eval_steps_per_second": 1.367,
"step": 300
},
{
"epoch": 0.009828571428571428,
"grad_norm": 3.2117464542388916,
"learning_rate": 9.99180274330503e-06,
"loss": 0.852,
"step": 301
},
{
"epoch": 0.009861224489795919,
"grad_norm": 2.7478814125061035,
"learning_rate": 9.991770084911823e-06,
"loss": 0.8454,
"step": 302
},
{
"epoch": 0.009893877551020409,
"grad_norm": 2.7774980068206787,
"learning_rate": 9.991737426518617e-06,
"loss": 0.8581,
"step": 303
},
{
"epoch": 0.009926530612244898,
"grad_norm": 2.691710948944092,
"learning_rate": 9.991704768125408e-06,
"loss": 0.8483,
"step": 304
},
{
"epoch": 0.009959183673469388,
"grad_norm": 2.803921937942505,
"learning_rate": 9.991672109732201e-06,
"loss": 0.8394,
"step": 305
},
{
"epoch": 0.009991836734693877,
"grad_norm": 2.9755606651306152,
"learning_rate": 9.991639451338995e-06,
"loss": 0.8245,
"step": 306
},
{
"epoch": 0.010024489795918367,
"grad_norm": 3.09531569480896,
"learning_rate": 9.991606792945788e-06,
"loss": 0.8701,
"step": 307
},
{
"epoch": 0.010057142857142857,
"grad_norm": 2.8512625694274902,
"learning_rate": 9.991574134552581e-06,
"loss": 0.8597,
"step": 308
},
{
"epoch": 0.010089795918367346,
"grad_norm": 2.7423555850982666,
"learning_rate": 9.991541476159374e-06,
"loss": 0.8116,
"step": 309
},
{
"epoch": 0.010122448979591837,
"grad_norm": 2.8539059162139893,
"learning_rate": 9.991508817766166e-06,
"loss": 0.841,
"step": 310
},
{
"epoch": 0.010155102040816327,
"grad_norm": 2.843325138092041,
"learning_rate": 9.99147615937296e-06,
"loss": 0.8491,
"step": 311
},
{
"epoch": 0.010187755102040816,
"grad_norm": 2.6367223262786865,
"learning_rate": 9.991443500979752e-06,
"loss": 0.8452,
"step": 312
},
{
"epoch": 0.010220408163265306,
"grad_norm": 2.6435811519622803,
"learning_rate": 9.991410842586546e-06,
"loss": 0.8415,
"step": 313
},
{
"epoch": 0.010253061224489796,
"grad_norm": 2.727060317993164,
"learning_rate": 9.991378184193339e-06,
"loss": 0.7985,
"step": 314
},
{
"epoch": 0.010285714285714285,
"grad_norm": 2.9816014766693115,
"learning_rate": 9.991345525800132e-06,
"loss": 0.8129,
"step": 315
},
{
"epoch": 0.010318367346938775,
"grad_norm": 3.3140499591827393,
"learning_rate": 9.991312867406925e-06,
"loss": 0.8822,
"step": 316
},
{
"epoch": 0.010351020408163266,
"grad_norm": 2.8612375259399414,
"learning_rate": 9.991280209013717e-06,
"loss": 0.7966,
"step": 317
},
{
"epoch": 0.010383673469387756,
"grad_norm": 2.692972183227539,
"learning_rate": 9.99124755062051e-06,
"loss": 0.8288,
"step": 318
},
{
"epoch": 0.010416326530612245,
"grad_norm": 2.980607271194458,
"learning_rate": 9.991214892227303e-06,
"loss": 0.8238,
"step": 319
},
{
"epoch": 0.010448979591836735,
"grad_norm": 3.0936498641967773,
"learning_rate": 9.991182233834097e-06,
"loss": 0.8338,
"step": 320
},
{
"epoch": 0.010481632653061224,
"grad_norm": 2.7482964992523193,
"learning_rate": 9.99114957544089e-06,
"loss": 0.8511,
"step": 321
},
{
"epoch": 0.010514285714285714,
"grad_norm": 2.6563820838928223,
"learning_rate": 9.991116917047681e-06,
"loss": 0.8568,
"step": 322
},
{
"epoch": 0.010546938775510203,
"grad_norm": 2.7284178733825684,
"learning_rate": 9.991084258654475e-06,
"loss": 0.8615,
"step": 323
},
{
"epoch": 0.010579591836734695,
"grad_norm": 2.7720155715942383,
"learning_rate": 9.991051600261268e-06,
"loss": 0.8459,
"step": 324
},
{
"epoch": 0.010612244897959184,
"grad_norm": 2.6759071350097656,
"learning_rate": 9.991018941868061e-06,
"loss": 0.8489,
"step": 325
},
{
"epoch": 0.010644897959183674,
"grad_norm": 2.8001279830932617,
"learning_rate": 9.990986283474854e-06,
"loss": 0.83,
"step": 326
},
{
"epoch": 0.010677551020408163,
"grad_norm": 2.7164299488067627,
"learning_rate": 9.990953625081646e-06,
"loss": 0.8059,
"step": 327
},
{
"epoch": 0.010710204081632653,
"grad_norm": 2.833035707473755,
"learning_rate": 9.990920966688439e-06,
"loss": 0.8394,
"step": 328
},
{
"epoch": 0.010742857142857143,
"grad_norm": 2.6928770542144775,
"learning_rate": 9.990888308295232e-06,
"loss": 0.8297,
"step": 329
},
{
"epoch": 0.010775510204081632,
"grad_norm": 2.7575488090515137,
"learning_rate": 9.990855649902026e-06,
"loss": 0.8149,
"step": 330
},
{
"epoch": 0.010808163265306122,
"grad_norm": 2.7353413105010986,
"learning_rate": 9.990822991508819e-06,
"loss": 0.839,
"step": 331
},
{
"epoch": 0.010840816326530613,
"grad_norm": 2.7132623195648193,
"learning_rate": 9.990790333115612e-06,
"loss": 0.8452,
"step": 332
},
{
"epoch": 0.010873469387755103,
"grad_norm": 2.8830623626708984,
"learning_rate": 9.990757674722404e-06,
"loss": 0.8156,
"step": 333
},
{
"epoch": 0.010906122448979592,
"grad_norm": 2.7021992206573486,
"learning_rate": 9.990725016329197e-06,
"loss": 0.8368,
"step": 334
},
{
"epoch": 0.010938775510204082,
"grad_norm": 2.6805949211120605,
"learning_rate": 9.99069235793599e-06,
"loss": 0.8402,
"step": 335
},
{
"epoch": 0.010971428571428571,
"grad_norm": 3.0374648571014404,
"learning_rate": 9.990659699542783e-06,
"loss": 0.8332,
"step": 336
},
{
"epoch": 0.01100408163265306,
"grad_norm": 2.9531807899475098,
"learning_rate": 9.990627041149577e-06,
"loss": 0.8451,
"step": 337
},
{
"epoch": 0.01103673469387755,
"grad_norm": 2.8179969787597656,
"learning_rate": 9.99059438275637e-06,
"loss": 0.8336,
"step": 338
},
{
"epoch": 0.011069387755102042,
"grad_norm": 3.139448642730713,
"learning_rate": 9.990561724363163e-06,
"loss": 0.865,
"step": 339
},
{
"epoch": 0.011102040816326531,
"grad_norm": 3.040363073348999,
"learning_rate": 9.990529065969955e-06,
"loss": 0.8387,
"step": 340
},
{
"epoch": 0.01113469387755102,
"grad_norm": 2.700171709060669,
"learning_rate": 9.990496407576748e-06,
"loss": 0.8219,
"step": 341
},
{
"epoch": 0.01116734693877551,
"grad_norm": 2.7326748371124268,
"learning_rate": 9.990463749183541e-06,
"loss": 0.7956,
"step": 342
},
{
"epoch": 0.0112,
"grad_norm": 2.7744908332824707,
"learning_rate": 9.990431090790334e-06,
"loss": 0.8133,
"step": 343
},
{
"epoch": 0.01123265306122449,
"grad_norm": 2.8565139770507812,
"learning_rate": 9.990398432397128e-06,
"loss": 0.825,
"step": 344
},
{
"epoch": 0.011265306122448979,
"grad_norm": 2.9909799098968506,
"learning_rate": 9.990365774003919e-06,
"loss": 0.8683,
"step": 345
},
{
"epoch": 0.011297959183673469,
"grad_norm": 2.642660140991211,
"learning_rate": 9.990333115610712e-06,
"loss": 0.8598,
"step": 346
},
{
"epoch": 0.01133061224489796,
"grad_norm": 2.7017788887023926,
"learning_rate": 9.990300457217506e-06,
"loss": 0.8164,
"step": 347
},
{
"epoch": 0.01136326530612245,
"grad_norm": 2.7391109466552734,
"learning_rate": 9.990267798824299e-06,
"loss": 0.8498,
"step": 348
},
{
"epoch": 0.011395918367346939,
"grad_norm": 2.7465527057647705,
"learning_rate": 9.990235140431092e-06,
"loss": 0.8225,
"step": 349
},
{
"epoch": 0.011428571428571429,
"grad_norm": 2.6583974361419678,
"learning_rate": 9.990202482037884e-06,
"loss": 0.8505,
"step": 350
},
{
"epoch": 0.011428571428571429,
"eval_loss": 0.8395382165908813,
"eval_runtime": 73.8611,
"eval_samples_per_second": 1.354,
"eval_steps_per_second": 1.354,
"step": 350
},
{
"epoch": 0.011461224489795918,
"grad_norm": 2.7981696128845215,
"learning_rate": 9.990169823644677e-06,
"loss": 0.8312,
"step": 351
},
{
"epoch": 0.011493877551020408,
"grad_norm": 2.924536943435669,
"learning_rate": 9.99013716525147e-06,
"loss": 0.8498,
"step": 352
},
{
"epoch": 0.011526530612244897,
"grad_norm": 2.819368362426758,
"learning_rate": 9.990104506858263e-06,
"loss": 0.8022,
"step": 353
},
{
"epoch": 0.011559183673469389,
"grad_norm": 2.9508090019226074,
"learning_rate": 9.990071848465056e-06,
"loss": 0.8363,
"step": 354
},
{
"epoch": 0.011591836734693878,
"grad_norm": 2.643820285797119,
"learning_rate": 9.99003919007185e-06,
"loss": 0.8172,
"step": 355
},
{
"epoch": 0.011624489795918368,
"grad_norm": 2.5882935523986816,
"learning_rate": 9.990006531678641e-06,
"loss": 0.8095,
"step": 356
},
{
"epoch": 0.011657142857142857,
"grad_norm": 2.968686103820801,
"learning_rate": 9.989973873285436e-06,
"loss": 0.8673,
"step": 357
},
{
"epoch": 0.011689795918367347,
"grad_norm": 2.8162810802459717,
"learning_rate": 9.989941214892228e-06,
"loss": 0.8693,
"step": 358
},
{
"epoch": 0.011722448979591836,
"grad_norm": 2.8215367794036865,
"learning_rate": 9.989908556499021e-06,
"loss": 0.8053,
"step": 359
},
{
"epoch": 0.011755102040816326,
"grad_norm": 2.7967352867126465,
"learning_rate": 9.989875898105814e-06,
"loss": 0.8094,
"step": 360
},
{
"epoch": 0.011787755102040815,
"grad_norm": 2.873476505279541,
"learning_rate": 9.989843239712607e-06,
"loss": 0.8174,
"step": 361
},
{
"epoch": 0.011820408163265307,
"grad_norm": 2.8655476570129395,
"learning_rate": 9.9898105813194e-06,
"loss": 0.7988,
"step": 362
},
{
"epoch": 0.011853061224489796,
"grad_norm": 2.794279098510742,
"learning_rate": 9.989777922926192e-06,
"loss": 0.8404,
"step": 363
},
{
"epoch": 0.011885714285714286,
"grad_norm": 2.8534135818481445,
"learning_rate": 9.989745264532985e-06,
"loss": 0.8474,
"step": 364
},
{
"epoch": 0.011918367346938775,
"grad_norm": 2.716606855392456,
"learning_rate": 9.989712606139779e-06,
"loss": 0.8256,
"step": 365
},
{
"epoch": 0.011951020408163265,
"grad_norm": 2.8670265674591064,
"learning_rate": 9.989679947746572e-06,
"loss": 0.834,
"step": 366
},
{
"epoch": 0.011983673469387755,
"grad_norm": 2.9551408290863037,
"learning_rate": 9.989647289353365e-06,
"loss": 0.8295,
"step": 367
},
{
"epoch": 0.012016326530612244,
"grad_norm": 3.209449529647827,
"learning_rate": 9.989614630960157e-06,
"loss": 0.8167,
"step": 368
},
{
"epoch": 0.012048979591836735,
"grad_norm": 2.8796558380126953,
"learning_rate": 9.98958197256695e-06,
"loss": 0.8245,
"step": 369
},
{
"epoch": 0.012081632653061225,
"grad_norm": 2.9785306453704834,
"learning_rate": 9.989549314173743e-06,
"loss": 0.7946,
"step": 370
},
{
"epoch": 0.012114285714285715,
"grad_norm": 2.8382294178009033,
"learning_rate": 9.989516655780536e-06,
"loss": 0.8392,
"step": 371
},
{
"epoch": 0.012146938775510204,
"grad_norm": 3.2182369232177734,
"learning_rate": 9.98948399738733e-06,
"loss": 0.8343,
"step": 372
},
{
"epoch": 0.012179591836734694,
"grad_norm": 2.9380509853363037,
"learning_rate": 9.989451338994123e-06,
"loss": 0.8417,
"step": 373
},
{
"epoch": 0.012212244897959183,
"grad_norm": 2.6060733795166016,
"learning_rate": 9.989418680600914e-06,
"loss": 0.8471,
"step": 374
},
{
"epoch": 0.012244897959183673,
"grad_norm": 3.0596020221710205,
"learning_rate": 9.989386022207708e-06,
"loss": 0.8408,
"step": 375
},
{
"epoch": 0.012277551020408162,
"grad_norm": 2.91416335105896,
"learning_rate": 9.989353363814501e-06,
"loss": 0.8648,
"step": 376
},
{
"epoch": 0.012310204081632654,
"grad_norm": 2.8798558712005615,
"learning_rate": 9.989320705421294e-06,
"loss": 0.7871,
"step": 377
},
{
"epoch": 0.012342857142857143,
"grad_norm": 2.9666454792022705,
"learning_rate": 9.989288047028087e-06,
"loss": 0.8461,
"step": 378
},
{
"epoch": 0.012375510204081633,
"grad_norm": 2.9145092964172363,
"learning_rate": 9.98925538863488e-06,
"loss": 0.8288,
"step": 379
},
{
"epoch": 0.012408163265306122,
"grad_norm": 2.844569444656372,
"learning_rate": 9.989222730241674e-06,
"loss": 0.8685,
"step": 380
},
{
"epoch": 0.012440816326530612,
"grad_norm": 2.790707588195801,
"learning_rate": 9.989190071848465e-06,
"loss": 0.8467,
"step": 381
},
{
"epoch": 0.012473469387755102,
"grad_norm": 2.7228894233703613,
"learning_rate": 9.989157413455259e-06,
"loss": 0.8149,
"step": 382
},
{
"epoch": 0.012506122448979591,
"grad_norm": 2.8930270671844482,
"learning_rate": 9.989124755062052e-06,
"loss": 0.8521,
"step": 383
},
{
"epoch": 0.012538775510204082,
"grad_norm": 2.7201409339904785,
"learning_rate": 9.989092096668845e-06,
"loss": 0.8303,
"step": 384
},
{
"epoch": 0.012571428571428572,
"grad_norm": 2.558375597000122,
"learning_rate": 9.989059438275638e-06,
"loss": 0.8165,
"step": 385
},
{
"epoch": 0.012604081632653062,
"grad_norm": 2.893265724182129,
"learning_rate": 9.98902677988243e-06,
"loss": 0.833,
"step": 386
},
{
"epoch": 0.012636734693877551,
"grad_norm": 2.6960701942443848,
"learning_rate": 9.988994121489223e-06,
"loss": 0.7972,
"step": 387
},
{
"epoch": 0.01266938775510204,
"grad_norm": 2.5734570026397705,
"learning_rate": 9.988961463096016e-06,
"loss": 0.8748,
"step": 388
},
{
"epoch": 0.01270204081632653,
"grad_norm": 2.993227243423462,
"learning_rate": 9.98892880470281e-06,
"loss": 0.8405,
"step": 389
},
{
"epoch": 0.01273469387755102,
"grad_norm": 3.2019479274749756,
"learning_rate": 9.988896146309603e-06,
"loss": 0.8423,
"step": 390
},
{
"epoch": 0.012767346938775511,
"grad_norm": 3.2083091735839844,
"learning_rate": 9.988863487916394e-06,
"loss": 0.8088,
"step": 391
},
{
"epoch": 0.0128,
"grad_norm": 3.1379964351654053,
"learning_rate": 9.988830829523188e-06,
"loss": 0.8681,
"step": 392
},
{
"epoch": 0.01283265306122449,
"grad_norm": 2.7037453651428223,
"learning_rate": 9.98879817112998e-06,
"loss": 0.8379,
"step": 393
},
{
"epoch": 0.01286530612244898,
"grad_norm": 2.828984022140503,
"learning_rate": 9.988765512736774e-06,
"loss": 0.8148,
"step": 394
},
{
"epoch": 0.01289795918367347,
"grad_norm": 2.9447624683380127,
"learning_rate": 9.988732854343567e-06,
"loss": 0.8263,
"step": 395
},
{
"epoch": 0.012930612244897959,
"grad_norm": 2.7229952812194824,
"learning_rate": 9.98870019595036e-06,
"loss": 0.8613,
"step": 396
},
{
"epoch": 0.012963265306122448,
"grad_norm": 2.820530652999878,
"learning_rate": 9.988667537557152e-06,
"loss": 0.8174,
"step": 397
},
{
"epoch": 0.012995918367346938,
"grad_norm": 2.897592067718506,
"learning_rate": 9.988634879163947e-06,
"loss": 0.8416,
"step": 398
},
{
"epoch": 0.01302857142857143,
"grad_norm": 3.032390832901001,
"learning_rate": 9.988602220770739e-06,
"loss": 0.842,
"step": 399
},
{
"epoch": 0.013061224489795919,
"grad_norm": 2.949514389038086,
"learning_rate": 9.988569562377532e-06,
"loss": 0.8283,
"step": 400
},
{
"epoch": 0.013061224489795919,
"eval_loss": 0.8356085419654846,
"eval_runtime": 76.0573,
"eval_samples_per_second": 1.315,
"eval_steps_per_second": 1.315,
"step": 400
},
{
"epoch": 0.013093877551020408,
"grad_norm": 2.811276912689209,
"learning_rate": 9.988536903984325e-06,
"loss": 0.7906,
"step": 401
},
{
"epoch": 0.013126530612244898,
"grad_norm": 2.9961533546447754,
"learning_rate": 9.988504245591118e-06,
"loss": 0.839,
"step": 402
},
{
"epoch": 0.013159183673469388,
"grad_norm": 3.0735530853271484,
"learning_rate": 9.988471587197911e-06,
"loss": 0.8053,
"step": 403
},
{
"epoch": 0.013191836734693877,
"grad_norm": 2.793745279312134,
"learning_rate": 9.988438928804703e-06,
"loss": 0.8231,
"step": 404
},
{
"epoch": 0.013224489795918367,
"grad_norm": 3.1552298069000244,
"learning_rate": 9.988406270411496e-06,
"loss": 0.8328,
"step": 405
},
{
"epoch": 0.013257142857142858,
"grad_norm": 2.984865427017212,
"learning_rate": 9.98837361201829e-06,
"loss": 0.811,
"step": 406
},
{
"epoch": 0.013289795918367348,
"grad_norm": 2.8718035221099854,
"learning_rate": 9.988340953625083e-06,
"loss": 0.853,
"step": 407
},
{
"epoch": 0.013322448979591837,
"grad_norm": 3.0865938663482666,
"learning_rate": 9.988308295231876e-06,
"loss": 0.8003,
"step": 408
},
{
"epoch": 0.013355102040816327,
"grad_norm": 2.7381820678710938,
"learning_rate": 9.988275636838668e-06,
"loss": 0.8184,
"step": 409
},
{
"epoch": 0.013387755102040816,
"grad_norm": 2.8170037269592285,
"learning_rate": 9.98824297844546e-06,
"loss": 0.8311,
"step": 410
},
{
"epoch": 0.013420408163265306,
"grad_norm": 2.9140095710754395,
"learning_rate": 9.988210320052254e-06,
"loss": 0.834,
"step": 411
},
{
"epoch": 0.013453061224489795,
"grad_norm": 2.933525800704956,
"learning_rate": 9.988177661659047e-06,
"loss": 0.8294,
"step": 412
},
{
"epoch": 0.013485714285714285,
"grad_norm": 2.8974623680114746,
"learning_rate": 9.98814500326584e-06,
"loss": 0.8704,
"step": 413
},
{
"epoch": 0.013518367346938776,
"grad_norm": 2.794966697692871,
"learning_rate": 9.988112344872634e-06,
"loss": 0.8579,
"step": 414
},
{
"epoch": 0.013551020408163266,
"grad_norm": 2.856208324432373,
"learning_rate": 9.988079686479425e-06,
"loss": 0.8239,
"step": 415
},
{
"epoch": 0.013583673469387755,
"grad_norm": 2.999901533126831,
"learning_rate": 9.988047028086218e-06,
"loss": 0.8166,
"step": 416
},
{
"epoch": 0.013616326530612245,
"grad_norm": 2.8142268657684326,
"learning_rate": 9.988014369693012e-06,
"loss": 0.8343,
"step": 417
},
{
"epoch": 0.013648979591836734,
"grad_norm": 2.808567523956299,
"learning_rate": 9.987981711299805e-06,
"loss": 0.8414,
"step": 418
},
{
"epoch": 0.013681632653061224,
"grad_norm": 2.6575918197631836,
"learning_rate": 9.987949052906598e-06,
"loss": 0.8256,
"step": 419
},
{
"epoch": 0.013714285714285714,
"grad_norm": 2.9415652751922607,
"learning_rate": 9.987916394513391e-06,
"loss": 0.811,
"step": 420
},
{
"epoch": 0.013746938775510205,
"grad_norm": 2.7710697650909424,
"learning_rate": 9.987883736120185e-06,
"loss": 0.8132,
"step": 421
},
{
"epoch": 0.013779591836734694,
"grad_norm": 2.6907401084899902,
"learning_rate": 9.987851077726976e-06,
"loss": 0.8022,
"step": 422
},
{
"epoch": 0.013812244897959184,
"grad_norm": 3.010432481765747,
"learning_rate": 9.98781841933377e-06,
"loss": 0.8196,
"step": 423
},
{
"epoch": 0.013844897959183674,
"grad_norm": 2.868598461151123,
"learning_rate": 9.987785760940563e-06,
"loss": 0.8234,
"step": 424
},
{
"epoch": 0.013877551020408163,
"grad_norm": 2.9146206378936768,
"learning_rate": 9.987753102547356e-06,
"loss": 0.8357,
"step": 425
},
{
"epoch": 0.013910204081632653,
"grad_norm": 3.148287534713745,
"learning_rate": 9.987720444154149e-06,
"loss": 0.8168,
"step": 426
},
{
"epoch": 0.013942857142857142,
"grad_norm": 3.0107545852661133,
"learning_rate": 9.98768778576094e-06,
"loss": 0.8313,
"step": 427
},
{
"epoch": 0.013975510204081632,
"grad_norm": 2.7397043704986572,
"learning_rate": 9.987655127367734e-06,
"loss": 0.825,
"step": 428
},
{
"epoch": 0.014008163265306123,
"grad_norm": 2.8118340969085693,
"learning_rate": 9.987622468974527e-06,
"loss": 0.8101,
"step": 429
},
{
"epoch": 0.014040816326530613,
"grad_norm": 2.862879514694214,
"learning_rate": 9.98758981058132e-06,
"loss": 0.8504,
"step": 430
},
{
"epoch": 0.014073469387755102,
"grad_norm": 3.0519521236419678,
"learning_rate": 9.987557152188114e-06,
"loss": 0.8337,
"step": 431
},
{
"epoch": 0.014106122448979592,
"grad_norm": 2.8947372436523438,
"learning_rate": 9.987524493794905e-06,
"loss": 0.8041,
"step": 432
},
{
"epoch": 0.014138775510204081,
"grad_norm": 2.7874197959899902,
"learning_rate": 9.987491835401698e-06,
"loss": 0.8143,
"step": 433
},
{
"epoch": 0.014171428571428571,
"grad_norm": 3.011812686920166,
"learning_rate": 9.987459177008492e-06,
"loss": 0.7936,
"step": 434
},
{
"epoch": 0.01420408163265306,
"grad_norm": 2.8212246894836426,
"learning_rate": 9.987426518615285e-06,
"loss": 0.8028,
"step": 435
},
{
"epoch": 0.014236734693877552,
"grad_norm": 2.7385244369506836,
"learning_rate": 9.987393860222078e-06,
"loss": 0.8292,
"step": 436
},
{
"epoch": 0.014269387755102041,
"grad_norm": 2.7048816680908203,
"learning_rate": 9.987361201828871e-06,
"loss": 0.7567,
"step": 437
},
{
"epoch": 0.014302040816326531,
"grad_norm": 3.001534938812256,
"learning_rate": 9.987328543435663e-06,
"loss": 0.7897,
"step": 438
},
{
"epoch": 0.01433469387755102,
"grad_norm": 3.1388514041900635,
"learning_rate": 9.987295885042456e-06,
"loss": 0.8373,
"step": 439
},
{
"epoch": 0.01436734693877551,
"grad_norm": 2.902858018875122,
"learning_rate": 9.98726322664925e-06,
"loss": 0.8342,
"step": 440
},
{
"epoch": 0.0144,
"grad_norm": 2.871511459350586,
"learning_rate": 9.987230568256043e-06,
"loss": 0.8107,
"step": 441
},
{
"epoch": 0.01443265306122449,
"grad_norm": 3.0739245414733887,
"learning_rate": 9.987197909862836e-06,
"loss": 0.8171,
"step": 442
},
{
"epoch": 0.014465306122448979,
"grad_norm": 2.9680142402648926,
"learning_rate": 9.987165251469629e-06,
"loss": 0.8492,
"step": 443
},
{
"epoch": 0.01449795918367347,
"grad_norm": 2.763535737991333,
"learning_rate": 9.987132593076422e-06,
"loss": 0.8201,
"step": 444
},
{
"epoch": 0.01453061224489796,
"grad_norm": 2.8160109519958496,
"learning_rate": 9.987099934683214e-06,
"loss": 0.8175,
"step": 445
},
{
"epoch": 0.01456326530612245,
"grad_norm": 2.746919870376587,
"learning_rate": 9.987067276290007e-06,
"loss": 0.8121,
"step": 446
},
{
"epoch": 0.014595918367346939,
"grad_norm": 2.9417343139648438,
"learning_rate": 9.9870346178968e-06,
"loss": 0.8211,
"step": 447
},
{
"epoch": 0.014628571428571428,
"grad_norm": 2.886810302734375,
"learning_rate": 9.987001959503594e-06,
"loss": 0.7889,
"step": 448
},
{
"epoch": 0.014661224489795918,
"grad_norm": 2.89780592918396,
"learning_rate": 9.986969301110387e-06,
"loss": 0.8386,
"step": 449
},
{
"epoch": 0.014693877551020407,
"grad_norm": 2.8015475273132324,
"learning_rate": 9.986936642717178e-06,
"loss": 0.8023,
"step": 450
},
{
"epoch": 0.014693877551020407,
"eval_loss": 0.83209228515625,
"eval_runtime": 76.371,
"eval_samples_per_second": 1.309,
"eval_steps_per_second": 1.309,
"step": 450
},
{
"epoch": 0.014726530612244899,
"grad_norm": 3.0286951065063477,
"learning_rate": 9.986903984323972e-06,
"loss": 0.795,
"step": 451
},
{
"epoch": 0.014759183673469388,
"grad_norm": 2.798675060272217,
"learning_rate": 9.986871325930765e-06,
"loss": 0.778,
"step": 452
},
{
"epoch": 0.014791836734693878,
"grad_norm": 2.9744062423706055,
"learning_rate": 9.986838667537558e-06,
"loss": 0.8034,
"step": 453
},
{
"epoch": 0.014824489795918367,
"grad_norm": 3.131807804107666,
"learning_rate": 9.986806009144351e-06,
"loss": 0.8147,
"step": 454
},
{
"epoch": 0.014857142857142857,
"grad_norm": 2.759378671646118,
"learning_rate": 9.986773350751145e-06,
"loss": 0.7798,
"step": 455
},
{
"epoch": 0.014889795918367347,
"grad_norm": 2.79740047454834,
"learning_rate": 9.986740692357936e-06,
"loss": 0.793,
"step": 456
},
{
"epoch": 0.014922448979591836,
"grad_norm": 3.367506742477417,
"learning_rate": 9.98670803396473e-06,
"loss": 0.8295,
"step": 457
},
{
"epoch": 0.014955102040816327,
"grad_norm": 2.965843677520752,
"learning_rate": 9.986675375571523e-06,
"loss": 0.7908,
"step": 458
},
{
"epoch": 0.014987755102040817,
"grad_norm": 2.815945625305176,
"learning_rate": 9.986642717178316e-06,
"loss": 0.821,
"step": 459
},
{
"epoch": 0.015020408163265307,
"grad_norm": 2.8080756664276123,
"learning_rate": 9.986610058785109e-06,
"loss": 0.7952,
"step": 460
},
{
"epoch": 0.015053061224489796,
"grad_norm": 3.2081828117370605,
"learning_rate": 9.9865774003919e-06,
"loss": 0.8306,
"step": 461
},
{
"epoch": 0.015085714285714286,
"grad_norm": 2.866701602935791,
"learning_rate": 9.986544741998695e-06,
"loss": 0.8438,
"step": 462
},
{
"epoch": 0.015118367346938775,
"grad_norm": 2.806741952896118,
"learning_rate": 9.986512083605487e-06,
"loss": 0.8082,
"step": 463
},
{
"epoch": 0.015151020408163265,
"grad_norm": 2.8352231979370117,
"learning_rate": 9.98647942521228e-06,
"loss": 0.7805,
"step": 464
},
{
"epoch": 0.015183673469387754,
"grad_norm": 2.838469982147217,
"learning_rate": 9.986446766819073e-06,
"loss": 0.844,
"step": 465
},
{
"epoch": 0.015216326530612246,
"grad_norm": 2.7868754863739014,
"learning_rate": 9.986414108425867e-06,
"loss": 0.8078,
"step": 466
},
{
"epoch": 0.015248979591836735,
"grad_norm": 2.809109687805176,
"learning_rate": 9.98638145003266e-06,
"loss": 0.8425,
"step": 467
},
{
"epoch": 0.015281632653061225,
"grad_norm": 2.7377090454101562,
"learning_rate": 9.986348791639451e-06,
"loss": 0.8332,
"step": 468
},
{
"epoch": 0.015314285714285714,
"grad_norm": 2.863290786743164,
"learning_rate": 9.986316133246245e-06,
"loss": 0.8352,
"step": 469
},
{
"epoch": 0.015346938775510204,
"grad_norm": 2.720506429672241,
"learning_rate": 9.986283474853038e-06,
"loss": 0.8114,
"step": 470
},
{
"epoch": 0.015379591836734693,
"grad_norm": 2.7052462100982666,
"learning_rate": 9.986250816459831e-06,
"loss": 0.8093,
"step": 471
},
{
"epoch": 0.015412244897959183,
"grad_norm": 3.0364644527435303,
"learning_rate": 9.986218158066624e-06,
"loss": 0.8156,
"step": 472
},
{
"epoch": 0.015444897959183674,
"grad_norm": 2.9919137954711914,
"learning_rate": 9.986185499673416e-06,
"loss": 0.7906,
"step": 473
},
{
"epoch": 0.015477551020408164,
"grad_norm": 3.102023124694824,
"learning_rate": 9.98615284128021e-06,
"loss": 0.8247,
"step": 474
},
{
"epoch": 0.015510204081632653,
"grad_norm": 2.8093349933624268,
"learning_rate": 9.986120182887002e-06,
"loss": 0.8339,
"step": 475
},
{
"epoch": 0.015542857142857143,
"grad_norm": 2.893361806869507,
"learning_rate": 9.986087524493796e-06,
"loss": 0.7698,
"step": 476
},
{
"epoch": 0.015575510204081633,
"grad_norm": 2.813911199569702,
"learning_rate": 9.986054866100589e-06,
"loss": 0.8204,
"step": 477
},
{
"epoch": 0.015608163265306122,
"grad_norm": 2.7959954738616943,
"learning_rate": 9.986022207707382e-06,
"loss": 0.8099,
"step": 478
},
{
"epoch": 0.015640816326530613,
"grad_norm": 2.637242317199707,
"learning_rate": 9.985989549314174e-06,
"loss": 0.8218,
"step": 479
},
{
"epoch": 0.0156734693877551,
"grad_norm": 2.694908380508423,
"learning_rate": 9.985956890920967e-06,
"loss": 0.843,
"step": 480
},
{
"epoch": 0.015706122448979593,
"grad_norm": 2.65104603767395,
"learning_rate": 9.98592423252776e-06,
"loss": 0.8142,
"step": 481
},
{
"epoch": 0.01573877551020408,
"grad_norm": 2.7440900802612305,
"learning_rate": 9.985891574134553e-06,
"loss": 0.8079,
"step": 482
},
{
"epoch": 0.01577142857142857,
"grad_norm": 2.7440459728240967,
"learning_rate": 9.985858915741347e-06,
"loss": 0.8106,
"step": 483
},
{
"epoch": 0.015804081632653063,
"grad_norm": 2.8990628719329834,
"learning_rate": 9.98582625734814e-06,
"loss": 0.7954,
"step": 484
},
{
"epoch": 0.01583673469387755,
"grad_norm": 2.8682186603546143,
"learning_rate": 9.985793598954933e-06,
"loss": 0.7892,
"step": 485
},
{
"epoch": 0.015869387755102042,
"grad_norm": 3.1332759857177734,
"learning_rate": 9.985760940561725e-06,
"loss": 0.8182,
"step": 486
},
{
"epoch": 0.01590204081632653,
"grad_norm": 2.9516282081604004,
"learning_rate": 9.985728282168518e-06,
"loss": 0.7985,
"step": 487
},
{
"epoch": 0.01593469387755102,
"grad_norm": 3.3294899463653564,
"learning_rate": 9.985695623775311e-06,
"loss": 0.8368,
"step": 488
},
{
"epoch": 0.01596734693877551,
"grad_norm": 2.95438814163208,
"learning_rate": 9.985662965382104e-06,
"loss": 0.8151,
"step": 489
},
{
"epoch": 0.016,
"grad_norm": 2.7308006286621094,
"learning_rate": 9.985630306988898e-06,
"loss": 0.827,
"step": 490
},
{
"epoch": 0.016032653061224488,
"grad_norm": 2.695218801498413,
"learning_rate": 9.985597648595689e-06,
"loss": 0.7846,
"step": 491
},
{
"epoch": 0.01606530612244898,
"grad_norm": 2.842834949493408,
"learning_rate": 9.985564990202482e-06,
"loss": 0.783,
"step": 492
},
{
"epoch": 0.01609795918367347,
"grad_norm": 2.944671869277954,
"learning_rate": 9.985532331809276e-06,
"loss": 0.8273,
"step": 493
},
{
"epoch": 0.01613061224489796,
"grad_norm": 2.8893423080444336,
"learning_rate": 9.985499673416069e-06,
"loss": 0.8131,
"step": 494
},
{
"epoch": 0.01616326530612245,
"grad_norm": 2.917097568511963,
"learning_rate": 9.985467015022862e-06,
"loss": 0.8491,
"step": 495
},
{
"epoch": 0.016195918367346938,
"grad_norm": 2.678619384765625,
"learning_rate": 9.985434356629654e-06,
"loss": 0.7789,
"step": 496
},
{
"epoch": 0.01622857142857143,
"grad_norm": 2.86249041557312,
"learning_rate": 9.985401698236447e-06,
"loss": 0.8461,
"step": 497
},
{
"epoch": 0.016261224489795917,
"grad_norm": 2.9272475242614746,
"learning_rate": 9.98536903984324e-06,
"loss": 0.8047,
"step": 498
},
{
"epoch": 0.016293877551020408,
"grad_norm": 2.894817590713501,
"learning_rate": 9.985336381450033e-06,
"loss": 0.8157,
"step": 499
},
{
"epoch": 0.0163265306122449,
"grad_norm": 2.9469430446624756,
"learning_rate": 9.985303723056827e-06,
"loss": 0.8409,
"step": 500
},
{
"epoch": 0.0163265306122449,
"eval_loss": 0.8293061852455139,
"eval_runtime": 76.9824,
"eval_samples_per_second": 1.299,
"eval_steps_per_second": 1.299,
"step": 500
},
{
"epoch": 0.016359183673469387,
"grad_norm": 2.7225022315979004,
"learning_rate": 9.98527106466362e-06,
"loss": 0.7978,
"step": 501
},
{
"epoch": 0.01639183673469388,
"grad_norm": 2.6931450366973877,
"learning_rate": 9.985238406270411e-06,
"loss": 0.8089,
"step": 502
},
{
"epoch": 0.016424489795918366,
"grad_norm": 2.697438955307007,
"learning_rate": 9.985205747877206e-06,
"loss": 0.8231,
"step": 503
},
{
"epoch": 0.016457142857142858,
"grad_norm": 2.997002124786377,
"learning_rate": 9.985173089483998e-06,
"loss": 0.8211,
"step": 504
},
{
"epoch": 0.016489795918367346,
"grad_norm": 2.952876567840576,
"learning_rate": 9.985140431090791e-06,
"loss": 0.786,
"step": 505
},
{
"epoch": 0.016522448979591837,
"grad_norm": 3.085167407989502,
"learning_rate": 9.985107772697584e-06,
"loss": 0.8261,
"step": 506
},
{
"epoch": 0.016555102040816328,
"grad_norm": 2.8937699794769287,
"learning_rate": 9.985075114304378e-06,
"loss": 0.8009,
"step": 507
},
{
"epoch": 0.016587755102040816,
"grad_norm": 2.934492588043213,
"learning_rate": 9.98504245591117e-06,
"loss": 0.8059,
"step": 508
},
{
"epoch": 0.016620408163265307,
"grad_norm": 2.9281036853790283,
"learning_rate": 9.985009797517962e-06,
"loss": 0.8488,
"step": 509
},
{
"epoch": 0.016653061224489795,
"grad_norm": 2.8904941082000732,
"learning_rate": 9.984977139124756e-06,
"loss": 0.8431,
"step": 510
},
{
"epoch": 0.016685714285714286,
"grad_norm": 2.7994799613952637,
"learning_rate": 9.984944480731549e-06,
"loss": 0.8339,
"step": 511
},
{
"epoch": 0.016718367346938774,
"grad_norm": 3.0456082820892334,
"learning_rate": 9.984911822338342e-06,
"loss": 0.8077,
"step": 512
},
{
"epoch": 0.016751020408163265,
"grad_norm": 2.969820261001587,
"learning_rate": 9.984879163945135e-06,
"loss": 0.8254,
"step": 513
},
{
"epoch": 0.016783673469387757,
"grad_norm": 2.691143035888672,
"learning_rate": 9.984846505551927e-06,
"loss": 0.8208,
"step": 514
},
{
"epoch": 0.016816326530612245,
"grad_norm": 3.0714621543884277,
"learning_rate": 9.98481384715872e-06,
"loss": 0.8094,
"step": 515
},
{
"epoch": 0.016848979591836736,
"grad_norm": 3.146872043609619,
"learning_rate": 9.984781188765513e-06,
"loss": 0.8295,
"step": 516
},
{
"epoch": 0.016881632653061224,
"grad_norm": 3.0401713848114014,
"learning_rate": 9.984748530372307e-06,
"loss": 0.8438,
"step": 517
},
{
"epoch": 0.016914285714285715,
"grad_norm": 3.1565845012664795,
"learning_rate": 9.9847158719791e-06,
"loss": 0.8238,
"step": 518
},
{
"epoch": 0.016946938775510203,
"grad_norm": 3.035167932510376,
"learning_rate": 9.984683213585893e-06,
"loss": 0.7907,
"step": 519
},
{
"epoch": 0.016979591836734694,
"grad_norm": 2.9173972606658936,
"learning_rate": 9.984650555192685e-06,
"loss": 0.8298,
"step": 520
},
{
"epoch": 0.017012244897959182,
"grad_norm": 3.0475914478302,
"learning_rate": 9.984617896799478e-06,
"loss": 0.7896,
"step": 521
},
{
"epoch": 0.017044897959183673,
"grad_norm": 2.8846919536590576,
"learning_rate": 9.984585238406271e-06,
"loss": 0.8268,
"step": 522
},
{
"epoch": 0.017077551020408165,
"grad_norm": 2.9851925373077393,
"learning_rate": 9.984552580013064e-06,
"loss": 0.8166,
"step": 523
},
{
"epoch": 0.017110204081632652,
"grad_norm": 2.960608720779419,
"learning_rate": 9.984519921619857e-06,
"loss": 0.838,
"step": 524
},
{
"epoch": 0.017142857142857144,
"grad_norm": 2.8663570880889893,
"learning_rate": 9.984487263226649e-06,
"loss": 0.8134,
"step": 525
},
{
"epoch": 0.01717551020408163,
"grad_norm": 3.2871882915496826,
"learning_rate": 9.984454604833444e-06,
"loss": 0.842,
"step": 526
},
{
"epoch": 0.017208163265306123,
"grad_norm": 2.9747695922851562,
"learning_rate": 9.984421946440235e-06,
"loss": 0.8371,
"step": 527
},
{
"epoch": 0.01724081632653061,
"grad_norm": 3.065366744995117,
"learning_rate": 9.984389288047029e-06,
"loss": 0.799,
"step": 528
},
{
"epoch": 0.017273469387755102,
"grad_norm": 3.3719398975372314,
"learning_rate": 9.984356629653822e-06,
"loss": 0.7875,
"step": 529
},
{
"epoch": 0.017306122448979593,
"grad_norm": 3.0407257080078125,
"learning_rate": 9.984323971260615e-06,
"loss": 0.833,
"step": 530
},
{
"epoch": 0.01733877551020408,
"grad_norm": 2.992464780807495,
"learning_rate": 9.984291312867408e-06,
"loss": 0.83,
"step": 531
},
{
"epoch": 0.017371428571428572,
"grad_norm": 2.655125379562378,
"learning_rate": 9.9842586544742e-06,
"loss": 0.8243,
"step": 532
},
{
"epoch": 0.01740408163265306,
"grad_norm": 2.7782280445098877,
"learning_rate": 9.984225996080993e-06,
"loss": 0.8279,
"step": 533
},
{
"epoch": 0.01743673469387755,
"grad_norm": 2.6449782848358154,
"learning_rate": 9.984193337687786e-06,
"loss": 0.8517,
"step": 534
},
{
"epoch": 0.01746938775510204,
"grad_norm": 3.011317491531372,
"learning_rate": 9.98416067929458e-06,
"loss": 0.7763,
"step": 535
},
{
"epoch": 0.01750204081632653,
"grad_norm": 2.915745496749878,
"learning_rate": 9.984128020901373e-06,
"loss": 0.8283,
"step": 536
},
{
"epoch": 0.017534693877551022,
"grad_norm": 2.7176997661590576,
"learning_rate": 9.984095362508164e-06,
"loss": 0.7846,
"step": 537
},
{
"epoch": 0.01756734693877551,
"grad_norm": 2.9343252182006836,
"learning_rate": 9.984062704114958e-06,
"loss": 0.7733,
"step": 538
},
{
"epoch": 0.0176,
"grad_norm": 3.041198492050171,
"learning_rate": 9.984030045721751e-06,
"loss": 0.8024,
"step": 539
},
{
"epoch": 0.01763265306122449,
"grad_norm": 2.8848764896392822,
"learning_rate": 9.983997387328544e-06,
"loss": 0.8064,
"step": 540
},
{
"epoch": 0.01766530612244898,
"grad_norm": 2.847885847091675,
"learning_rate": 9.983964728935337e-06,
"loss": 0.7638,
"step": 541
},
{
"epoch": 0.017697959183673468,
"grad_norm": 2.7816402912139893,
"learning_rate": 9.98393207054213e-06,
"loss": 0.7839,
"step": 542
},
{
"epoch": 0.01773061224489796,
"grad_norm": 2.8443193435668945,
"learning_rate": 9.983899412148922e-06,
"loss": 0.8005,
"step": 543
},
{
"epoch": 0.01776326530612245,
"grad_norm": 2.923306941986084,
"learning_rate": 9.983866753755717e-06,
"loss": 0.8048,
"step": 544
},
{
"epoch": 0.01779591836734694,
"grad_norm": 3.0685269832611084,
"learning_rate": 9.983834095362509e-06,
"loss": 0.7676,
"step": 545
},
{
"epoch": 0.01782857142857143,
"grad_norm": 2.841459274291992,
"learning_rate": 9.983801436969302e-06,
"loss": 0.7828,
"step": 546
},
{
"epoch": 0.017861224489795918,
"grad_norm": 3.1108107566833496,
"learning_rate": 9.983768778576095e-06,
"loss": 0.8287,
"step": 547
},
{
"epoch": 0.01789387755102041,
"grad_norm": 2.8307478427886963,
"learning_rate": 9.983736120182888e-06,
"loss": 0.7956,
"step": 548
},
{
"epoch": 0.017926530612244897,
"grad_norm": 2.83941650390625,
"learning_rate": 9.983703461789682e-06,
"loss": 0.7963,
"step": 549
},
{
"epoch": 0.017959183673469388,
"grad_norm": 2.710853338241577,
"learning_rate": 9.983670803396473e-06,
"loss": 0.8023,
"step": 550
},
{
"epoch": 0.017959183673469388,
"eval_loss": 0.8259984850883484,
"eval_runtime": 98.251,
"eval_samples_per_second": 1.018,
"eval_steps_per_second": 1.018,
"step": 550
},
{
"epoch": 0.01799183673469388,
"grad_norm": 2.7674920558929443,
"learning_rate": 9.983638145003266e-06,
"loss": 0.8121,
"step": 551
},
{
"epoch": 0.018024489795918367,
"grad_norm": 2.756206512451172,
"learning_rate": 9.98360548661006e-06,
"loss": 0.7886,
"step": 552
},
{
"epoch": 0.01805714285714286,
"grad_norm": 3.000251054763794,
"learning_rate": 9.983572828216853e-06,
"loss": 0.8181,
"step": 553
},
{
"epoch": 0.018089795918367346,
"grad_norm": 2.7608935832977295,
"learning_rate": 9.983540169823646e-06,
"loss": 0.8047,
"step": 554
},
{
"epoch": 0.018122448979591838,
"grad_norm": 2.7886579036712646,
"learning_rate": 9.983507511430438e-06,
"loss": 0.8023,
"step": 555
},
{
"epoch": 0.018155102040816325,
"grad_norm": 2.7562673091888428,
"learning_rate": 9.983474853037231e-06,
"loss": 0.7935,
"step": 556
},
{
"epoch": 0.018187755102040817,
"grad_norm": 2.8150694370269775,
"learning_rate": 9.983442194644024e-06,
"loss": 0.8252,
"step": 557
},
{
"epoch": 0.018220408163265304,
"grad_norm": 3.069070339202881,
"learning_rate": 9.983409536250817e-06,
"loss": 0.8352,
"step": 558
},
{
"epoch": 0.018253061224489796,
"grad_norm": 3.065143585205078,
"learning_rate": 9.98337687785761e-06,
"loss": 0.8322,
"step": 559
},
{
"epoch": 0.018285714285714287,
"grad_norm": 2.808361530303955,
"learning_rate": 9.983344219464404e-06,
"loss": 0.8316,
"step": 560
},
{
"epoch": 0.018318367346938775,
"grad_norm": 3.086597204208374,
"learning_rate": 9.983311561071195e-06,
"loss": 0.8063,
"step": 561
},
{
"epoch": 0.018351020408163266,
"grad_norm": 2.8510525226593018,
"learning_rate": 9.983278902677989e-06,
"loss": 0.7735,
"step": 562
},
{
"epoch": 0.018383673469387754,
"grad_norm": 2.8289899826049805,
"learning_rate": 9.983246244284782e-06,
"loss": 0.831,
"step": 563
},
{
"epoch": 0.018416326530612245,
"grad_norm": 2.6663060188293457,
"learning_rate": 9.983213585891575e-06,
"loss": 0.8109,
"step": 564
},
{
"epoch": 0.018448979591836733,
"grad_norm": 2.8458051681518555,
"learning_rate": 9.983180927498368e-06,
"loss": 0.8069,
"step": 565
},
{
"epoch": 0.018481632653061224,
"grad_norm": 3.2364461421966553,
"learning_rate": 9.98314826910516e-06,
"loss": 0.8214,
"step": 566
},
{
"epoch": 0.018514285714285716,
"grad_norm": 3.0581226348876953,
"learning_rate": 9.983115610711955e-06,
"loss": 0.8115,
"step": 567
},
{
"epoch": 0.018546938775510204,
"grad_norm": 2.7239277362823486,
"learning_rate": 9.983082952318746e-06,
"loss": 0.8248,
"step": 568
},
{
"epoch": 0.018579591836734695,
"grad_norm": 2.827193260192871,
"learning_rate": 9.98305029392554e-06,
"loss": 0.8282,
"step": 569
},
{
"epoch": 0.018612244897959183,
"grad_norm": 3.018669605255127,
"learning_rate": 9.983017635532333e-06,
"loss": 0.8196,
"step": 570
},
{
"epoch": 0.018644897959183674,
"grad_norm": 2.9022018909454346,
"learning_rate": 9.982984977139126e-06,
"loss": 0.7748,
"step": 571
},
{
"epoch": 0.018677551020408162,
"grad_norm": 2.7094552516937256,
"learning_rate": 9.98295231874592e-06,
"loss": 0.8381,
"step": 572
},
{
"epoch": 0.018710204081632653,
"grad_norm": 3.020749568939209,
"learning_rate": 9.98291966035271e-06,
"loss": 0.7954,
"step": 573
},
{
"epoch": 0.018742857142857144,
"grad_norm": 2.9567782878875732,
"learning_rate": 9.982887001959504e-06,
"loss": 0.8197,
"step": 574
},
{
"epoch": 0.018775510204081632,
"grad_norm": 2.939793348312378,
"learning_rate": 9.982854343566297e-06,
"loss": 0.7891,
"step": 575
},
{
"epoch": 0.018808163265306124,
"grad_norm": 2.851130485534668,
"learning_rate": 9.98282168517309e-06,
"loss": 0.8669,
"step": 576
},
{
"epoch": 0.01884081632653061,
"grad_norm": 2.745495080947876,
"learning_rate": 9.982789026779884e-06,
"loss": 0.8028,
"step": 577
},
{
"epoch": 0.018873469387755103,
"grad_norm": 2.83685302734375,
"learning_rate": 9.982756368386675e-06,
"loss": 0.7981,
"step": 578
},
{
"epoch": 0.01890612244897959,
"grad_norm": 2.845992088317871,
"learning_rate": 9.982723709993469e-06,
"loss": 0.7752,
"step": 579
},
{
"epoch": 0.018938775510204082,
"grad_norm": 3.1280617713928223,
"learning_rate": 9.982691051600262e-06,
"loss": 0.824,
"step": 580
},
{
"epoch": 0.018971428571428573,
"grad_norm": 2.8821194171905518,
"learning_rate": 9.982658393207055e-06,
"loss": 0.7925,
"step": 581
},
{
"epoch": 0.01900408163265306,
"grad_norm": 2.827406167984009,
"learning_rate": 9.982625734813848e-06,
"loss": 0.8311,
"step": 582
},
{
"epoch": 0.019036734693877552,
"grad_norm": 2.7628026008605957,
"learning_rate": 9.982593076420641e-06,
"loss": 0.8228,
"step": 583
},
{
"epoch": 0.01906938775510204,
"grad_norm": 2.8712172508239746,
"learning_rate": 9.982560418027433e-06,
"loss": 0.8014,
"step": 584
},
{
"epoch": 0.01910204081632653,
"grad_norm": 2.85799241065979,
"learning_rate": 9.982527759634226e-06,
"loss": 0.8281,
"step": 585
},
{
"epoch": 0.01913469387755102,
"grad_norm": 2.917358160018921,
"learning_rate": 9.98249510124102e-06,
"loss": 0.8049,
"step": 586
},
{
"epoch": 0.01916734693877551,
"grad_norm": 2.8612101078033447,
"learning_rate": 9.982462442847813e-06,
"loss": 0.7645,
"step": 587
},
{
"epoch": 0.0192,
"grad_norm": 2.8899152278900146,
"learning_rate": 9.982429784454606e-06,
"loss": 0.7655,
"step": 588
},
{
"epoch": 0.01923265306122449,
"grad_norm": 2.761504888534546,
"learning_rate": 9.9823971260614e-06,
"loss": 0.784,
"step": 589
},
{
"epoch": 0.01926530612244898,
"grad_norm": 3.1624319553375244,
"learning_rate": 9.982364467668192e-06,
"loss": 0.8538,
"step": 590
},
{
"epoch": 0.01929795918367347,
"grad_norm": 2.840982675552368,
"learning_rate": 9.982331809274984e-06,
"loss": 0.7879,
"step": 591
},
{
"epoch": 0.01933061224489796,
"grad_norm": 2.8063557147979736,
"learning_rate": 9.982299150881777e-06,
"loss": 0.7858,
"step": 592
},
{
"epoch": 0.019363265306122448,
"grad_norm": 2.761817455291748,
"learning_rate": 9.98226649248857e-06,
"loss": 0.7932,
"step": 593
},
{
"epoch": 0.01939591836734694,
"grad_norm": 2.845803737640381,
"learning_rate": 9.982233834095364e-06,
"loss": 0.8315,
"step": 594
},
{
"epoch": 0.019428571428571427,
"grad_norm": 2.820237398147583,
"learning_rate": 9.982201175702157e-06,
"loss": 0.8115,
"step": 595
},
{
"epoch": 0.01946122448979592,
"grad_norm": 2.8963451385498047,
"learning_rate": 9.982168517308948e-06,
"loss": 0.7965,
"step": 596
},
{
"epoch": 0.01949387755102041,
"grad_norm": 2.8076157569885254,
"learning_rate": 9.982135858915742e-06,
"loss": 0.8205,
"step": 597
},
{
"epoch": 0.019526530612244897,
"grad_norm": 2.9265244007110596,
"learning_rate": 9.982103200522535e-06,
"loss": 0.7975,
"step": 598
},
{
"epoch": 0.01955918367346939,
"grad_norm": 3.044168472290039,
"learning_rate": 9.982070542129328e-06,
"loss": 0.7738,
"step": 599
},
{
"epoch": 0.019591836734693877,
"grad_norm": 2.991462469100952,
"learning_rate": 9.982037883736121e-06,
"loss": 0.808,
"step": 600
},
{
"epoch": 0.019591836734693877,
"eval_loss": 0.8216409087181091,
"eval_runtime": 96.6305,
"eval_samples_per_second": 1.035,
"eval_steps_per_second": 1.035,
"step": 600
},
{
"epoch": 0.019624489795918368,
"grad_norm": 2.856177568435669,
"learning_rate": 9.982005225342915e-06,
"loss": 0.8002,
"step": 601
},
{
"epoch": 0.019657142857142856,
"grad_norm": 2.881652355194092,
"learning_rate": 9.981972566949706e-06,
"loss": 0.8012,
"step": 602
},
{
"epoch": 0.019689795918367347,
"grad_norm": 2.9178857803344727,
"learning_rate": 9.9819399085565e-06,
"loss": 0.7724,
"step": 603
},
{
"epoch": 0.019722448979591838,
"grad_norm": 2.7211968898773193,
"learning_rate": 9.981907250163293e-06,
"loss": 0.798,
"step": 604
},
{
"epoch": 0.019755102040816326,
"grad_norm": 3.050656795501709,
"learning_rate": 9.981874591770086e-06,
"loss": 0.8564,
"step": 605
},
{
"epoch": 0.019787755102040817,
"grad_norm": 2.9224750995635986,
"learning_rate": 9.981841933376879e-06,
"loss": 0.8058,
"step": 606
},
{
"epoch": 0.019820408163265305,
"grad_norm": 2.9750068187713623,
"learning_rate": 9.98180927498367e-06,
"loss": 0.7774,
"step": 607
},
{
"epoch": 0.019853061224489797,
"grad_norm": 3.0843162536621094,
"learning_rate": 9.981776616590466e-06,
"loss": 0.8017,
"step": 608
},
{
"epoch": 0.019885714285714284,
"grad_norm": 3.043372631072998,
"learning_rate": 9.981743958197257e-06,
"loss": 0.7763,
"step": 609
},
{
"epoch": 0.019918367346938776,
"grad_norm": 2.8992819786071777,
"learning_rate": 9.98171129980405e-06,
"loss": 0.8483,
"step": 610
},
{
"epoch": 0.019951020408163267,
"grad_norm": 2.5121912956237793,
"learning_rate": 9.981678641410844e-06,
"loss": 0.8014,
"step": 611
},
{
"epoch": 0.019983673469387755,
"grad_norm": 2.8126091957092285,
"learning_rate": 9.981645983017637e-06,
"loss": 0.8,
"step": 612
},
{
"epoch": 0.020016326530612246,
"grad_norm": 2.760281801223755,
"learning_rate": 9.98161332462443e-06,
"loss": 0.8089,
"step": 613
},
{
"epoch": 0.020048979591836734,
"grad_norm": 2.75252103805542,
"learning_rate": 9.981580666231222e-06,
"loss": 0.7991,
"step": 614
},
{
"epoch": 0.020081632653061225,
"grad_norm": 2.7450368404388428,
"learning_rate": 9.981548007838015e-06,
"loss": 0.7995,
"step": 615
},
{
"epoch": 0.020114285714285713,
"grad_norm": 2.752061367034912,
"learning_rate": 9.981515349444808e-06,
"loss": 0.8167,
"step": 616
},
{
"epoch": 0.020146938775510204,
"grad_norm": 2.882789134979248,
"learning_rate": 9.981482691051601e-06,
"loss": 0.8438,
"step": 617
},
{
"epoch": 0.020179591836734692,
"grad_norm": 2.8994760513305664,
"learning_rate": 9.981450032658395e-06,
"loss": 0.7794,
"step": 618
},
{
"epoch": 0.020212244897959183,
"grad_norm": 2.925473928451538,
"learning_rate": 9.981417374265186e-06,
"loss": 0.798,
"step": 619
},
{
"epoch": 0.020244897959183675,
"grad_norm": 2.909290075302124,
"learning_rate": 9.98138471587198e-06,
"loss": 0.8487,
"step": 620
},
{
"epoch": 0.020277551020408163,
"grad_norm": 3.004021644592285,
"learning_rate": 9.981352057478773e-06,
"loss": 0.8214,
"step": 621
},
{
"epoch": 0.020310204081632654,
"grad_norm": 2.850106954574585,
"learning_rate": 9.981319399085566e-06,
"loss": 0.8158,
"step": 622
},
{
"epoch": 0.02034285714285714,
"grad_norm": 2.903092622756958,
"learning_rate": 9.981286740692359e-06,
"loss": 0.8078,
"step": 623
},
{
"epoch": 0.020375510204081633,
"grad_norm": 2.9145193099975586,
"learning_rate": 9.981254082299152e-06,
"loss": 0.8224,
"step": 624
},
{
"epoch": 0.02040816326530612,
"grad_norm": 3.0473721027374268,
"learning_rate": 9.981221423905944e-06,
"loss": 0.7681,
"step": 625
},
{
"epoch": 0.020440816326530612,
"grad_norm": 2.8615925312042236,
"learning_rate": 9.981188765512737e-06,
"loss": 0.815,
"step": 626
},
{
"epoch": 0.020473469387755103,
"grad_norm": 2.8572824001312256,
"learning_rate": 9.98115610711953e-06,
"loss": 0.7672,
"step": 627
},
{
"epoch": 0.02050612244897959,
"grad_norm": 2.9113521575927734,
"learning_rate": 9.981123448726324e-06,
"loss": 0.7952,
"step": 628
},
{
"epoch": 0.020538775510204083,
"grad_norm": 3.0004873275756836,
"learning_rate": 9.981090790333117e-06,
"loss": 0.8181,
"step": 629
},
{
"epoch": 0.02057142857142857,
"grad_norm": 2.9250986576080322,
"learning_rate": 9.981058131939908e-06,
"loss": 0.7681,
"step": 630
},
{
"epoch": 0.02060408163265306,
"grad_norm": 3.1994357109069824,
"learning_rate": 9.981025473546703e-06,
"loss": 0.8292,
"step": 631
},
{
"epoch": 0.02063673469387755,
"grad_norm": 3.061445713043213,
"learning_rate": 9.980992815153495e-06,
"loss": 0.8241,
"step": 632
},
{
"epoch": 0.02066938775510204,
"grad_norm": 3.1536760330200195,
"learning_rate": 9.980960156760288e-06,
"loss": 0.7889,
"step": 633
},
{
"epoch": 0.020702040816326532,
"grad_norm": 2.7358641624450684,
"learning_rate": 9.980927498367081e-06,
"loss": 0.8167,
"step": 634
},
{
"epoch": 0.02073469387755102,
"grad_norm": 2.952697992324829,
"learning_rate": 9.980894839973874e-06,
"loss": 0.7741,
"step": 635
},
{
"epoch": 0.02076734693877551,
"grad_norm": 3.003535747528076,
"learning_rate": 9.980862181580668e-06,
"loss": 0.8101,
"step": 636
},
{
"epoch": 0.0208,
"grad_norm": 2.895209789276123,
"learning_rate": 9.98082952318746e-06,
"loss": 0.8025,
"step": 637
},
{
"epoch": 0.02083265306122449,
"grad_norm": 2.974769115447998,
"learning_rate": 9.980796864794252e-06,
"loss": 0.8236,
"step": 638
},
{
"epoch": 0.020865306122448978,
"grad_norm": 2.8859550952911377,
"learning_rate": 9.980764206401046e-06,
"loss": 0.8214,
"step": 639
},
{
"epoch": 0.02089795918367347,
"grad_norm": 3.1916568279266357,
"learning_rate": 9.980731548007839e-06,
"loss": 0.7618,
"step": 640
},
{
"epoch": 0.02093061224489796,
"grad_norm": 3.1790311336517334,
"learning_rate": 9.980698889614632e-06,
"loss": 0.779,
"step": 641
},
{
"epoch": 0.02096326530612245,
"grad_norm": 9.082115173339844,
"learning_rate": 9.980666231221424e-06,
"loss": 0.7972,
"step": 642
},
{
"epoch": 0.02099591836734694,
"grad_norm": 3.0865111351013184,
"learning_rate": 9.980633572828217e-06,
"loss": 0.7972,
"step": 643
},
{
"epoch": 0.021028571428571428,
"grad_norm": 3.3228487968444824,
"learning_rate": 9.98060091443501e-06,
"loss": 0.8245,
"step": 644
},
{
"epoch": 0.02106122448979592,
"grad_norm": 2.7116079330444336,
"learning_rate": 9.980568256041803e-06,
"loss": 0.8129,
"step": 645
},
{
"epoch": 0.021093877551020407,
"grad_norm": 2.9123730659484863,
"learning_rate": 9.980535597648597e-06,
"loss": 0.832,
"step": 646
},
{
"epoch": 0.021126530612244898,
"grad_norm": 3.1061103343963623,
"learning_rate": 9.98050293925539e-06,
"loss": 0.8067,
"step": 647
},
{
"epoch": 0.02115918367346939,
"grad_norm": 3.2617485523223877,
"learning_rate": 9.980470280862181e-06,
"loss": 0.7897,
"step": 648
},
{
"epoch": 0.021191836734693877,
"grad_norm": 3.0000083446502686,
"learning_rate": 9.980437622468976e-06,
"loss": 0.7663,
"step": 649
},
{
"epoch": 0.02122448979591837,
"grad_norm": 3.029799461364746,
"learning_rate": 9.980404964075768e-06,
"loss": 0.7688,
"step": 650
},
{
"epoch": 0.02122448979591837,
"eval_loss": 0.8194996118545532,
"eval_runtime": 74.2652,
"eval_samples_per_second": 1.347,
"eval_steps_per_second": 1.347,
"step": 650
},
{
"epoch": 0.021257142857142856,
"grad_norm": 2.762091875076294,
"learning_rate": 9.980372305682561e-06,
"loss": 0.8051,
"step": 651
},
{
"epoch": 0.021289795918367348,
"grad_norm": 3.030320167541504,
"learning_rate": 9.980339647289354e-06,
"loss": 0.8201,
"step": 652
},
{
"epoch": 0.021322448979591836,
"grad_norm": 2.9875736236572266,
"learning_rate": 9.980306988896148e-06,
"loss": 0.7841,
"step": 653
},
{
"epoch": 0.021355102040816327,
"grad_norm": 2.847587823867798,
"learning_rate": 9.980274330502941e-06,
"loss": 0.7807,
"step": 654
},
{
"epoch": 0.021387755102040815,
"grad_norm": 3.0852084159851074,
"learning_rate": 9.980241672109732e-06,
"loss": 0.7606,
"step": 655
},
{
"epoch": 0.021420408163265306,
"grad_norm": 3.159208297729492,
"learning_rate": 9.980209013716526e-06,
"loss": 0.8011,
"step": 656
},
{
"epoch": 0.021453061224489797,
"grad_norm": 3.063185214996338,
"learning_rate": 9.980176355323319e-06,
"loss": 0.7839,
"step": 657
},
{
"epoch": 0.021485714285714285,
"grad_norm": 2.9843311309814453,
"learning_rate": 9.980143696930112e-06,
"loss": 0.7961,
"step": 658
},
{
"epoch": 0.021518367346938776,
"grad_norm": 3.0662293434143066,
"learning_rate": 9.980111038536905e-06,
"loss": 0.8065,
"step": 659
},
{
"epoch": 0.021551020408163264,
"grad_norm": 2.9915835857391357,
"learning_rate": 9.980078380143697e-06,
"loss": 0.8022,
"step": 660
},
{
"epoch": 0.021583673469387755,
"grad_norm": 2.7856531143188477,
"learning_rate": 9.98004572175049e-06,
"loss": 0.7853,
"step": 661
},
{
"epoch": 0.021616326530612243,
"grad_norm": 2.8300893306732178,
"learning_rate": 9.980013063357283e-06,
"loss": 0.8097,
"step": 662
},
{
"epoch": 0.021648979591836735,
"grad_norm": 2.7422966957092285,
"learning_rate": 9.979980404964077e-06,
"loss": 0.8307,
"step": 663
},
{
"epoch": 0.021681632653061226,
"grad_norm": 2.80092453956604,
"learning_rate": 9.97994774657087e-06,
"loss": 0.7932,
"step": 664
},
{
"epoch": 0.021714285714285714,
"grad_norm": 2.8894505500793457,
"learning_rate": 9.979915088177663e-06,
"loss": 0.8042,
"step": 665
},
{
"epoch": 0.021746938775510205,
"grad_norm": 2.7492778301239014,
"learning_rate": 9.979882429784455e-06,
"loss": 0.7817,
"step": 666
},
{
"epoch": 0.021779591836734693,
"grad_norm": 2.734226703643799,
"learning_rate": 9.979849771391248e-06,
"loss": 0.7984,
"step": 667
},
{
"epoch": 0.021812244897959184,
"grad_norm": 2.7127978801727295,
"learning_rate": 9.979817112998041e-06,
"loss": 0.7955,
"step": 668
},
{
"epoch": 0.021844897959183672,
"grad_norm": 2.881192207336426,
"learning_rate": 9.979784454604834e-06,
"loss": 0.8257,
"step": 669
},
{
"epoch": 0.021877551020408163,
"grad_norm": 3.02278995513916,
"learning_rate": 9.979751796211628e-06,
"loss": 0.7892,
"step": 670
},
{
"epoch": 0.021910204081632655,
"grad_norm": 3.157317876815796,
"learning_rate": 9.979719137818419e-06,
"loss": 0.8227,
"step": 671
},
{
"epoch": 0.021942857142857142,
"grad_norm": 2.876741409301758,
"learning_rate": 9.979686479425214e-06,
"loss": 0.8116,
"step": 672
},
{
"epoch": 0.021975510204081634,
"grad_norm": 3.1728057861328125,
"learning_rate": 9.979653821032006e-06,
"loss": 0.8207,
"step": 673
},
{
"epoch": 0.02200816326530612,
"grad_norm": 3.2653119564056396,
"learning_rate": 9.979621162638799e-06,
"loss": 0.8087,
"step": 674
},
{
"epoch": 0.022040816326530613,
"grad_norm": 3.0209734439849854,
"learning_rate": 9.979588504245592e-06,
"loss": 0.7958,
"step": 675
},
{
"epoch": 0.0220734693877551,
"grad_norm": 2.9444637298583984,
"learning_rate": 9.979555845852385e-06,
"loss": 0.8031,
"step": 676
},
{
"epoch": 0.022106122448979592,
"grad_norm": 2.8222014904022217,
"learning_rate": 9.979523187459179e-06,
"loss": 0.7842,
"step": 677
},
{
"epoch": 0.022138775510204083,
"grad_norm": 3.034346342086792,
"learning_rate": 9.97949052906597e-06,
"loss": 0.7979,
"step": 678
},
{
"epoch": 0.02217142857142857,
"grad_norm": 2.9454312324523926,
"learning_rate": 9.979457870672763e-06,
"loss": 0.8209,
"step": 679
},
{
"epoch": 0.022204081632653062,
"grad_norm": 2.9699203968048096,
"learning_rate": 9.979425212279557e-06,
"loss": 0.7783,
"step": 680
},
{
"epoch": 0.02223673469387755,
"grad_norm": 3.0586094856262207,
"learning_rate": 9.97939255388635e-06,
"loss": 0.8257,
"step": 681
},
{
"epoch": 0.02226938775510204,
"grad_norm": 3.1194658279418945,
"learning_rate": 9.979359895493143e-06,
"loss": 0.7824,
"step": 682
},
{
"epoch": 0.02230204081632653,
"grad_norm": 2.950078248977661,
"learning_rate": 9.979327237099935e-06,
"loss": 0.7866,
"step": 683
},
{
"epoch": 0.02233469387755102,
"grad_norm": 3.135443925857544,
"learning_rate": 9.979294578706728e-06,
"loss": 0.786,
"step": 684
},
{
"epoch": 0.02236734693877551,
"grad_norm": 3.1549785137176514,
"learning_rate": 9.979261920313521e-06,
"loss": 0.8063,
"step": 685
},
{
"epoch": 0.0224,
"grad_norm": 3.064702033996582,
"learning_rate": 9.979229261920314e-06,
"loss": 0.7789,
"step": 686
},
{
"epoch": 0.02243265306122449,
"grad_norm": 2.9797706604003906,
"learning_rate": 9.979196603527108e-06,
"loss": 0.8022,
"step": 687
},
{
"epoch": 0.02246530612244898,
"grad_norm": 3.0248265266418457,
"learning_rate": 9.9791639451339e-06,
"loss": 0.7935,
"step": 688
},
{
"epoch": 0.02249795918367347,
"grad_norm": 2.865000009536743,
"learning_rate": 9.979131286740692e-06,
"loss": 0.7539,
"step": 689
},
{
"epoch": 0.022530612244897958,
"grad_norm": 2.7899985313415527,
"learning_rate": 9.979098628347487e-06,
"loss": 0.766,
"step": 690
},
{
"epoch": 0.02256326530612245,
"grad_norm": 3.016523838043213,
"learning_rate": 9.979065969954279e-06,
"loss": 0.7904,
"step": 691
},
{
"epoch": 0.022595918367346937,
"grad_norm": 2.954990863800049,
"learning_rate": 9.979033311561072e-06,
"loss": 0.8276,
"step": 692
},
{
"epoch": 0.02262857142857143,
"grad_norm": 2.932527780532837,
"learning_rate": 9.979000653167865e-06,
"loss": 0.7614,
"step": 693
},
{
"epoch": 0.02266122448979592,
"grad_norm": 2.892082691192627,
"learning_rate": 9.978967994774657e-06,
"loss": 0.8026,
"step": 694
},
{
"epoch": 0.022693877551020408,
"grad_norm": 2.8288159370422363,
"learning_rate": 9.978935336381452e-06,
"loss": 0.7956,
"step": 695
},
{
"epoch": 0.0227265306122449,
"grad_norm": 2.8160433769226074,
"learning_rate": 9.978902677988243e-06,
"loss": 0.7765,
"step": 696
},
{
"epoch": 0.022759183673469387,
"grad_norm": 2.8383493423461914,
"learning_rate": 9.978870019595036e-06,
"loss": 0.7926,
"step": 697
},
{
"epoch": 0.022791836734693878,
"grad_norm": 3.058093786239624,
"learning_rate": 9.97883736120183e-06,
"loss": 0.7878,
"step": 698
},
{
"epoch": 0.022824489795918366,
"grad_norm": 3.0352556705474854,
"learning_rate": 9.978804702808623e-06,
"loss": 0.7885,
"step": 699
},
{
"epoch": 0.022857142857142857,
"grad_norm": 2.9252421855926514,
"learning_rate": 9.978772044415416e-06,
"loss": 0.7759,
"step": 700
},
{
"epoch": 0.022857142857142857,
"eval_loss": 0.8187767267227173,
"eval_runtime": 74.1789,
"eval_samples_per_second": 1.348,
"eval_steps_per_second": 1.348,
"step": 700
},
{
"epoch": 0.02288979591836735,
"grad_norm": 2.860393762588501,
"learning_rate": 9.978739386022208e-06,
"loss": 0.8177,
"step": 701
},
{
"epoch": 0.022922448979591836,
"grad_norm": 2.960848331451416,
"learning_rate": 9.978706727629001e-06,
"loss": 0.8298,
"step": 702
},
{
"epoch": 0.022955102040816328,
"grad_norm": 2.8897740840911865,
"learning_rate": 9.978674069235794e-06,
"loss": 0.7625,
"step": 703
},
{
"epoch": 0.022987755102040815,
"grad_norm": 3.0318431854248047,
"learning_rate": 9.978641410842587e-06,
"loss": 0.7759,
"step": 704
},
{
"epoch": 0.023020408163265307,
"grad_norm": 2.959118366241455,
"learning_rate": 9.97860875244938e-06,
"loss": 0.7921,
"step": 705
},
{
"epoch": 0.023053061224489795,
"grad_norm": 2.871290922164917,
"learning_rate": 9.978576094056174e-06,
"loss": 0.7943,
"step": 706
},
{
"epoch": 0.023085714285714286,
"grad_norm": 2.699939489364624,
"learning_rate": 9.978543435662965e-06,
"loss": 0.7708,
"step": 707
},
{
"epoch": 0.023118367346938777,
"grad_norm": 2.8150787353515625,
"learning_rate": 9.978510777269759e-06,
"loss": 0.7888,
"step": 708
},
{
"epoch": 0.023151020408163265,
"grad_norm": 2.9636101722717285,
"learning_rate": 9.978478118876552e-06,
"loss": 0.7861,
"step": 709
},
{
"epoch": 0.023183673469387756,
"grad_norm": 3.122624397277832,
"learning_rate": 9.978445460483345e-06,
"loss": 0.7967,
"step": 710
},
{
"epoch": 0.023216326530612244,
"grad_norm": 3.070082902908325,
"learning_rate": 9.978412802090138e-06,
"loss": 0.793,
"step": 711
},
{
"epoch": 0.023248979591836735,
"grad_norm": 3.063530206680298,
"learning_rate": 9.97838014369693e-06,
"loss": 0.8115,
"step": 712
},
{
"epoch": 0.023281632653061223,
"grad_norm": 2.8320116996765137,
"learning_rate": 9.978347485303725e-06,
"loss": 0.831,
"step": 713
},
{
"epoch": 0.023314285714285714,
"grad_norm": 2.9792158603668213,
"learning_rate": 9.978314826910516e-06,
"loss": 0.745,
"step": 714
},
{
"epoch": 0.023346938775510206,
"grad_norm": 2.9203131198883057,
"learning_rate": 9.97828216851731e-06,
"loss": 0.7912,
"step": 715
},
{
"epoch": 0.023379591836734694,
"grad_norm": 3.151167631149292,
"learning_rate": 9.978249510124103e-06,
"loss": 0.7687,
"step": 716
},
{
"epoch": 0.023412244897959185,
"grad_norm": 2.8722281455993652,
"learning_rate": 9.978216851730896e-06,
"loss": 0.7975,
"step": 717
},
{
"epoch": 0.023444897959183673,
"grad_norm": 2.989642858505249,
"learning_rate": 9.97818419333769e-06,
"loss": 0.8136,
"step": 718
},
{
"epoch": 0.023477551020408164,
"grad_norm": 3.2293035984039307,
"learning_rate": 9.978151534944481e-06,
"loss": 0.8469,
"step": 719
},
{
"epoch": 0.023510204081632652,
"grad_norm": 3.016899585723877,
"learning_rate": 9.978118876551274e-06,
"loss": 0.7901,
"step": 720
},
{
"epoch": 0.023542857142857143,
"grad_norm": 3.194211006164551,
"learning_rate": 9.978086218158067e-06,
"loss": 0.8112,
"step": 721
},
{
"epoch": 0.02357551020408163,
"grad_norm": 2.810654640197754,
"learning_rate": 9.97805355976486e-06,
"loss": 0.7878,
"step": 722
},
{
"epoch": 0.023608163265306122,
"grad_norm": 2.788594961166382,
"learning_rate": 9.978020901371654e-06,
"loss": 0.8245,
"step": 723
},
{
"epoch": 0.023640816326530614,
"grad_norm": 2.976698875427246,
"learning_rate": 9.977988242978445e-06,
"loss": 0.7869,
"step": 724
},
{
"epoch": 0.0236734693877551,
"grad_norm": 2.7467331886291504,
"learning_rate": 9.977955584585239e-06,
"loss": 0.7837,
"step": 725
},
{
"epoch": 0.023706122448979593,
"grad_norm": 2.8149194717407227,
"learning_rate": 9.977922926192032e-06,
"loss": 0.7713,
"step": 726
},
{
"epoch": 0.02373877551020408,
"grad_norm": 3.07743501663208,
"learning_rate": 9.977890267798825e-06,
"loss": 0.7519,
"step": 727
},
{
"epoch": 0.023771428571428572,
"grad_norm": 2.779167413711548,
"learning_rate": 9.977857609405618e-06,
"loss": 0.8015,
"step": 728
},
{
"epoch": 0.02380408163265306,
"grad_norm": 2.861452579498291,
"learning_rate": 9.977824951012412e-06,
"loss": 0.8097,
"step": 729
},
{
"epoch": 0.02383673469387755,
"grad_norm": 2.9489521980285645,
"learning_rate": 9.977792292619203e-06,
"loss": 0.7988,
"step": 730
},
{
"epoch": 0.023869387755102042,
"grad_norm": 2.7786002159118652,
"learning_rate": 9.977759634225996e-06,
"loss": 0.7828,
"step": 731
},
{
"epoch": 0.02390204081632653,
"grad_norm": 2.9412615299224854,
"learning_rate": 9.97772697583279e-06,
"loss": 0.7684,
"step": 732
},
{
"epoch": 0.02393469387755102,
"grad_norm": 2.9088659286499023,
"learning_rate": 9.977694317439583e-06,
"loss": 0.7899,
"step": 733
},
{
"epoch": 0.02396734693877551,
"grad_norm": 2.811553955078125,
"learning_rate": 9.977661659046376e-06,
"loss": 0.7705,
"step": 734
},
{
"epoch": 0.024,
"grad_norm": 2.7078611850738525,
"learning_rate": 9.977629000653168e-06,
"loss": 0.7803,
"step": 735
},
{
"epoch": 0.02403265306122449,
"grad_norm": 2.9198622703552246,
"learning_rate": 9.977596342259963e-06,
"loss": 0.794,
"step": 736
},
{
"epoch": 0.02406530612244898,
"grad_norm": 3.1856372356414795,
"learning_rate": 9.977563683866754e-06,
"loss": 0.8146,
"step": 737
},
{
"epoch": 0.02409795918367347,
"grad_norm": 2.914483070373535,
"learning_rate": 9.977531025473547e-06,
"loss": 0.813,
"step": 738
},
{
"epoch": 0.02413061224489796,
"grad_norm": 2.837502956390381,
"learning_rate": 9.97749836708034e-06,
"loss": 0.7807,
"step": 739
},
{
"epoch": 0.02416326530612245,
"grad_norm": 2.7452003955841064,
"learning_rate": 9.977465708687134e-06,
"loss": 0.8272,
"step": 740
},
{
"epoch": 0.024195918367346938,
"grad_norm": 2.922658681869507,
"learning_rate": 9.977433050293927e-06,
"loss": 0.8096,
"step": 741
},
{
"epoch": 0.02422857142857143,
"grad_norm": 2.9923369884490967,
"learning_rate": 9.977400391900719e-06,
"loss": 0.8232,
"step": 742
},
{
"epoch": 0.024261224489795917,
"grad_norm": 2.879516363143921,
"learning_rate": 9.977367733507512e-06,
"loss": 0.8128,
"step": 743
},
{
"epoch": 0.02429387755102041,
"grad_norm": 2.7809741497039795,
"learning_rate": 9.977335075114305e-06,
"loss": 0.7824,
"step": 744
},
{
"epoch": 0.0243265306122449,
"grad_norm": 2.8293521404266357,
"learning_rate": 9.977302416721098e-06,
"loss": 0.7637,
"step": 745
},
{
"epoch": 0.024359183673469387,
"grad_norm": 2.900975227355957,
"learning_rate": 9.977269758327891e-06,
"loss": 0.8065,
"step": 746
},
{
"epoch": 0.02439183673469388,
"grad_norm": 2.8670005798339844,
"learning_rate": 9.977237099934683e-06,
"loss": 0.7747,
"step": 747
},
{
"epoch": 0.024424489795918367,
"grad_norm": 2.987907886505127,
"learning_rate": 9.977204441541476e-06,
"loss": 0.7907,
"step": 748
},
{
"epoch": 0.024457142857142858,
"grad_norm": 2.8523452281951904,
"learning_rate": 9.97717178314827e-06,
"loss": 0.781,
"step": 749
},
{
"epoch": 0.024489795918367346,
"grad_norm": 2.9037506580352783,
"learning_rate": 9.977139124755063e-06,
"loss": 0.7505,
"step": 750
},
{
"epoch": 0.024489795918367346,
"eval_loss": 0.8159348964691162,
"eval_runtime": 74.1956,
"eval_samples_per_second": 1.348,
"eval_steps_per_second": 1.348,
"step": 750
},
{
"epoch": 0.024522448979591837,
"grad_norm": 2.8492870330810547,
"learning_rate": 9.977106466361856e-06,
"loss": 0.8186,
"step": 751
},
{
"epoch": 0.024555102040816325,
"grad_norm": 3.0403406620025635,
"learning_rate": 9.97707380796865e-06,
"loss": 0.8152,
"step": 752
},
{
"epoch": 0.024587755102040816,
"grad_norm": 2.8419723510742188,
"learning_rate": 9.97704114957544e-06,
"loss": 0.7908,
"step": 753
},
{
"epoch": 0.024620408163265307,
"grad_norm": 3.0176303386688232,
"learning_rate": 9.977008491182236e-06,
"loss": 0.7921,
"step": 754
},
{
"epoch": 0.024653061224489795,
"grad_norm": 2.928659439086914,
"learning_rate": 9.976975832789027e-06,
"loss": 0.7784,
"step": 755
},
{
"epoch": 0.024685714285714287,
"grad_norm": 2.800809621810913,
"learning_rate": 9.97694317439582e-06,
"loss": 0.7362,
"step": 756
},
{
"epoch": 0.024718367346938774,
"grad_norm": 2.698071002960205,
"learning_rate": 9.976910516002614e-06,
"loss": 0.8046,
"step": 757
},
{
"epoch": 0.024751020408163266,
"grad_norm": 2.8206839561462402,
"learning_rate": 9.976877857609407e-06,
"loss": 0.8169,
"step": 758
},
{
"epoch": 0.024783673469387753,
"grad_norm": 2.927217960357666,
"learning_rate": 9.9768451992162e-06,
"loss": 0.7816,
"step": 759
},
{
"epoch": 0.024816326530612245,
"grad_norm": 2.9845449924468994,
"learning_rate": 9.976812540822992e-06,
"loss": 0.801,
"step": 760
},
{
"epoch": 0.024848979591836736,
"grad_norm": 2.881765604019165,
"learning_rate": 9.976779882429785e-06,
"loss": 0.8052,
"step": 761
},
{
"epoch": 0.024881632653061224,
"grad_norm": 2.9570679664611816,
"learning_rate": 9.976747224036578e-06,
"loss": 0.8044,
"step": 762
},
{
"epoch": 0.024914285714285715,
"grad_norm": 3.069812774658203,
"learning_rate": 9.976714565643371e-06,
"loss": 0.8166,
"step": 763
},
{
"epoch": 0.024946938775510203,
"grad_norm": 3.0275254249572754,
"learning_rate": 9.976681907250165e-06,
"loss": 0.7835,
"step": 764
},
{
"epoch": 0.024979591836734694,
"grad_norm": 3.007643461227417,
"learning_rate": 9.976649248856956e-06,
"loss": 0.8088,
"step": 765
},
{
"epoch": 0.025012244897959182,
"grad_norm": 2.788733720779419,
"learning_rate": 9.97661659046375e-06,
"loss": 0.7852,
"step": 766
},
{
"epoch": 0.025044897959183673,
"grad_norm": 3.187893867492676,
"learning_rate": 9.976583932070543e-06,
"loss": 0.7887,
"step": 767
},
{
"epoch": 0.025077551020408165,
"grad_norm": 3.0475502014160156,
"learning_rate": 9.976551273677336e-06,
"loss": 0.7837,
"step": 768
},
{
"epoch": 0.025110204081632653,
"grad_norm": 2.9664242267608643,
"learning_rate": 9.976518615284129e-06,
"loss": 0.8021,
"step": 769
},
{
"epoch": 0.025142857142857144,
"grad_norm": 3.1118979454040527,
"learning_rate": 9.976485956890922e-06,
"loss": 0.7992,
"step": 770
},
{
"epoch": 0.02517551020408163,
"grad_norm": 2.909301280975342,
"learning_rate": 9.976453298497714e-06,
"loss": 0.743,
"step": 771
},
{
"epoch": 0.025208163265306123,
"grad_norm": 2.9557557106018066,
"learning_rate": 9.976420640104507e-06,
"loss": 0.8007,
"step": 772
},
{
"epoch": 0.02524081632653061,
"grad_norm": 3.108922243118286,
"learning_rate": 9.9763879817113e-06,
"loss": 0.8127,
"step": 773
},
{
"epoch": 0.025273469387755102,
"grad_norm": 2.927797555923462,
"learning_rate": 9.976355323318094e-06,
"loss": 0.7577,
"step": 774
},
{
"epoch": 0.025306122448979593,
"grad_norm": 2.929468870162964,
"learning_rate": 9.976322664924887e-06,
"loss": 0.7662,
"step": 775
},
{
"epoch": 0.02533877551020408,
"grad_norm": 3.0343034267425537,
"learning_rate": 9.976290006531678e-06,
"loss": 0.8283,
"step": 776
},
{
"epoch": 0.025371428571428573,
"grad_norm": 3.136800765991211,
"learning_rate": 9.976257348138473e-06,
"loss": 0.7935,
"step": 777
},
{
"epoch": 0.02540408163265306,
"grad_norm": 2.916213035583496,
"learning_rate": 9.976224689745265e-06,
"loss": 0.7651,
"step": 778
},
{
"epoch": 0.02543673469387755,
"grad_norm": 2.929622173309326,
"learning_rate": 9.976192031352058e-06,
"loss": 0.7677,
"step": 779
},
{
"epoch": 0.02546938775510204,
"grad_norm": 3.2419509887695312,
"learning_rate": 9.976159372958851e-06,
"loss": 0.7789,
"step": 780
},
{
"epoch": 0.02550204081632653,
"grad_norm": 2.8888614177703857,
"learning_rate": 9.976126714565645e-06,
"loss": 0.7735,
"step": 781
},
{
"epoch": 0.025534693877551022,
"grad_norm": 3.1505699157714844,
"learning_rate": 9.976094056172438e-06,
"loss": 0.8128,
"step": 782
},
{
"epoch": 0.02556734693877551,
"grad_norm": 3.030273675918579,
"learning_rate": 9.97606139777923e-06,
"loss": 0.7678,
"step": 783
},
{
"epoch": 0.0256,
"grad_norm": 3.0003929138183594,
"learning_rate": 9.976028739386023e-06,
"loss": 0.8076,
"step": 784
},
{
"epoch": 0.02563265306122449,
"grad_norm": 2.957667350769043,
"learning_rate": 9.975996080992816e-06,
"loss": 0.7886,
"step": 785
},
{
"epoch": 0.02566530612244898,
"grad_norm": 3.0816612243652344,
"learning_rate": 9.975963422599609e-06,
"loss": 0.7871,
"step": 786
},
{
"epoch": 0.025697959183673468,
"grad_norm": 2.947930097579956,
"learning_rate": 9.975930764206402e-06,
"loss": 0.8288,
"step": 787
},
{
"epoch": 0.02573061224489796,
"grad_norm": 3.121185302734375,
"learning_rate": 9.975898105813194e-06,
"loss": 0.793,
"step": 788
},
{
"epoch": 0.025763265306122447,
"grad_norm": 3.0452706813812256,
"learning_rate": 9.975865447419987e-06,
"loss": 0.7819,
"step": 789
},
{
"epoch": 0.02579591836734694,
"grad_norm": 2.903444528579712,
"learning_rate": 9.97583278902678e-06,
"loss": 0.798,
"step": 790
},
{
"epoch": 0.02582857142857143,
"grad_norm": 3.088996410369873,
"learning_rate": 9.975800130633574e-06,
"loss": 0.8171,
"step": 791
},
{
"epoch": 0.025861224489795918,
"grad_norm": 2.957293748855591,
"learning_rate": 9.975767472240367e-06,
"loss": 0.7862,
"step": 792
},
{
"epoch": 0.02589387755102041,
"grad_norm": 2.7569499015808105,
"learning_rate": 9.97573481384716e-06,
"loss": 0.7899,
"step": 793
},
{
"epoch": 0.025926530612244897,
"grad_norm": 2.931257724761963,
"learning_rate": 9.975702155453952e-06,
"loss": 0.8051,
"step": 794
},
{
"epoch": 0.025959183673469388,
"grad_norm": 2.9962334632873535,
"learning_rate": 9.975669497060746e-06,
"loss": 0.7914,
"step": 795
},
{
"epoch": 0.025991836734693876,
"grad_norm": 3.016993761062622,
"learning_rate": 9.975636838667538e-06,
"loss": 0.8011,
"step": 796
},
{
"epoch": 0.026024489795918367,
"grad_norm": 3.066710948944092,
"learning_rate": 9.975604180274331e-06,
"loss": 0.7885,
"step": 797
},
{
"epoch": 0.02605714285714286,
"grad_norm": 2.9981632232666016,
"learning_rate": 9.975571521881125e-06,
"loss": 0.7968,
"step": 798
},
{
"epoch": 0.026089795918367346,
"grad_norm": 2.971733331680298,
"learning_rate": 9.975538863487916e-06,
"loss": 0.7997,
"step": 799
},
{
"epoch": 0.026122448979591838,
"grad_norm": 3.0647032260894775,
"learning_rate": 9.975506205094711e-06,
"loss": 0.8122,
"step": 800
},
{
"epoch": 0.026122448979591838,
"eval_loss": 0.8141899108886719,
"eval_runtime": 73.6209,
"eval_samples_per_second": 1.358,
"eval_steps_per_second": 1.358,
"step": 800
},
{
"epoch": 0.026155102040816326,
"grad_norm": 3.015854597091675,
"learning_rate": 9.975473546701503e-06,
"loss": 0.8062,
"step": 801
},
{
"epoch": 0.026187755102040817,
"grad_norm": 3.065284013748169,
"learning_rate": 9.975440888308296e-06,
"loss": 0.7973,
"step": 802
},
{
"epoch": 0.026220408163265305,
"grad_norm": 2.8390414714813232,
"learning_rate": 9.975408229915089e-06,
"loss": 0.7943,
"step": 803
},
{
"epoch": 0.026253061224489796,
"grad_norm": 3.254967451095581,
"learning_rate": 9.975375571521882e-06,
"loss": 0.8059,
"step": 804
},
{
"epoch": 0.026285714285714287,
"grad_norm": 3.0705697536468506,
"learning_rate": 9.975342913128675e-06,
"loss": 0.7871,
"step": 805
},
{
"epoch": 0.026318367346938775,
"grad_norm": 2.900188446044922,
"learning_rate": 9.975310254735467e-06,
"loss": 0.7672,
"step": 806
},
{
"epoch": 0.026351020408163266,
"grad_norm": 2.8207056522369385,
"learning_rate": 9.97527759634226e-06,
"loss": 0.7921,
"step": 807
},
{
"epoch": 0.026383673469387754,
"grad_norm": 2.891328811645508,
"learning_rate": 9.975244937949053e-06,
"loss": 0.8177,
"step": 808
},
{
"epoch": 0.026416326530612246,
"grad_norm": 2.974545955657959,
"learning_rate": 9.975212279555847e-06,
"loss": 0.7773,
"step": 809
},
{
"epoch": 0.026448979591836733,
"grad_norm": 2.8445703983306885,
"learning_rate": 9.97517962116264e-06,
"loss": 0.8131,
"step": 810
},
{
"epoch": 0.026481632653061225,
"grad_norm": 2.93383526802063,
"learning_rate": 9.975146962769433e-06,
"loss": 0.7693,
"step": 811
},
{
"epoch": 0.026514285714285716,
"grad_norm": 3.0080511569976807,
"learning_rate": 9.975114304376225e-06,
"loss": 0.7593,
"step": 812
},
{
"epoch": 0.026546938775510204,
"grad_norm": 2.967305898666382,
"learning_rate": 9.975081645983018e-06,
"loss": 0.7773,
"step": 813
},
{
"epoch": 0.026579591836734695,
"grad_norm": 2.9201626777648926,
"learning_rate": 9.975048987589811e-06,
"loss": 0.7908,
"step": 814
},
{
"epoch": 0.026612244897959183,
"grad_norm": 3.0869758129119873,
"learning_rate": 9.975016329196604e-06,
"loss": 0.8015,
"step": 815
},
{
"epoch": 0.026644897959183674,
"grad_norm": 3.120110511779785,
"learning_rate": 9.974983670803398e-06,
"loss": 0.7993,
"step": 816
},
{
"epoch": 0.026677551020408162,
"grad_norm": 3.1260945796966553,
"learning_rate": 9.97495101241019e-06,
"loss": 0.8407,
"step": 817
},
{
"epoch": 0.026710204081632653,
"grad_norm": 2.858689785003662,
"learning_rate": 9.974918354016984e-06,
"loss": 0.8001,
"step": 818
},
{
"epoch": 0.02674285714285714,
"grad_norm": 2.854126214981079,
"learning_rate": 9.974885695623776e-06,
"loss": 0.754,
"step": 819
},
{
"epoch": 0.026775510204081632,
"grad_norm": 2.932650566101074,
"learning_rate": 9.974853037230569e-06,
"loss": 0.8051,
"step": 820
},
{
"epoch": 0.026808163265306124,
"grad_norm": 3.063242197036743,
"learning_rate": 9.974820378837362e-06,
"loss": 0.8114,
"step": 821
},
{
"epoch": 0.02684081632653061,
"grad_norm": 2.8537795543670654,
"learning_rate": 9.974787720444155e-06,
"loss": 0.7842,
"step": 822
},
{
"epoch": 0.026873469387755103,
"grad_norm": 2.986083745956421,
"learning_rate": 9.974755062050949e-06,
"loss": 0.7949,
"step": 823
},
{
"epoch": 0.02690612244897959,
"grad_norm": 3.0178754329681396,
"learning_rate": 9.97472240365774e-06,
"loss": 0.8005,
"step": 824
},
{
"epoch": 0.026938775510204082,
"grad_norm": 3.1746835708618164,
"learning_rate": 9.974689745264533e-06,
"loss": 0.7994,
"step": 825
},
{
"epoch": 0.02697142857142857,
"grad_norm": 2.8827731609344482,
"learning_rate": 9.974657086871327e-06,
"loss": 0.7764,
"step": 826
},
{
"epoch": 0.02700408163265306,
"grad_norm": 2.7529618740081787,
"learning_rate": 9.97462442847812e-06,
"loss": 0.8092,
"step": 827
},
{
"epoch": 0.027036734693877552,
"grad_norm": 2.8854501247406006,
"learning_rate": 9.974591770084913e-06,
"loss": 0.7713,
"step": 828
},
{
"epoch": 0.02706938775510204,
"grad_norm": 2.8492770195007324,
"learning_rate": 9.974559111691705e-06,
"loss": 0.75,
"step": 829
},
{
"epoch": 0.02710204081632653,
"grad_norm": 2.9942445755004883,
"learning_rate": 9.974526453298498e-06,
"loss": 0.7838,
"step": 830
},
{
"epoch": 0.02713469387755102,
"grad_norm": 2.9380862712860107,
"learning_rate": 9.974493794905291e-06,
"loss": 0.8111,
"step": 831
},
{
"epoch": 0.02716734693877551,
"grad_norm": 2.8200275897979736,
"learning_rate": 9.974461136512084e-06,
"loss": 0.7818,
"step": 832
},
{
"epoch": 0.0272,
"grad_norm": 3.0098073482513428,
"learning_rate": 9.974428478118878e-06,
"loss": 0.7786,
"step": 833
},
{
"epoch": 0.02723265306122449,
"grad_norm": 2.96885347366333,
"learning_rate": 9.97439581972567e-06,
"loss": 0.762,
"step": 834
},
{
"epoch": 0.02726530612244898,
"grad_norm": 2.9910387992858887,
"learning_rate": 9.974363161332462e-06,
"loss": 0.8142,
"step": 835
},
{
"epoch": 0.02729795918367347,
"grad_norm": 3.1752495765686035,
"learning_rate": 9.974330502939257e-06,
"loss": 0.8031,
"step": 836
},
{
"epoch": 0.02733061224489796,
"grad_norm": 2.9318926334381104,
"learning_rate": 9.974297844546049e-06,
"loss": 0.7873,
"step": 837
},
{
"epoch": 0.027363265306122448,
"grad_norm": 2.8536527156829834,
"learning_rate": 9.974265186152842e-06,
"loss": 0.7878,
"step": 838
},
{
"epoch": 0.02739591836734694,
"grad_norm": 2.7921385765075684,
"learning_rate": 9.974232527759635e-06,
"loss": 0.8042,
"step": 839
},
{
"epoch": 0.027428571428571427,
"grad_norm": 2.97587251663208,
"learning_rate": 9.974199869366427e-06,
"loss": 0.8044,
"step": 840
},
{
"epoch": 0.02746122448979592,
"grad_norm": 2.879415273666382,
"learning_rate": 9.974167210973222e-06,
"loss": 0.7902,
"step": 841
},
{
"epoch": 0.02749387755102041,
"grad_norm": 2.930534839630127,
"learning_rate": 9.974134552580013e-06,
"loss": 0.7809,
"step": 842
},
{
"epoch": 0.027526530612244898,
"grad_norm": 2.8264150619506836,
"learning_rate": 9.974101894186807e-06,
"loss": 0.7565,
"step": 843
},
{
"epoch": 0.02755918367346939,
"grad_norm": 2.879453182220459,
"learning_rate": 9.9740692357936e-06,
"loss": 0.8139,
"step": 844
},
{
"epoch": 0.027591836734693877,
"grad_norm": 2.805103063583374,
"learning_rate": 9.974036577400393e-06,
"loss": 0.798,
"step": 845
},
{
"epoch": 0.027624489795918368,
"grad_norm": 2.8710720539093018,
"learning_rate": 9.974003919007186e-06,
"loss": 0.7412,
"step": 846
},
{
"epoch": 0.027657142857142856,
"grad_norm": 2.9521493911743164,
"learning_rate": 9.973971260613978e-06,
"loss": 0.7943,
"step": 847
},
{
"epoch": 0.027689795918367347,
"grad_norm": 3.062432050704956,
"learning_rate": 9.973938602220771e-06,
"loss": 0.7884,
"step": 848
},
{
"epoch": 0.02772244897959184,
"grad_norm": 2.98431658744812,
"learning_rate": 9.973905943827564e-06,
"loss": 0.7643,
"step": 849
},
{
"epoch": 0.027755102040816326,
"grad_norm": 2.852004289627075,
"learning_rate": 9.973873285434358e-06,
"loss": 0.8126,
"step": 850
},
{
"epoch": 0.027755102040816326,
"eval_loss": 0.8143028020858765,
"eval_runtime": 85.3794,
"eval_samples_per_second": 1.171,
"eval_steps_per_second": 1.171,
"step": 850
},
{
"epoch": 0.027787755102040818,
"grad_norm": 3.2032933235168457,
"learning_rate": 9.97384062704115e-06,
"loss": 0.7847,
"step": 851
},
{
"epoch": 0.027820408163265305,
"grad_norm": 3.0217716693878174,
"learning_rate": 9.973807968647944e-06,
"loss": 0.7812,
"step": 852
},
{
"epoch": 0.027853061224489797,
"grad_norm": 2.961899995803833,
"learning_rate": 9.973775310254736e-06,
"loss": 0.7917,
"step": 853
},
{
"epoch": 0.027885714285714285,
"grad_norm": 2.9481585025787354,
"learning_rate": 9.973742651861529e-06,
"loss": 0.7866,
"step": 854
},
{
"epoch": 0.027918367346938776,
"grad_norm": 2.8054749965667725,
"learning_rate": 9.973709993468322e-06,
"loss": 0.7808,
"step": 855
},
{
"epoch": 0.027951020408163264,
"grad_norm": 3.106367349624634,
"learning_rate": 9.973677335075115e-06,
"loss": 0.8281,
"step": 856
},
{
"epoch": 0.027983673469387755,
"grad_norm": 3.7496819496154785,
"learning_rate": 9.973644676681908e-06,
"loss": 0.7851,
"step": 857
},
{
"epoch": 0.028016326530612246,
"grad_norm": 2.7985293865203857,
"learning_rate": 9.9736120182887e-06,
"loss": 0.7843,
"step": 858
},
{
"epoch": 0.028048979591836734,
"grad_norm": 2.8734278678894043,
"learning_rate": 9.973579359895495e-06,
"loss": 0.7482,
"step": 859
},
{
"epoch": 0.028081632653061225,
"grad_norm": 3.0287320613861084,
"learning_rate": 9.973546701502287e-06,
"loss": 0.7901,
"step": 860
},
{
"epoch": 0.028114285714285713,
"grad_norm": 3.2069811820983887,
"learning_rate": 9.97351404310908e-06,
"loss": 0.7762,
"step": 861
},
{
"epoch": 0.028146938775510204,
"grad_norm": 3.193441152572632,
"learning_rate": 9.973481384715873e-06,
"loss": 0.7867,
"step": 862
},
{
"epoch": 0.028179591836734692,
"grad_norm": 2.932454824447632,
"learning_rate": 9.973448726322665e-06,
"loss": 0.801,
"step": 863
},
{
"epoch": 0.028212244897959184,
"grad_norm": 3.009593963623047,
"learning_rate": 9.97341606792946e-06,
"loss": 0.7612,
"step": 864
},
{
"epoch": 0.028244897959183675,
"grad_norm": 3.2580459117889404,
"learning_rate": 9.973383409536251e-06,
"loss": 0.7295,
"step": 865
},
{
"epoch": 0.028277551020408163,
"grad_norm": 2.934058904647827,
"learning_rate": 9.973350751143044e-06,
"loss": 0.7766,
"step": 866
},
{
"epoch": 0.028310204081632654,
"grad_norm": 3.0753285884857178,
"learning_rate": 9.973318092749837e-06,
"loss": 0.7647,
"step": 867
},
{
"epoch": 0.028342857142857142,
"grad_norm": 3.1178665161132812,
"learning_rate": 9.97328543435663e-06,
"loss": 0.8071,
"step": 868
},
{
"epoch": 0.028375510204081633,
"grad_norm": 3.235008955001831,
"learning_rate": 9.973252775963424e-06,
"loss": 0.7899,
"step": 869
},
{
"epoch": 0.02840816326530612,
"grad_norm": 3.198587417602539,
"learning_rate": 9.973220117570215e-06,
"loss": 0.7918,
"step": 870
},
{
"epoch": 0.028440816326530612,
"grad_norm": 3.1364388465881348,
"learning_rate": 9.973187459177009e-06,
"loss": 0.8178,
"step": 871
},
{
"epoch": 0.028473469387755104,
"grad_norm": 3.1181514263153076,
"learning_rate": 9.973154800783802e-06,
"loss": 0.7922,
"step": 872
},
{
"epoch": 0.02850612244897959,
"grad_norm": 3.0627710819244385,
"learning_rate": 9.973122142390595e-06,
"loss": 0.8063,
"step": 873
},
{
"epoch": 0.028538775510204083,
"grad_norm": 3.042802572250366,
"learning_rate": 9.973089483997388e-06,
"loss": 0.8304,
"step": 874
},
{
"epoch": 0.02857142857142857,
"grad_norm": 2.935323715209961,
"learning_rate": 9.973056825604182e-06,
"loss": 0.7983,
"step": 875
},
{
"epoch": 0.028604081632653062,
"grad_norm": 2.9549400806427,
"learning_rate": 9.973024167210973e-06,
"loss": 0.8063,
"step": 876
},
{
"epoch": 0.02863673469387755,
"grad_norm": 2.7875542640686035,
"learning_rate": 9.972991508817766e-06,
"loss": 0.7981,
"step": 877
},
{
"epoch": 0.02866938775510204,
"grad_norm": 2.9394519329071045,
"learning_rate": 9.97295885042456e-06,
"loss": 0.792,
"step": 878
},
{
"epoch": 0.028702040816326532,
"grad_norm": 3.0494203567504883,
"learning_rate": 9.972926192031353e-06,
"loss": 0.8128,
"step": 879
},
{
"epoch": 0.02873469387755102,
"grad_norm": 2.9686169624328613,
"learning_rate": 9.972893533638146e-06,
"loss": 0.7797,
"step": 880
},
{
"epoch": 0.02876734693877551,
"grad_norm": 2.964944839477539,
"learning_rate": 9.972860875244938e-06,
"loss": 0.7494,
"step": 881
},
{
"epoch": 0.0288,
"grad_norm": 2.9027297496795654,
"learning_rate": 9.972828216851733e-06,
"loss": 0.7526,
"step": 882
},
{
"epoch": 0.02883265306122449,
"grad_norm": 3.0997474193573,
"learning_rate": 9.972795558458524e-06,
"loss": 0.7878,
"step": 883
},
{
"epoch": 0.02886530612244898,
"grad_norm": 2.999011516571045,
"learning_rate": 9.972762900065317e-06,
"loss": 0.785,
"step": 884
},
{
"epoch": 0.02889795918367347,
"grad_norm": 3.1621158123016357,
"learning_rate": 9.97273024167211e-06,
"loss": 0.8338,
"step": 885
},
{
"epoch": 0.028930612244897957,
"grad_norm": 2.943216323852539,
"learning_rate": 9.972697583278904e-06,
"loss": 0.7844,
"step": 886
},
{
"epoch": 0.02896326530612245,
"grad_norm": 3.0958337783813477,
"learning_rate": 9.972664924885697e-06,
"loss": 0.7895,
"step": 887
},
{
"epoch": 0.02899591836734694,
"grad_norm": 2.982701539993286,
"learning_rate": 9.972632266492489e-06,
"loss": 0.7783,
"step": 888
},
{
"epoch": 0.029028571428571428,
"grad_norm": 2.9064059257507324,
"learning_rate": 9.972599608099282e-06,
"loss": 0.7915,
"step": 889
},
{
"epoch": 0.02906122448979592,
"grad_norm": 2.8960936069488525,
"learning_rate": 9.972566949706075e-06,
"loss": 0.7914,
"step": 890
},
{
"epoch": 0.029093877551020407,
"grad_norm": 2.881122350692749,
"learning_rate": 9.972534291312868e-06,
"loss": 0.7673,
"step": 891
},
{
"epoch": 0.0291265306122449,
"grad_norm": 2.9828569889068604,
"learning_rate": 9.972501632919662e-06,
"loss": 0.8089,
"step": 892
},
{
"epoch": 0.029159183673469386,
"grad_norm": 3.0593910217285156,
"learning_rate": 9.972468974526453e-06,
"loss": 0.8079,
"step": 893
},
{
"epoch": 0.029191836734693877,
"grad_norm": 2.993431568145752,
"learning_rate": 9.972436316133246e-06,
"loss": 0.7855,
"step": 894
},
{
"epoch": 0.02922448979591837,
"grad_norm": 3.132528066635132,
"learning_rate": 9.97240365774004e-06,
"loss": 0.7948,
"step": 895
},
{
"epoch": 0.029257142857142857,
"grad_norm": 3.0954887866973877,
"learning_rate": 9.972370999346833e-06,
"loss": 0.765,
"step": 896
},
{
"epoch": 0.029289795918367348,
"grad_norm": 3.0066490173339844,
"learning_rate": 9.972338340953626e-06,
"loss": 0.7616,
"step": 897
},
{
"epoch": 0.029322448979591836,
"grad_norm": 3.05830717086792,
"learning_rate": 9.97230568256042e-06,
"loss": 0.7585,
"step": 898
},
{
"epoch": 0.029355102040816327,
"grad_norm": 3.2171273231506348,
"learning_rate": 9.972273024167211e-06,
"loss": 0.8138,
"step": 899
},
{
"epoch": 0.029387755102040815,
"grad_norm": 2.9297590255737305,
"learning_rate": 9.972240365774006e-06,
"loss": 0.762,
"step": 900
},
{
"epoch": 0.029387755102040815,
"eval_loss": 0.8132917284965515,
"eval_runtime": 76.0316,
"eval_samples_per_second": 1.315,
"eval_steps_per_second": 1.315,
"step": 900
},
{
"epoch": 0.029420408163265306,
"grad_norm": 2.8318891525268555,
"learning_rate": 9.972207707380797e-06,
"loss": 0.7571,
"step": 901
},
{
"epoch": 0.029453061224489797,
"grad_norm": 3.222458600997925,
"learning_rate": 9.97217504898759e-06,
"loss": 0.7869,
"step": 902
},
{
"epoch": 0.029485714285714285,
"grad_norm": 3.2787888050079346,
"learning_rate": 9.972142390594384e-06,
"loss": 0.7922,
"step": 903
},
{
"epoch": 0.029518367346938777,
"grad_norm": 2.9158124923706055,
"learning_rate": 9.972109732201175e-06,
"loss": 0.8008,
"step": 904
},
{
"epoch": 0.029551020408163264,
"grad_norm": 2.931711435317993,
"learning_rate": 9.97207707380797e-06,
"loss": 0.7958,
"step": 905
},
{
"epoch": 0.029583673469387756,
"grad_norm": 3.1747889518737793,
"learning_rate": 9.972044415414762e-06,
"loss": 0.7953,
"step": 906
},
{
"epoch": 0.029616326530612243,
"grad_norm": 3.23126482963562,
"learning_rate": 9.972011757021555e-06,
"loss": 0.8133,
"step": 907
},
{
"epoch": 0.029648979591836735,
"grad_norm": 2.8572208881378174,
"learning_rate": 9.971979098628348e-06,
"loss": 0.7832,
"step": 908
},
{
"epoch": 0.029681632653061226,
"grad_norm": 3.004584312438965,
"learning_rate": 9.971946440235142e-06,
"loss": 0.8025,
"step": 909
},
{
"epoch": 0.029714285714285714,
"grad_norm": 2.7122766971588135,
"learning_rate": 9.971913781841935e-06,
"loss": 0.7819,
"step": 910
},
{
"epoch": 0.029746938775510205,
"grad_norm": 2.9016056060791016,
"learning_rate": 9.971881123448726e-06,
"loss": 0.7852,
"step": 911
},
{
"epoch": 0.029779591836734693,
"grad_norm": 2.985293388366699,
"learning_rate": 9.97184846505552e-06,
"loss": 0.8294,
"step": 912
},
{
"epoch": 0.029812244897959184,
"grad_norm": 2.9722225666046143,
"learning_rate": 9.971815806662313e-06,
"loss": 0.7924,
"step": 913
},
{
"epoch": 0.029844897959183672,
"grad_norm": 2.8796117305755615,
"learning_rate": 9.971783148269106e-06,
"loss": 0.7549,
"step": 914
},
{
"epoch": 0.029877551020408163,
"grad_norm": 3.0465636253356934,
"learning_rate": 9.9717504898759e-06,
"loss": 0.7875,
"step": 915
},
{
"epoch": 0.029910204081632655,
"grad_norm": 3.179034948348999,
"learning_rate": 9.971717831482692e-06,
"loss": 0.774,
"step": 916
},
{
"epoch": 0.029942857142857143,
"grad_norm": 3.2824554443359375,
"learning_rate": 9.971685173089484e-06,
"loss": 0.788,
"step": 917
},
{
"epoch": 0.029975510204081634,
"grad_norm": 3.2578866481781006,
"learning_rate": 9.971652514696277e-06,
"loss": 0.7672,
"step": 918
},
{
"epoch": 0.03000816326530612,
"grad_norm": 3.1106038093566895,
"learning_rate": 9.97161985630307e-06,
"loss": 0.8133,
"step": 919
},
{
"epoch": 0.030040816326530613,
"grad_norm": 2.948033094406128,
"learning_rate": 9.971587197909864e-06,
"loss": 0.7819,
"step": 920
},
{
"epoch": 0.0300734693877551,
"grad_norm": 3.094900131225586,
"learning_rate": 9.971554539516657e-06,
"loss": 0.7833,
"step": 921
},
{
"epoch": 0.030106122448979592,
"grad_norm": 2.9197640419006348,
"learning_rate": 9.971521881123449e-06,
"loss": 0.7804,
"step": 922
},
{
"epoch": 0.03013877551020408,
"grad_norm": 3.011596441268921,
"learning_rate": 9.971489222730243e-06,
"loss": 0.7706,
"step": 923
},
{
"epoch": 0.03017142857142857,
"grad_norm": 2.9059317111968994,
"learning_rate": 9.971456564337035e-06,
"loss": 0.7594,
"step": 924
},
{
"epoch": 0.030204081632653063,
"grad_norm": 3.198932409286499,
"learning_rate": 9.971423905943828e-06,
"loss": 0.7876,
"step": 925
},
{
"epoch": 0.03023673469387755,
"grad_norm": 3.1033425331115723,
"learning_rate": 9.971391247550621e-06,
"loss": 0.7628,
"step": 926
},
{
"epoch": 0.03026938775510204,
"grad_norm": 3.210116147994995,
"learning_rate": 9.971358589157413e-06,
"loss": 0.748,
"step": 927
},
{
"epoch": 0.03030204081632653,
"grad_norm": 2.9255874156951904,
"learning_rate": 9.971325930764208e-06,
"loss": 0.7654,
"step": 928
},
{
"epoch": 0.03033469387755102,
"grad_norm": 2.949495553970337,
"learning_rate": 9.971293272371e-06,
"loss": 0.7875,
"step": 929
},
{
"epoch": 0.03036734693877551,
"grad_norm": 2.8590776920318604,
"learning_rate": 9.971260613977793e-06,
"loss": 0.7365,
"step": 930
},
{
"epoch": 0.0304,
"grad_norm": 2.9772932529449463,
"learning_rate": 9.971227955584586e-06,
"loss": 0.7683,
"step": 931
},
{
"epoch": 0.03043265306122449,
"grad_norm": 3.1256070137023926,
"learning_rate": 9.97119529719138e-06,
"loss": 0.8071,
"step": 932
},
{
"epoch": 0.03046530612244898,
"grad_norm": 3.2088918685913086,
"learning_rate": 9.971162638798172e-06,
"loss": 0.7676,
"step": 933
},
{
"epoch": 0.03049795918367347,
"grad_norm": 2.9030396938323975,
"learning_rate": 9.971129980404964e-06,
"loss": 0.8101,
"step": 934
},
{
"epoch": 0.030530612244897958,
"grad_norm": 2.968740940093994,
"learning_rate": 9.971097322011757e-06,
"loss": 0.7716,
"step": 935
},
{
"epoch": 0.03056326530612245,
"grad_norm": 2.748076915740967,
"learning_rate": 9.97106466361855e-06,
"loss": 0.7889,
"step": 936
},
{
"epoch": 0.030595918367346937,
"grad_norm": 2.941471815109253,
"learning_rate": 9.971032005225344e-06,
"loss": 0.7453,
"step": 937
},
{
"epoch": 0.03062857142857143,
"grad_norm": 2.9281535148620605,
"learning_rate": 9.970999346832137e-06,
"loss": 0.7891,
"step": 938
},
{
"epoch": 0.03066122448979592,
"grad_norm": 2.85978627204895,
"learning_rate": 9.97096668843893e-06,
"loss": 0.7772,
"step": 939
},
{
"epoch": 0.030693877551020408,
"grad_norm": 3.195918560028076,
"learning_rate": 9.970934030045722e-06,
"loss": 0.8068,
"step": 940
},
{
"epoch": 0.0307265306122449,
"grad_norm": 3.1167869567871094,
"learning_rate": 9.970901371652517e-06,
"loss": 0.7549,
"step": 941
},
{
"epoch": 0.030759183673469387,
"grad_norm": 2.9762868881225586,
"learning_rate": 9.970868713259308e-06,
"loss": 0.7923,
"step": 942
},
{
"epoch": 0.030791836734693878,
"grad_norm": 3.1077170372009277,
"learning_rate": 9.970836054866101e-06,
"loss": 0.7636,
"step": 943
},
{
"epoch": 0.030824489795918366,
"grad_norm": 3.062537431716919,
"learning_rate": 9.970803396472895e-06,
"loss": 0.7796,
"step": 944
},
{
"epoch": 0.030857142857142857,
"grad_norm": 3.2072839736938477,
"learning_rate": 9.970770738079686e-06,
"loss": 0.793,
"step": 945
},
{
"epoch": 0.03088979591836735,
"grad_norm": 3.26112699508667,
"learning_rate": 9.970738079686481e-06,
"loss": 0.8107,
"step": 946
},
{
"epoch": 0.030922448979591836,
"grad_norm": 2.9786770343780518,
"learning_rate": 9.970705421293273e-06,
"loss": 0.8025,
"step": 947
},
{
"epoch": 0.030955102040816328,
"grad_norm": 3.0765345096588135,
"learning_rate": 9.970672762900066e-06,
"loss": 0.8047,
"step": 948
},
{
"epoch": 0.030987755102040816,
"grad_norm": 2.958984375,
"learning_rate": 9.970640104506859e-06,
"loss": 0.762,
"step": 949
},
{
"epoch": 0.031020408163265307,
"grad_norm": 3.08040452003479,
"learning_rate": 9.970607446113652e-06,
"loss": 0.7928,
"step": 950
},
{
"epoch": 0.031020408163265307,
"eval_loss": 0.8111075758934021,
"eval_runtime": 74.5279,
"eval_samples_per_second": 1.342,
"eval_steps_per_second": 1.342,
"step": 950
},
{
"epoch": 0.031053061224489795,
"grad_norm": 2.965144395828247,
"learning_rate": 9.970574787720446e-06,
"loss": 0.8011,
"step": 951
},
{
"epoch": 0.031085714285714286,
"grad_norm": 2.9179527759552,
"learning_rate": 9.970542129327237e-06,
"loss": 0.7703,
"step": 952
},
{
"epoch": 0.031118367346938774,
"grad_norm": 2.9587607383728027,
"learning_rate": 9.97050947093403e-06,
"loss": 0.7536,
"step": 953
},
{
"epoch": 0.031151020408163265,
"grad_norm": 2.9393138885498047,
"learning_rate": 9.970476812540824e-06,
"loss": 0.7922,
"step": 954
},
{
"epoch": 0.031183673469387756,
"grad_norm": 3.013162612915039,
"learning_rate": 9.970444154147617e-06,
"loss": 0.7668,
"step": 955
},
{
"epoch": 0.031216326530612244,
"grad_norm": 2.8993349075317383,
"learning_rate": 9.97041149575441e-06,
"loss": 0.7559,
"step": 956
},
{
"epoch": 0.031248979591836736,
"grad_norm": 3.0861432552337646,
"learning_rate": 9.970378837361203e-06,
"loss": 0.7632,
"step": 957
},
{
"epoch": 0.03128163265306123,
"grad_norm": 2.9803247451782227,
"learning_rate": 9.970346178967995e-06,
"loss": 0.7933,
"step": 958
},
{
"epoch": 0.03131428571428571,
"grad_norm": 3.0155863761901855,
"learning_rate": 9.970313520574788e-06,
"loss": 0.8215,
"step": 959
},
{
"epoch": 0.0313469387755102,
"grad_norm": 3.0787782669067383,
"learning_rate": 9.970280862181581e-06,
"loss": 0.7936,
"step": 960
},
{
"epoch": 0.031379591836734694,
"grad_norm": 3.0619027614593506,
"learning_rate": 9.970248203788375e-06,
"loss": 0.7958,
"step": 961
},
{
"epoch": 0.031412244897959185,
"grad_norm": 2.7557358741760254,
"learning_rate": 9.970215545395168e-06,
"loss": 0.761,
"step": 962
},
{
"epoch": 0.031444897959183676,
"grad_norm": 2.8766896724700928,
"learning_rate": 9.97018288700196e-06,
"loss": 0.7563,
"step": 963
},
{
"epoch": 0.03147755102040816,
"grad_norm": 2.976086139678955,
"learning_rate": 9.970150228608754e-06,
"loss": 0.7839,
"step": 964
},
{
"epoch": 0.03151020408163265,
"grad_norm": 2.788193941116333,
"learning_rate": 9.970117570215546e-06,
"loss": 0.789,
"step": 965
},
{
"epoch": 0.03154285714285714,
"grad_norm": 3.0085878372192383,
"learning_rate": 9.970084911822339e-06,
"loss": 0.771,
"step": 966
},
{
"epoch": 0.031575510204081635,
"grad_norm": 3.1194570064544678,
"learning_rate": 9.970052253429132e-06,
"loss": 0.7889,
"step": 967
},
{
"epoch": 0.031608163265306126,
"grad_norm": 2.8725526332855225,
"learning_rate": 9.970019595035924e-06,
"loss": 0.7882,
"step": 968
},
{
"epoch": 0.03164081632653061,
"grad_norm": 2.982837677001953,
"learning_rate": 9.969986936642719e-06,
"loss": 0.7674,
"step": 969
},
{
"epoch": 0.0316734693877551,
"grad_norm": 2.9641151428222656,
"learning_rate": 9.96995427824951e-06,
"loss": 0.7871,
"step": 970
},
{
"epoch": 0.03170612244897959,
"grad_norm": 2.973365068435669,
"learning_rate": 9.969921619856304e-06,
"loss": 0.7939,
"step": 971
},
{
"epoch": 0.031738775510204084,
"grad_norm": 3.0161428451538086,
"learning_rate": 9.969888961463097e-06,
"loss": 0.7865,
"step": 972
},
{
"epoch": 0.03177142857142857,
"grad_norm": 3.369993209838867,
"learning_rate": 9.96985630306989e-06,
"loss": 0.8231,
"step": 973
},
{
"epoch": 0.03180408163265306,
"grad_norm": 3.0183799266815186,
"learning_rate": 9.969823644676683e-06,
"loss": 0.7853,
"step": 974
},
{
"epoch": 0.03183673469387755,
"grad_norm": 3.244966983795166,
"learning_rate": 9.969790986283475e-06,
"loss": 0.7607,
"step": 975
},
{
"epoch": 0.03186938775510204,
"grad_norm": 3.0273585319519043,
"learning_rate": 9.969758327890268e-06,
"loss": 0.77,
"step": 976
},
{
"epoch": 0.031902040816326534,
"grad_norm": 3.168429136276245,
"learning_rate": 9.969725669497061e-06,
"loss": 0.7537,
"step": 977
},
{
"epoch": 0.03193469387755102,
"grad_norm": 3.006373882293701,
"learning_rate": 9.969693011103854e-06,
"loss": 0.7927,
"step": 978
},
{
"epoch": 0.03196734693877551,
"grad_norm": 3.1100571155548096,
"learning_rate": 9.969660352710648e-06,
"loss": 0.7846,
"step": 979
},
{
"epoch": 0.032,
"grad_norm": 2.941803216934204,
"learning_rate": 9.969627694317441e-06,
"loss": 0.7821,
"step": 980
},
{
"epoch": 0.03203265306122449,
"grad_norm": 3.077153444290161,
"learning_rate": 9.969595035924232e-06,
"loss": 0.8004,
"step": 981
},
{
"epoch": 0.032065306122448976,
"grad_norm": 2.9901161193847656,
"learning_rate": 9.969562377531026e-06,
"loss": 0.7848,
"step": 982
},
{
"epoch": 0.03209795918367347,
"grad_norm": 3.0473811626434326,
"learning_rate": 9.969529719137819e-06,
"loss": 0.7823,
"step": 983
},
{
"epoch": 0.03213061224489796,
"grad_norm": 2.894256353378296,
"learning_rate": 9.969497060744612e-06,
"loss": 0.7757,
"step": 984
},
{
"epoch": 0.03216326530612245,
"grad_norm": 3.0119597911834717,
"learning_rate": 9.969464402351405e-06,
"loss": 0.7799,
"step": 985
},
{
"epoch": 0.03219591836734694,
"grad_norm": 3.020019292831421,
"learning_rate": 9.969431743958197e-06,
"loss": 0.7817,
"step": 986
},
{
"epoch": 0.032228571428571426,
"grad_norm": 3.0470404624938965,
"learning_rate": 9.969399085564992e-06,
"loss": 0.7873,
"step": 987
},
{
"epoch": 0.03226122448979592,
"grad_norm": 2.9759464263916016,
"learning_rate": 9.969366427171783e-06,
"loss": 0.7857,
"step": 988
},
{
"epoch": 0.03229387755102041,
"grad_norm": 3.1986935138702393,
"learning_rate": 9.969333768778577e-06,
"loss": 0.786,
"step": 989
},
{
"epoch": 0.0323265306122449,
"grad_norm": 3.0360336303710938,
"learning_rate": 9.96930111038537e-06,
"loss": 0.8174,
"step": 990
},
{
"epoch": 0.03235918367346939,
"grad_norm": 3.093979597091675,
"learning_rate": 9.969268451992163e-06,
"loss": 0.7619,
"step": 991
},
{
"epoch": 0.032391836734693875,
"grad_norm": 2.935920238494873,
"learning_rate": 9.969235793598956e-06,
"loss": 0.7661,
"step": 992
},
{
"epoch": 0.03242448979591837,
"grad_norm": 3.1429708003997803,
"learning_rate": 9.969203135205748e-06,
"loss": 0.7981,
"step": 993
},
{
"epoch": 0.03245714285714286,
"grad_norm": 2.8831772804260254,
"learning_rate": 9.969170476812541e-06,
"loss": 0.7969,
"step": 994
},
{
"epoch": 0.03248979591836735,
"grad_norm": 2.8683125972747803,
"learning_rate": 9.969137818419334e-06,
"loss": 0.7688,
"step": 995
},
{
"epoch": 0.032522448979591834,
"grad_norm": 3.026094436645508,
"learning_rate": 9.969105160026128e-06,
"loss": 0.782,
"step": 996
},
{
"epoch": 0.032555102040816325,
"grad_norm": 3.065110445022583,
"learning_rate": 9.969072501632921e-06,
"loss": 0.782,
"step": 997
},
{
"epoch": 0.032587755102040816,
"grad_norm": 2.9125099182128906,
"learning_rate": 9.969039843239714e-06,
"loss": 0.8056,
"step": 998
},
{
"epoch": 0.03262040816326531,
"grad_norm": 2.978609800338745,
"learning_rate": 9.969007184846506e-06,
"loss": 0.7839,
"step": 999
},
{
"epoch": 0.0326530612244898,
"grad_norm": 3.037384510040283,
"learning_rate": 9.968974526453299e-06,
"loss": 0.7877,
"step": 1000
},
{
"epoch": 0.0326530612244898,
"eval_loss": 0.8111925721168518,
"eval_runtime": 72.9629,
"eval_samples_per_second": 1.371,
"eval_steps_per_second": 1.371,
"step": 1000
},
{
"epoch": 0.03268571428571428,
"grad_norm": 2.9627959728240967,
"learning_rate": 9.968941868060092e-06,
"loss": 0.7855,
"step": 1001
},
{
"epoch": 0.032718367346938775,
"grad_norm": 3.2459042072296143,
"learning_rate": 9.968909209666885e-06,
"loss": 0.7897,
"step": 1002
},
{
"epoch": 0.032751020408163266,
"grad_norm": 3.0746281147003174,
"learning_rate": 9.968876551273679e-06,
"loss": 0.7415,
"step": 1003
},
{
"epoch": 0.03278367346938776,
"grad_norm": 3.005556583404541,
"learning_rate": 9.96884389288047e-06,
"loss": 0.8075,
"step": 1004
},
{
"epoch": 0.03281632653061224,
"grad_norm": 2.9875948429107666,
"learning_rate": 9.968811234487265e-06,
"loss": 0.7982,
"step": 1005
},
{
"epoch": 0.03284897959183673,
"grad_norm": 2.8231279850006104,
"learning_rate": 9.968778576094057e-06,
"loss": 0.758,
"step": 1006
},
{
"epoch": 0.032881632653061224,
"grad_norm": 2.8995821475982666,
"learning_rate": 9.96874591770085e-06,
"loss": 0.7653,
"step": 1007
},
{
"epoch": 0.032914285714285715,
"grad_norm": 3.0024149417877197,
"learning_rate": 9.968713259307643e-06,
"loss": 0.7777,
"step": 1008
},
{
"epoch": 0.03294693877551021,
"grad_norm": 3.019351005554199,
"learning_rate": 9.968680600914435e-06,
"loss": 0.7921,
"step": 1009
},
{
"epoch": 0.03297959183673469,
"grad_norm": 2.916240930557251,
"learning_rate": 9.96864794252123e-06,
"loss": 0.75,
"step": 1010
},
{
"epoch": 0.03301224489795918,
"grad_norm": 3.005613088607788,
"learning_rate": 9.968615284128021e-06,
"loss": 0.7943,
"step": 1011
},
{
"epoch": 0.033044897959183674,
"grad_norm": 2.9690959453582764,
"learning_rate": 9.968582625734814e-06,
"loss": 0.7855,
"step": 1012
},
{
"epoch": 0.033077551020408165,
"grad_norm": 3.0370657444000244,
"learning_rate": 9.968549967341608e-06,
"loss": 0.7567,
"step": 1013
},
{
"epoch": 0.033110204081632656,
"grad_norm": 3.010715961456299,
"learning_rate": 9.9685173089484e-06,
"loss": 0.7545,
"step": 1014
},
{
"epoch": 0.03314285714285714,
"grad_norm": 3.1019158363342285,
"learning_rate": 9.968484650555194e-06,
"loss": 0.7881,
"step": 1015
},
{
"epoch": 0.03317551020408163,
"grad_norm": 2.9990556240081787,
"learning_rate": 9.968451992161986e-06,
"loss": 0.7892,
"step": 1016
},
{
"epoch": 0.03320816326530612,
"grad_norm": 2.9631762504577637,
"learning_rate": 9.968419333768779e-06,
"loss": 0.7459,
"step": 1017
},
{
"epoch": 0.033240816326530614,
"grad_norm": 3.1492321491241455,
"learning_rate": 9.968386675375572e-06,
"loss": 0.796,
"step": 1018
},
{
"epoch": 0.0332734693877551,
"grad_norm": 3.0348122119903564,
"learning_rate": 9.968354016982365e-06,
"loss": 0.795,
"step": 1019
},
{
"epoch": 0.03330612244897959,
"grad_norm": 3.2612578868865967,
"learning_rate": 9.968321358589159e-06,
"loss": 0.7781,
"step": 1020
},
{
"epoch": 0.03333877551020408,
"grad_norm": 3.0605742931365967,
"learning_rate": 9.968288700195952e-06,
"loss": 0.7329,
"step": 1021
},
{
"epoch": 0.03337142857142857,
"grad_norm": 3.0548489093780518,
"learning_rate": 9.968256041802743e-06,
"loss": 0.7601,
"step": 1022
},
{
"epoch": 0.033404081632653064,
"grad_norm": 2.946451187133789,
"learning_rate": 9.968223383409537e-06,
"loss": 0.7756,
"step": 1023
},
{
"epoch": 0.03343673469387755,
"grad_norm": 3.0444273948669434,
"learning_rate": 9.96819072501633e-06,
"loss": 0.762,
"step": 1024
},
{
"epoch": 0.03346938775510204,
"grad_norm": 3.207759380340576,
"learning_rate": 9.968158066623123e-06,
"loss": 0.7755,
"step": 1025
},
{
"epoch": 0.03350204081632653,
"grad_norm": 3.0307605266571045,
"learning_rate": 9.968125408229916e-06,
"loss": 0.7604,
"step": 1026
},
{
"epoch": 0.03353469387755102,
"grad_norm": 2.975473165512085,
"learning_rate": 9.968092749836708e-06,
"loss": 0.7929,
"step": 1027
},
{
"epoch": 0.033567346938775514,
"grad_norm": 3.00290846824646,
"learning_rate": 9.968060091443503e-06,
"loss": 0.8294,
"step": 1028
},
{
"epoch": 0.0336,
"grad_norm": 3.235747814178467,
"learning_rate": 9.968027433050294e-06,
"loss": 0.7807,
"step": 1029
},
{
"epoch": 0.03363265306122449,
"grad_norm": 3.031163215637207,
"learning_rate": 9.967994774657088e-06,
"loss": 0.7549,
"step": 1030
},
{
"epoch": 0.03366530612244898,
"grad_norm": 2.9528584480285645,
"learning_rate": 9.96796211626388e-06,
"loss": 0.7542,
"step": 1031
},
{
"epoch": 0.03369795918367347,
"grad_norm": 3.0693178176879883,
"learning_rate": 9.967929457870672e-06,
"loss": 0.7773,
"step": 1032
},
{
"epoch": 0.033730612244897956,
"grad_norm": 3.159449338912964,
"learning_rate": 9.967896799477467e-06,
"loss": 0.7998,
"step": 1033
},
{
"epoch": 0.03376326530612245,
"grad_norm": 3.0627989768981934,
"learning_rate": 9.967864141084259e-06,
"loss": 0.7624,
"step": 1034
},
{
"epoch": 0.03379591836734694,
"grad_norm": 3.088207721710205,
"learning_rate": 9.967831482691052e-06,
"loss": 0.7757,
"step": 1035
},
{
"epoch": 0.03382857142857143,
"grad_norm": 3.1921637058258057,
"learning_rate": 9.967798824297845e-06,
"loss": 0.7923,
"step": 1036
},
{
"epoch": 0.03386122448979592,
"grad_norm": 3.039994955062866,
"learning_rate": 9.967766165904638e-06,
"loss": 0.7628,
"step": 1037
},
{
"epoch": 0.033893877551020406,
"grad_norm": 3.221714973449707,
"learning_rate": 9.967733507511432e-06,
"loss": 0.7684,
"step": 1038
},
{
"epoch": 0.0339265306122449,
"grad_norm": 2.9829065799713135,
"learning_rate": 9.967700849118223e-06,
"loss": 0.7956,
"step": 1039
},
{
"epoch": 0.03395918367346939,
"grad_norm": 2.9304864406585693,
"learning_rate": 9.967668190725016e-06,
"loss": 0.7865,
"step": 1040
},
{
"epoch": 0.03399183673469388,
"grad_norm": 3.0041542053222656,
"learning_rate": 9.96763553233181e-06,
"loss": 0.8113,
"step": 1041
},
{
"epoch": 0.034024489795918364,
"grad_norm": 3.1308107376098633,
"learning_rate": 9.967602873938603e-06,
"loss": 0.7622,
"step": 1042
},
{
"epoch": 0.034057142857142855,
"grad_norm": 3.0761616230010986,
"learning_rate": 9.967570215545396e-06,
"loss": 0.7822,
"step": 1043
},
{
"epoch": 0.03408979591836735,
"grad_norm": 3.073542594909668,
"learning_rate": 9.96753755715219e-06,
"loss": 0.782,
"step": 1044
},
{
"epoch": 0.03412244897959184,
"grad_norm": 2.954620361328125,
"learning_rate": 9.967504898758981e-06,
"loss": 0.7686,
"step": 1045
},
{
"epoch": 0.03415510204081633,
"grad_norm": 3.1652603149414062,
"learning_rate": 9.967472240365776e-06,
"loss": 0.7743,
"step": 1046
},
{
"epoch": 0.034187755102040814,
"grad_norm": 2.8833162784576416,
"learning_rate": 9.967439581972567e-06,
"loss": 0.7741,
"step": 1047
},
{
"epoch": 0.034220408163265305,
"grad_norm": 2.910767078399658,
"learning_rate": 9.96740692357936e-06,
"loss": 0.7768,
"step": 1048
},
{
"epoch": 0.034253061224489796,
"grad_norm": 3.0124878883361816,
"learning_rate": 9.967374265186154e-06,
"loss": 0.7656,
"step": 1049
},
{
"epoch": 0.03428571428571429,
"grad_norm": 3.200578451156616,
"learning_rate": 9.967341606792945e-06,
"loss": 0.7979,
"step": 1050
},
{
"epoch": 0.03428571428571429,
"eval_loss": 0.809950053691864,
"eval_runtime": 78.3268,
"eval_samples_per_second": 1.277,
"eval_steps_per_second": 1.277,
"step": 1050
},
{
"epoch": 0.03431836734693878,
"grad_norm": 3.1058106422424316,
"learning_rate": 9.96730894839974e-06,
"loss": 0.7929,
"step": 1051
},
{
"epoch": 0.03435102040816326,
"grad_norm": 2.981287956237793,
"learning_rate": 9.967276290006532e-06,
"loss": 0.7922,
"step": 1052
},
{
"epoch": 0.034383673469387754,
"grad_norm": 3.0732994079589844,
"learning_rate": 9.967243631613325e-06,
"loss": 0.7598,
"step": 1053
},
{
"epoch": 0.034416326530612246,
"grad_norm": 2.965428352355957,
"learning_rate": 9.967210973220118e-06,
"loss": 0.7796,
"step": 1054
},
{
"epoch": 0.03444897959183674,
"grad_norm": 3.047433853149414,
"learning_rate": 9.967178314826912e-06,
"loss": 0.7806,
"step": 1055
},
{
"epoch": 0.03448163265306122,
"grad_norm": 2.9787747859954834,
"learning_rate": 9.967145656433705e-06,
"loss": 0.7548,
"step": 1056
},
{
"epoch": 0.03451428571428571,
"grad_norm": 3.3675661087036133,
"learning_rate": 9.967112998040496e-06,
"loss": 0.7764,
"step": 1057
},
{
"epoch": 0.034546938775510204,
"grad_norm": 2.90116024017334,
"learning_rate": 9.96708033964729e-06,
"loss": 0.7819,
"step": 1058
},
{
"epoch": 0.034579591836734695,
"grad_norm": 2.965744972229004,
"learning_rate": 9.967047681254083e-06,
"loss": 0.7603,
"step": 1059
},
{
"epoch": 0.034612244897959187,
"grad_norm": 3.116170644760132,
"learning_rate": 9.967015022860876e-06,
"loss": 0.7885,
"step": 1060
},
{
"epoch": 0.03464489795918367,
"grad_norm": 3.1526293754577637,
"learning_rate": 9.96698236446767e-06,
"loss": 0.7752,
"step": 1061
},
{
"epoch": 0.03467755102040816,
"grad_norm": 3.0937249660491943,
"learning_rate": 9.966949706074463e-06,
"loss": 0.7845,
"step": 1062
},
{
"epoch": 0.034710204081632653,
"grad_norm": 3.069850444793701,
"learning_rate": 9.966917047681254e-06,
"loss": 0.7645,
"step": 1063
},
{
"epoch": 0.034742857142857145,
"grad_norm": 2.979881763458252,
"learning_rate": 9.966884389288047e-06,
"loss": 0.7687,
"step": 1064
},
{
"epoch": 0.034775510204081636,
"grad_norm": 3.1711268424987793,
"learning_rate": 9.96685173089484e-06,
"loss": 0.8136,
"step": 1065
},
{
"epoch": 0.03480816326530612,
"grad_norm": 3.197852373123169,
"learning_rate": 9.966819072501634e-06,
"loss": 0.8152,
"step": 1066
},
{
"epoch": 0.03484081632653061,
"grad_norm": 3.098540782928467,
"learning_rate": 9.966786414108427e-06,
"loss": 0.7645,
"step": 1067
},
{
"epoch": 0.0348734693877551,
"grad_norm": 3.1754651069641113,
"learning_rate": 9.966753755715219e-06,
"loss": 0.7663,
"step": 1068
},
{
"epoch": 0.034906122448979594,
"grad_norm": 3.0712950229644775,
"learning_rate": 9.966721097322014e-06,
"loss": 0.7851,
"step": 1069
},
{
"epoch": 0.03493877551020408,
"grad_norm": 3.285234212875366,
"learning_rate": 9.966688438928805e-06,
"loss": 0.774,
"step": 1070
},
{
"epoch": 0.03497142857142857,
"grad_norm": 3.2550666332244873,
"learning_rate": 9.966655780535598e-06,
"loss": 0.751,
"step": 1071
},
{
"epoch": 0.03500408163265306,
"grad_norm": 3.238384246826172,
"learning_rate": 9.966623122142392e-06,
"loss": 0.7666,
"step": 1072
},
{
"epoch": 0.03503673469387755,
"grad_norm": 3.0862374305725098,
"learning_rate": 9.966590463749183e-06,
"loss": 0.7981,
"step": 1073
},
{
"epoch": 0.035069387755102044,
"grad_norm": 3.172941207885742,
"learning_rate": 9.966557805355978e-06,
"loss": 0.7549,
"step": 1074
},
{
"epoch": 0.03510204081632653,
"grad_norm": 3.060302495956421,
"learning_rate": 9.96652514696277e-06,
"loss": 0.7835,
"step": 1075
},
{
"epoch": 0.03513469387755102,
"grad_norm": 3.122617244720459,
"learning_rate": 9.966492488569563e-06,
"loss": 0.7952,
"step": 1076
},
{
"epoch": 0.03516734693877551,
"grad_norm": 2.904637336730957,
"learning_rate": 9.966459830176356e-06,
"loss": 0.7595,
"step": 1077
},
{
"epoch": 0.0352,
"grad_norm": 3.1672565937042236,
"learning_rate": 9.96642717178315e-06,
"loss": 0.7926,
"step": 1078
},
{
"epoch": 0.035232653061224486,
"grad_norm": 3.0276424884796143,
"learning_rate": 9.966394513389943e-06,
"loss": 0.7801,
"step": 1079
},
{
"epoch": 0.03526530612244898,
"grad_norm": 3.0199339389801025,
"learning_rate": 9.966361854996734e-06,
"loss": 0.7977,
"step": 1080
},
{
"epoch": 0.03529795918367347,
"grad_norm": 3.1580443382263184,
"learning_rate": 9.966329196603527e-06,
"loss": 0.7638,
"step": 1081
},
{
"epoch": 0.03533061224489796,
"grad_norm": 2.9399259090423584,
"learning_rate": 9.96629653821032e-06,
"loss": 0.7594,
"step": 1082
},
{
"epoch": 0.03536326530612245,
"grad_norm": 3.1555962562561035,
"learning_rate": 9.966263879817114e-06,
"loss": 0.7688,
"step": 1083
},
{
"epoch": 0.035395918367346936,
"grad_norm": 3.03648042678833,
"learning_rate": 9.966231221423907e-06,
"loss": 0.8,
"step": 1084
},
{
"epoch": 0.03542857142857143,
"grad_norm": 3.142136573791504,
"learning_rate": 9.9661985630307e-06,
"loss": 0.7502,
"step": 1085
},
{
"epoch": 0.03546122448979592,
"grad_norm": 3.103031635284424,
"learning_rate": 9.966165904637492e-06,
"loss": 0.7979,
"step": 1086
},
{
"epoch": 0.03549387755102041,
"grad_norm": 3.0145888328552246,
"learning_rate": 9.966133246244287e-06,
"loss": 0.7669,
"step": 1087
},
{
"epoch": 0.0355265306122449,
"grad_norm": 3.1782283782958984,
"learning_rate": 9.966100587851078e-06,
"loss": 0.7677,
"step": 1088
},
{
"epoch": 0.035559183673469386,
"grad_norm": 3.102893829345703,
"learning_rate": 9.966067929457871e-06,
"loss": 0.7651,
"step": 1089
},
{
"epoch": 0.03559183673469388,
"grad_norm": 3.0069074630737305,
"learning_rate": 9.966035271064665e-06,
"loss": 0.7729,
"step": 1090
},
{
"epoch": 0.03562448979591837,
"grad_norm": 3.022388458251953,
"learning_rate": 9.966002612671456e-06,
"loss": 0.7537,
"step": 1091
},
{
"epoch": 0.03565714285714286,
"grad_norm": 3.0949289798736572,
"learning_rate": 9.965969954278251e-06,
"loss": 0.77,
"step": 1092
},
{
"epoch": 0.035689795918367344,
"grad_norm": 3.48496413230896,
"learning_rate": 9.965937295885043e-06,
"loss": 0.7671,
"step": 1093
},
{
"epoch": 0.035722448979591835,
"grad_norm": 3.1143581867218018,
"learning_rate": 9.965904637491836e-06,
"loss": 0.7451,
"step": 1094
},
{
"epoch": 0.035755102040816326,
"grad_norm": 2.93046498298645,
"learning_rate": 9.96587197909863e-06,
"loss": 0.7878,
"step": 1095
},
{
"epoch": 0.03578775510204082,
"grad_norm": 3.0415232181549072,
"learning_rate": 9.96583932070542e-06,
"loss": 0.7465,
"step": 1096
},
{
"epoch": 0.03582040816326531,
"grad_norm": 3.1707770824432373,
"learning_rate": 9.965806662312216e-06,
"loss": 0.7548,
"step": 1097
},
{
"epoch": 0.03585306122448979,
"grad_norm": 3.0290682315826416,
"learning_rate": 9.965774003919007e-06,
"loss": 0.7703,
"step": 1098
},
{
"epoch": 0.035885714285714285,
"grad_norm": 2.925924777984619,
"learning_rate": 9.9657413455258e-06,
"loss": 0.7653,
"step": 1099
},
{
"epoch": 0.035918367346938776,
"grad_norm": 3.2360856533050537,
"learning_rate": 9.965708687132594e-06,
"loss": 0.7705,
"step": 1100
},
{
"epoch": 0.035918367346938776,
"eval_loss": 0.8095739483833313,
"eval_runtime": 95.0564,
"eval_samples_per_second": 1.052,
"eval_steps_per_second": 1.052,
"step": 1100
},
{
"epoch": 0.03595102040816327,
"grad_norm": 3.276266574859619,
"learning_rate": 9.965676028739387e-06,
"loss": 0.7926,
"step": 1101
},
{
"epoch": 0.03598367346938776,
"grad_norm": 3.069849967956543,
"learning_rate": 9.96564337034618e-06,
"loss": 0.7966,
"step": 1102
},
{
"epoch": 0.03601632653061224,
"grad_norm": 3.484065532684326,
"learning_rate": 9.965610711952973e-06,
"loss": 0.7744,
"step": 1103
},
{
"epoch": 0.036048979591836734,
"grad_norm": 3.468071699142456,
"learning_rate": 9.965578053559765e-06,
"loss": 0.7575,
"step": 1104
},
{
"epoch": 0.036081632653061226,
"grad_norm": 3.8356101512908936,
"learning_rate": 9.965545395166558e-06,
"loss": 0.7817,
"step": 1105
},
{
"epoch": 0.03611428571428572,
"grad_norm": 3.2512879371643066,
"learning_rate": 9.965512736773351e-06,
"loss": 0.7356,
"step": 1106
},
{
"epoch": 0.0361469387755102,
"grad_norm": 3.004352569580078,
"learning_rate": 9.965480078380145e-06,
"loss": 0.7286,
"step": 1107
},
{
"epoch": 0.03617959183673469,
"grad_norm": 3.194286823272705,
"learning_rate": 9.965447419986938e-06,
"loss": 0.7474,
"step": 1108
},
{
"epoch": 0.036212244897959184,
"grad_norm": 3.509319543838501,
"learning_rate": 9.96541476159373e-06,
"loss": 0.7888,
"step": 1109
},
{
"epoch": 0.036244897959183675,
"grad_norm": 3.528698205947876,
"learning_rate": 9.965382103200524e-06,
"loss": 0.794,
"step": 1110
},
{
"epoch": 0.036277551020408166,
"grad_norm": 3.2880067825317383,
"learning_rate": 9.965349444807316e-06,
"loss": 0.7408,
"step": 1111
},
{
"epoch": 0.03631020408163265,
"grad_norm": 3.031257152557373,
"learning_rate": 9.965316786414109e-06,
"loss": 0.7847,
"step": 1112
},
{
"epoch": 0.03634285714285714,
"grad_norm": 3.2594382762908936,
"learning_rate": 9.965284128020902e-06,
"loss": 0.7911,
"step": 1113
},
{
"epoch": 0.03637551020408163,
"grad_norm": 3.2341842651367188,
"learning_rate": 9.965251469627694e-06,
"loss": 0.7412,
"step": 1114
},
{
"epoch": 0.036408163265306125,
"grad_norm": 3.251246213912964,
"learning_rate": 9.965218811234489e-06,
"loss": 0.7549,
"step": 1115
},
{
"epoch": 0.03644081632653061,
"grad_norm": 3.0365469455718994,
"learning_rate": 9.96518615284128e-06,
"loss": 0.78,
"step": 1116
},
{
"epoch": 0.0364734693877551,
"grad_norm": 3.059936285018921,
"learning_rate": 9.965153494448074e-06,
"loss": 0.805,
"step": 1117
},
{
"epoch": 0.03650612244897959,
"grad_norm": 3.0952272415161133,
"learning_rate": 9.965120836054867e-06,
"loss": 0.7563,
"step": 1118
},
{
"epoch": 0.03653877551020408,
"grad_norm": 3.136528491973877,
"learning_rate": 9.96508817766166e-06,
"loss": 0.7767,
"step": 1119
},
{
"epoch": 0.036571428571428574,
"grad_norm": 2.979304313659668,
"learning_rate": 9.965055519268453e-06,
"loss": 0.8002,
"step": 1120
},
{
"epoch": 0.03660408163265306,
"grad_norm": 2.8801610469818115,
"learning_rate": 9.965022860875245e-06,
"loss": 0.7793,
"step": 1121
},
{
"epoch": 0.03663673469387755,
"grad_norm": 2.9416167736053467,
"learning_rate": 9.964990202482038e-06,
"loss": 0.7703,
"step": 1122
},
{
"epoch": 0.03666938775510204,
"grad_norm": 3.1201093196868896,
"learning_rate": 9.964957544088831e-06,
"loss": 0.7699,
"step": 1123
},
{
"epoch": 0.03670204081632653,
"grad_norm": 3.1061949729919434,
"learning_rate": 9.964924885695625e-06,
"loss": 0.7974,
"step": 1124
},
{
"epoch": 0.036734693877551024,
"grad_norm": 3.0530447959899902,
"learning_rate": 9.964892227302418e-06,
"loss": 0.7576,
"step": 1125
},
{
"epoch": 0.03676734693877551,
"grad_norm": 3.002209186553955,
"learning_rate": 9.964859568909211e-06,
"loss": 0.7699,
"step": 1126
},
{
"epoch": 0.0368,
"grad_norm": 3.1953022480010986,
"learning_rate": 9.964826910516003e-06,
"loss": 0.7153,
"step": 1127
},
{
"epoch": 0.03683265306122449,
"grad_norm": 3.091996192932129,
"learning_rate": 9.964794252122796e-06,
"loss": 0.7368,
"step": 1128
},
{
"epoch": 0.03686530612244898,
"grad_norm": 3.173081398010254,
"learning_rate": 9.964761593729589e-06,
"loss": 0.7839,
"step": 1129
},
{
"epoch": 0.036897959183673466,
"grad_norm": 3.0133304595947266,
"learning_rate": 9.964728935336382e-06,
"loss": 0.7827,
"step": 1130
},
{
"epoch": 0.03693061224489796,
"grad_norm": 3.1596643924713135,
"learning_rate": 9.964696276943176e-06,
"loss": 0.7649,
"step": 1131
},
{
"epoch": 0.03696326530612245,
"grad_norm": 3.2854256629943848,
"learning_rate": 9.964663618549967e-06,
"loss": 0.751,
"step": 1132
},
{
"epoch": 0.03699591836734694,
"grad_norm": 2.8992016315460205,
"learning_rate": 9.964630960156762e-06,
"loss": 0.7535,
"step": 1133
},
{
"epoch": 0.03702857142857143,
"grad_norm": 2.9541709423065186,
"learning_rate": 9.964598301763554e-06,
"loss": 0.7433,
"step": 1134
},
{
"epoch": 0.037061224489795916,
"grad_norm": 2.9385082721710205,
"learning_rate": 9.964565643370347e-06,
"loss": 0.8005,
"step": 1135
},
{
"epoch": 0.03709387755102041,
"grad_norm": 3.042072057723999,
"learning_rate": 9.96453298497714e-06,
"loss": 0.7819,
"step": 1136
},
{
"epoch": 0.0371265306122449,
"grad_norm": 2.9819841384887695,
"learning_rate": 9.964500326583932e-06,
"loss": 0.7911,
"step": 1137
},
{
"epoch": 0.03715918367346939,
"grad_norm": 3.242607831954956,
"learning_rate": 9.964467668190726e-06,
"loss": 0.7926,
"step": 1138
},
{
"epoch": 0.037191836734693874,
"grad_norm": 2.9996449947357178,
"learning_rate": 9.964435009797518e-06,
"loss": 0.7904,
"step": 1139
},
{
"epoch": 0.037224489795918365,
"grad_norm": 3.1378376483917236,
"learning_rate": 9.964402351404311e-06,
"loss": 0.8,
"step": 1140
},
{
"epoch": 0.03725714285714286,
"grad_norm": 2.9573757648468018,
"learning_rate": 9.964369693011105e-06,
"loss": 0.7589,
"step": 1141
},
{
"epoch": 0.03728979591836735,
"grad_norm": 3.122229814529419,
"learning_rate": 9.964337034617898e-06,
"loss": 0.7852,
"step": 1142
},
{
"epoch": 0.03732244897959184,
"grad_norm": 3.3379294872283936,
"learning_rate": 9.964304376224691e-06,
"loss": 0.7868,
"step": 1143
},
{
"epoch": 0.037355102040816324,
"grad_norm": 3.2145421504974365,
"learning_rate": 9.964271717831484e-06,
"loss": 0.7516,
"step": 1144
},
{
"epoch": 0.037387755102040815,
"grad_norm": 3.068854808807373,
"learning_rate": 9.964239059438276e-06,
"loss": 0.7539,
"step": 1145
},
{
"epoch": 0.037420408163265306,
"grad_norm": 3.1995043754577637,
"learning_rate": 9.964206401045069e-06,
"loss": 0.7927,
"step": 1146
},
{
"epoch": 0.0374530612244898,
"grad_norm": 3.0854032039642334,
"learning_rate": 9.964173742651862e-06,
"loss": 0.7608,
"step": 1147
},
{
"epoch": 0.03748571428571429,
"grad_norm": 3.3977510929107666,
"learning_rate": 9.964141084258655e-06,
"loss": 0.779,
"step": 1148
},
{
"epoch": 0.03751836734693877,
"grad_norm": 3.0493321418762207,
"learning_rate": 9.964108425865449e-06,
"loss": 0.7329,
"step": 1149
},
{
"epoch": 0.037551020408163265,
"grad_norm": 2.8519468307495117,
"learning_rate": 9.96407576747224e-06,
"loss": 0.7954,
"step": 1150
},
{
"epoch": 0.037551020408163265,
"eval_loss": 0.8091681003570557,
"eval_runtime": 85.0018,
"eval_samples_per_second": 1.176,
"eval_steps_per_second": 1.176,
"step": 1150
},
{
"epoch": 0.037583673469387756,
"grad_norm": 2.9106438159942627,
"learning_rate": 9.964043109079035e-06,
"loss": 0.7796,
"step": 1151
},
{
"epoch": 0.03761632653061225,
"grad_norm": 3.1851396560668945,
"learning_rate": 9.964010450685827e-06,
"loss": 0.7718,
"step": 1152
},
{
"epoch": 0.03764897959183673,
"grad_norm": 3.101126194000244,
"learning_rate": 9.96397779229262e-06,
"loss": 0.8017,
"step": 1153
},
{
"epoch": 0.03768163265306122,
"grad_norm": 3.1867284774780273,
"learning_rate": 9.963945133899413e-06,
"loss": 0.7973,
"step": 1154
},
{
"epoch": 0.037714285714285714,
"grad_norm": 2.98366379737854,
"learning_rate": 9.963912475506205e-06,
"loss": 0.8105,
"step": 1155
},
{
"epoch": 0.037746938775510205,
"grad_norm": 2.998530387878418,
"learning_rate": 9.963879817113e-06,
"loss": 0.7738,
"step": 1156
},
{
"epoch": 0.0377795918367347,
"grad_norm": 3.0798680782318115,
"learning_rate": 9.963847158719791e-06,
"loss": 0.7693,
"step": 1157
},
{
"epoch": 0.03781224489795918,
"grad_norm": 3.2727184295654297,
"learning_rate": 9.963814500326584e-06,
"loss": 0.7271,
"step": 1158
},
{
"epoch": 0.03784489795918367,
"grad_norm": 3.05658221244812,
"learning_rate": 9.963781841933378e-06,
"loss": 0.7759,
"step": 1159
},
{
"epoch": 0.037877551020408164,
"grad_norm": 3.0974197387695312,
"learning_rate": 9.963749183540171e-06,
"loss": 0.7844,
"step": 1160
},
{
"epoch": 0.037910204081632655,
"grad_norm": 3.1191930770874023,
"learning_rate": 9.963716525146964e-06,
"loss": 0.8066,
"step": 1161
},
{
"epoch": 0.037942857142857146,
"grad_norm": 3.0258705615997314,
"learning_rate": 9.963683866753756e-06,
"loss": 0.7373,
"step": 1162
},
{
"epoch": 0.03797551020408163,
"grad_norm": 3.1126925945281982,
"learning_rate": 9.963651208360549e-06,
"loss": 0.7577,
"step": 1163
},
{
"epoch": 0.03800816326530612,
"grad_norm": 3.1630795001983643,
"learning_rate": 9.963618549967342e-06,
"loss": 0.7333,
"step": 1164
},
{
"epoch": 0.03804081632653061,
"grad_norm": 3.100018262863159,
"learning_rate": 9.963585891574135e-06,
"loss": 0.7697,
"step": 1165
},
{
"epoch": 0.038073469387755104,
"grad_norm": 3.369335889816284,
"learning_rate": 9.963553233180929e-06,
"loss": 0.7788,
"step": 1166
},
{
"epoch": 0.03810612244897959,
"grad_norm": 3.357944965362549,
"learning_rate": 9.963520574787722e-06,
"loss": 0.7744,
"step": 1167
},
{
"epoch": 0.03813877551020408,
"grad_norm": 3.253232717514038,
"learning_rate": 9.963487916394513e-06,
"loss": 0.752,
"step": 1168
},
{
"epoch": 0.03817142857142857,
"grad_norm": 2.9491517543792725,
"learning_rate": 9.963455258001307e-06,
"loss": 0.7657,
"step": 1169
},
{
"epoch": 0.03820408163265306,
"grad_norm": 3.1153948307037354,
"learning_rate": 9.9634225996081e-06,
"loss": 0.7545,
"step": 1170
},
{
"epoch": 0.038236734693877554,
"grad_norm": 3.0334460735321045,
"learning_rate": 9.963389941214893e-06,
"loss": 0.7999,
"step": 1171
},
{
"epoch": 0.03826938775510204,
"grad_norm": 3.0638813972473145,
"learning_rate": 9.963357282821686e-06,
"loss": 0.7922,
"step": 1172
},
{
"epoch": 0.03830204081632653,
"grad_norm": 3.0759074687957764,
"learning_rate": 9.963324624428478e-06,
"loss": 0.7521,
"step": 1173
},
{
"epoch": 0.03833469387755102,
"grad_norm": 3.0248312950134277,
"learning_rate": 9.963291966035273e-06,
"loss": 0.7695,
"step": 1174
},
{
"epoch": 0.03836734693877551,
"grad_norm": 3.101550817489624,
"learning_rate": 9.963259307642064e-06,
"loss": 0.7446,
"step": 1175
},
{
"epoch": 0.0384,
"grad_norm": 3.0817489624023438,
"learning_rate": 9.963226649248858e-06,
"loss": 0.7653,
"step": 1176
},
{
"epoch": 0.03843265306122449,
"grad_norm": 3.051537036895752,
"learning_rate": 9.96319399085565e-06,
"loss": 0.7723,
"step": 1177
},
{
"epoch": 0.03846530612244898,
"grad_norm": 3.030776023864746,
"learning_rate": 9.963161332462442e-06,
"loss": 0.7762,
"step": 1178
},
{
"epoch": 0.03849795918367347,
"grad_norm": 3.1081271171569824,
"learning_rate": 9.963128674069237e-06,
"loss": 0.8198,
"step": 1179
},
{
"epoch": 0.03853061224489796,
"grad_norm": 3.1396594047546387,
"learning_rate": 9.963096015676029e-06,
"loss": 0.7389,
"step": 1180
},
{
"epoch": 0.038563265306122446,
"grad_norm": 2.977698564529419,
"learning_rate": 9.963063357282822e-06,
"loss": 0.7961,
"step": 1181
},
{
"epoch": 0.03859591836734694,
"grad_norm": 3.048220157623291,
"learning_rate": 9.963030698889615e-06,
"loss": 0.8025,
"step": 1182
},
{
"epoch": 0.03862857142857143,
"grad_norm": 3.087770700454712,
"learning_rate": 9.962998040496409e-06,
"loss": 0.6993,
"step": 1183
},
{
"epoch": 0.03866122448979592,
"grad_norm": 3.0148611068725586,
"learning_rate": 9.962965382103202e-06,
"loss": 0.7555,
"step": 1184
},
{
"epoch": 0.03869387755102041,
"grad_norm": 3.4241836071014404,
"learning_rate": 9.962932723709993e-06,
"loss": 0.7876,
"step": 1185
},
{
"epoch": 0.038726530612244896,
"grad_norm": 3.200383424758911,
"learning_rate": 9.962900065316787e-06,
"loss": 0.7739,
"step": 1186
},
{
"epoch": 0.03875918367346939,
"grad_norm": 3.0834004878997803,
"learning_rate": 9.96286740692358e-06,
"loss": 0.7524,
"step": 1187
},
{
"epoch": 0.03879183673469388,
"grad_norm": 2.9796390533447266,
"learning_rate": 9.962834748530373e-06,
"loss": 0.7495,
"step": 1188
},
{
"epoch": 0.03882448979591837,
"grad_norm": 3.1456053256988525,
"learning_rate": 9.962802090137166e-06,
"loss": 0.7939,
"step": 1189
},
{
"epoch": 0.038857142857142854,
"grad_norm": 3.227130651473999,
"learning_rate": 9.96276943174396e-06,
"loss": 0.7283,
"step": 1190
},
{
"epoch": 0.038889795918367345,
"grad_norm": 3.2170636653900146,
"learning_rate": 9.962736773350751e-06,
"loss": 0.7481,
"step": 1191
},
{
"epoch": 0.03892244897959184,
"grad_norm": 6.70438814163208,
"learning_rate": 9.962704114957546e-06,
"loss": 0.7565,
"step": 1192
},
{
"epoch": 0.03895510204081633,
"grad_norm": 3.2916693687438965,
"learning_rate": 9.962671456564338e-06,
"loss": 0.7732,
"step": 1193
},
{
"epoch": 0.03898775510204082,
"grad_norm": 3.4526243209838867,
"learning_rate": 9.96263879817113e-06,
"loss": 0.7563,
"step": 1194
},
{
"epoch": 0.039020408163265304,
"grad_norm": 3.188997507095337,
"learning_rate": 9.962606139777924e-06,
"loss": 0.771,
"step": 1195
},
{
"epoch": 0.039053061224489795,
"grad_norm": 3.0594732761383057,
"learning_rate": 9.962573481384716e-06,
"loss": 0.7433,
"step": 1196
},
{
"epoch": 0.039085714285714286,
"grad_norm": 3.239617109298706,
"learning_rate": 9.96254082299151e-06,
"loss": 0.7823,
"step": 1197
},
{
"epoch": 0.03911836734693878,
"grad_norm": 2.983693838119507,
"learning_rate": 9.962508164598302e-06,
"loss": 0.7557,
"step": 1198
},
{
"epoch": 0.03915102040816327,
"grad_norm": 2.9180068969726562,
"learning_rate": 9.962475506205095e-06,
"loss": 0.7915,
"step": 1199
},
{
"epoch": 0.03918367346938775,
"grad_norm": 3.102025270462036,
"learning_rate": 9.962442847811888e-06,
"loss": 0.764,
"step": 1200
},
{
"epoch": 0.03918367346938775,
"eval_loss": 0.807881772518158,
"eval_runtime": 85.0203,
"eval_samples_per_second": 1.176,
"eval_steps_per_second": 1.176,
"step": 1200
}
],
"logging_steps": 1,
"max_steps": 306250,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.305701335141549e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}