bbytxt's picture
Training in progress, step 150, checkpoint
ad10459 verified
raw
history blame
26.4 kB
{
"best_metric": 6.682973861694336,
"best_model_checkpoint": "miner_id_24/checkpoint-150",
"epoch": 0.96,
"eval_steps": 25,
"global_step": 150,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0064,
"grad_norm": 0.9495358467102051,
"learning_rate": 2.9999999999999997e-05,
"loss": 6.9341,
"step": 1
},
{
"epoch": 0.0064,
"eval_loss": 6.940670013427734,
"eval_runtime": 0.4624,
"eval_samples_per_second": 108.123,
"eval_steps_per_second": 15.137,
"step": 1
},
{
"epoch": 0.0128,
"grad_norm": 0.8018835186958313,
"learning_rate": 5.9999999999999995e-05,
"loss": 6.9556,
"step": 2
},
{
"epoch": 0.0192,
"grad_norm": 0.9049068093299866,
"learning_rate": 8.999999999999999e-05,
"loss": 6.9653,
"step": 3
},
{
"epoch": 0.0256,
"grad_norm": 0.7154098749160767,
"learning_rate": 0.00011999999999999999,
"loss": 6.9551,
"step": 4
},
{
"epoch": 0.032,
"grad_norm": 0.7742681503295898,
"learning_rate": 0.00015,
"loss": 6.9498,
"step": 5
},
{
"epoch": 0.0384,
"grad_norm": 0.8399883508682251,
"learning_rate": 0.00017999999999999998,
"loss": 6.94,
"step": 6
},
{
"epoch": 0.0448,
"grad_norm": 0.5200990438461304,
"learning_rate": 0.00020999999999999998,
"loss": 6.9413,
"step": 7
},
{
"epoch": 0.0512,
"grad_norm": 0.6034143567085266,
"learning_rate": 0.00023999999999999998,
"loss": 6.9438,
"step": 8
},
{
"epoch": 0.0576,
"grad_norm": 0.5844438672065735,
"learning_rate": 0.00027,
"loss": 6.9437,
"step": 9
},
{
"epoch": 0.064,
"grad_norm": 0.6551675796508789,
"learning_rate": 0.0003,
"loss": 6.943,
"step": 10
},
{
"epoch": 0.0704,
"grad_norm": 0.6681671142578125,
"learning_rate": 0.0002999794957488703,
"loss": 6.9383,
"step": 11
},
{
"epoch": 0.0768,
"grad_norm": 0.6732079386711121,
"learning_rate": 0.0002999179886011389,
"loss": 6.9406,
"step": 12
},
{
"epoch": 0.0832,
"grad_norm": 0.6102283596992493,
"learning_rate": 0.0002998154953722457,
"loss": 6.9383,
"step": 13
},
{
"epoch": 0.0896,
"grad_norm": 0.6549004912376404,
"learning_rate": 0.00029967204408281613,
"loss": 6.9284,
"step": 14
},
{
"epoch": 0.096,
"grad_norm": 0.5962654948234558,
"learning_rate": 0.00029948767395100045,
"loss": 6.9047,
"step": 15
},
{
"epoch": 0.1024,
"grad_norm": 0.649928867816925,
"learning_rate": 0.0002992624353817517,
"loss": 6.9212,
"step": 16
},
{
"epoch": 0.1088,
"grad_norm": 0.504554808139801,
"learning_rate": 0.0002989963899530457,
"loss": 6.9179,
"step": 17
},
{
"epoch": 0.1152,
"grad_norm": 0.597122848033905,
"learning_rate": 0.00029868961039904624,
"loss": 6.9112,
"step": 18
},
{
"epoch": 0.1216,
"grad_norm": 0.5620957612991333,
"learning_rate": 0.00029834218059022024,
"loss": 6.9046,
"step": 19
},
{
"epoch": 0.128,
"grad_norm": 0.6724961400032043,
"learning_rate": 0.00029795419551040833,
"loss": 6.8993,
"step": 20
},
{
"epoch": 0.1344,
"grad_norm": 0.5969198346138,
"learning_rate": 0.00029752576123085736,
"loss": 6.8979,
"step": 21
},
{
"epoch": 0.1408,
"grad_norm": 0.5506598949432373,
"learning_rate": 0.0002970569948812214,
"loss": 6.8887,
"step": 22
},
{
"epoch": 0.1472,
"grad_norm": 0.7546066641807556,
"learning_rate": 0.0002965480246175399,
"loss": 6.8619,
"step": 23
},
{
"epoch": 0.1536,
"grad_norm": 0.6300695538520813,
"learning_rate": 0.0002959989895872009,
"loss": 6.8506,
"step": 24
},
{
"epoch": 0.16,
"grad_norm": 0.5741549134254456,
"learning_rate": 0.0002954100398908995,
"loss": 6.8638,
"step": 25
},
{
"epoch": 0.16,
"eval_loss": 6.874875545501709,
"eval_runtime": 0.047,
"eval_samples_per_second": 1063.777,
"eval_steps_per_second": 148.929,
"step": 25
},
{
"epoch": 0.1664,
"grad_norm": 0.5633116364479065,
"learning_rate": 0.0002947813365416023,
"loss": 6.8588,
"step": 26
},
{
"epoch": 0.1728,
"grad_norm": 0.6096696853637695,
"learning_rate": 0.0002941130514205272,
"loss": 6.8797,
"step": 27
},
{
"epoch": 0.1792,
"grad_norm": 0.5343248844146729,
"learning_rate": 0.0002934053672301536,
"loss": 6.8702,
"step": 28
},
{
"epoch": 0.1856,
"grad_norm": 0.6776606440544128,
"learning_rate": 0.00029265847744427303,
"loss": 6.84,
"step": 29
},
{
"epoch": 0.192,
"grad_norm": 0.5798454880714417,
"learning_rate": 0.00029187258625509513,
"loss": 6.8618,
"step": 30
},
{
"epoch": 0.1984,
"grad_norm": 0.6195558905601501,
"learning_rate": 0.00029104790851742417,
"loss": 6.8286,
"step": 31
},
{
"epoch": 0.2048,
"grad_norm": 0.556268036365509,
"learning_rate": 0.0002901846696899191,
"loss": 6.8496,
"step": 32
},
{
"epoch": 0.2112,
"grad_norm": 0.5736655592918396,
"learning_rate": 0.00028928310577345606,
"loss": 6.8319,
"step": 33
},
{
"epoch": 0.2176,
"grad_norm": 0.5479353666305542,
"learning_rate": 0.0002883434632466077,
"loss": 6.8609,
"step": 34
},
{
"epoch": 0.224,
"grad_norm": 0.8071677684783936,
"learning_rate": 0.00028736599899825856,
"loss": 6.9076,
"step": 35
},
{
"epoch": 0.2304,
"grad_norm": 0.775070071220398,
"learning_rate": 0.00028635098025737434,
"loss": 6.9059,
"step": 36
},
{
"epoch": 0.2368,
"grad_norm": 0.7527448534965515,
"learning_rate": 0.00028529868451994384,
"loss": 6.8908,
"step": 37
},
{
"epoch": 0.2432,
"grad_norm": 0.911148726940155,
"learning_rate": 0.0002842093994731145,
"loss": 6.8912,
"step": 38
},
{
"epoch": 0.2496,
"grad_norm": 1.4041032791137695,
"learning_rate": 0.00028308342291654174,
"loss": 6.8862,
"step": 39
},
{
"epoch": 0.256,
"grad_norm": 0.9864165186882019,
"learning_rate": 0.00028192106268097334,
"loss": 6.9131,
"step": 40
},
{
"epoch": 0.2624,
"grad_norm": 0.8453418612480164,
"learning_rate": 0.00028072263654409154,
"loss": 6.9059,
"step": 41
},
{
"epoch": 0.2688,
"grad_norm": 0.77274489402771,
"learning_rate": 0.0002794884721436361,
"loss": 6.8597,
"step": 42
},
{
"epoch": 0.2752,
"grad_norm": 0.7312727570533752,
"learning_rate": 0.00027821890688783083,
"loss": 6.8523,
"step": 43
},
{
"epoch": 0.2816,
"grad_norm": 0.7589954137802124,
"learning_rate": 0.0002769142878631403,
"loss": 6.8514,
"step": 44
},
{
"epoch": 0.288,
"grad_norm": 0.588603138923645,
"learning_rate": 0.00027557497173937923,
"loss": 6.7881,
"step": 45
},
{
"epoch": 0.2944,
"grad_norm": 0.632996678352356,
"learning_rate": 0.000274201324672203,
"loss": 6.7755,
"step": 46
},
{
"epoch": 0.3008,
"grad_norm": 0.575196385383606,
"learning_rate": 0.00027279372220300385,
"loss": 6.79,
"step": 47
},
{
"epoch": 0.3072,
"grad_norm": 0.5218028426170349,
"learning_rate": 0.0002713525491562421,
"loss": 6.7852,
"step": 48
},
{
"epoch": 0.3136,
"grad_norm": 0.5754166841506958,
"learning_rate": 0.00026987819953423867,
"loss": 6.7727,
"step": 49
},
{
"epoch": 0.32,
"grad_norm": 0.5350818634033203,
"learning_rate": 0.00026837107640945905,
"loss": 6.7583,
"step": 50
},
{
"epoch": 0.32,
"eval_loss": 6.796112060546875,
"eval_runtime": 0.0455,
"eval_samples_per_second": 1098.877,
"eval_steps_per_second": 153.843,
"step": 50
},
{
"epoch": 0.3264,
"grad_norm": 0.4826822578907013,
"learning_rate": 0.0002668315918143169,
"loss": 6.7629,
"step": 51
},
{
"epoch": 0.3328,
"grad_norm": 0.508161187171936,
"learning_rate": 0.00026526016662852886,
"loss": 6.7403,
"step": 52
},
{
"epoch": 0.3392,
"grad_norm": 0.5200028419494629,
"learning_rate": 0.00026365723046405023,
"loss": 6.7561,
"step": 53
},
{
"epoch": 0.3456,
"grad_norm": 0.49363598227500916,
"learning_rate": 0.0002620232215476231,
"loss": 6.7514,
"step": 54
},
{
"epoch": 0.352,
"grad_norm": 0.46784576773643494,
"learning_rate": 0.0002603585866009697,
"loss": 6.7524,
"step": 55
},
{
"epoch": 0.3584,
"grad_norm": 0.46129316091537476,
"learning_rate": 0.00025866378071866334,
"loss": 6.7433,
"step": 56
},
{
"epoch": 0.3648,
"grad_norm": 0.5088637471199036,
"learning_rate": 0.00025693926724370956,
"loss": 6.7277,
"step": 57
},
{
"epoch": 0.3712,
"grad_norm": 0.5694301724433899,
"learning_rate": 0.00025518551764087326,
"loss": 6.7277,
"step": 58
},
{
"epoch": 0.3776,
"grad_norm": 0.41760918498039246,
"learning_rate": 0.00025340301136778483,
"loss": 6.7531,
"step": 59
},
{
"epoch": 0.384,
"grad_norm": 0.4982942044734955,
"learning_rate": 0.00025159223574386114,
"loss": 6.729,
"step": 60
},
{
"epoch": 0.3904,
"grad_norm": 0.4909391701221466,
"learning_rate": 0.0002497536858170772,
"loss": 6.7228,
"step": 61
},
{
"epoch": 0.3968,
"grad_norm": 0.5273060202598572,
"learning_rate": 0.00024788786422862526,
"loss": 6.7134,
"step": 62
},
{
"epoch": 0.4032,
"grad_norm": 0.4253634512424469,
"learning_rate": 0.00024599528107549745,
"loss": 6.7223,
"step": 63
},
{
"epoch": 0.4096,
"grad_norm": 0.47950324416160583,
"learning_rate": 0.00024407645377103054,
"loss": 6.718,
"step": 64
},
{
"epoch": 0.416,
"grad_norm": 0.46385979652404785,
"learning_rate": 0.00024213190690345018,
"loss": 6.725,
"step": 65
},
{
"epoch": 0.4224,
"grad_norm": 0.540244996547699,
"learning_rate": 0.00024016217209245374,
"loss": 6.7151,
"step": 66
},
{
"epoch": 0.4288,
"grad_norm": 0.4589102864265442,
"learning_rate": 0.00023816778784387094,
"loss": 6.7114,
"step": 67
},
{
"epoch": 0.4352,
"grad_norm": 0.4630209505558014,
"learning_rate": 0.0002361492994024415,
"loss": 6.7337,
"step": 68
},
{
"epoch": 0.4416,
"grad_norm": 0.47441697120666504,
"learning_rate": 0.0002341072586027509,
"loss": 6.7308,
"step": 69
},
{
"epoch": 0.448,
"grad_norm": 0.49404001235961914,
"learning_rate": 0.00023204222371836405,
"loss": 6.7455,
"step": 70
},
{
"epoch": 0.4544,
"grad_norm": 0.44997942447662354,
"learning_rate": 0.00022995475930919905,
"loss": 6.739,
"step": 71
},
{
"epoch": 0.4608,
"grad_norm": 0.45261090993881226,
"learning_rate": 0.00022784543606718227,
"loss": 6.7068,
"step": 72
},
{
"epoch": 0.4672,
"grad_norm": 0.44797801971435547,
"learning_rate": 0.00022571483066022657,
"loss": 6.7162,
"step": 73
},
{
"epoch": 0.4736,
"grad_norm": 0.6951871514320374,
"learning_rate": 0.0002235635255745762,
"loss": 6.7901,
"step": 74
},
{
"epoch": 0.48,
"grad_norm": 0.7108526825904846,
"learning_rate": 0.00022139210895556104,
"loss": 6.8099,
"step": 75
},
{
"epoch": 0.48,
"eval_loss": 6.746245861053467,
"eval_runtime": 0.0445,
"eval_samples_per_second": 1123.84,
"eval_steps_per_second": 157.338,
"step": 75
},
{
"epoch": 0.4864,
"grad_norm": 0.7808616161346436,
"learning_rate": 0.00021920117444680317,
"loss": 6.8143,
"step": 76
},
{
"epoch": 0.4928,
"grad_norm": 0.7759772539138794,
"learning_rate": 0.00021699132102792097,
"loss": 6.8032,
"step": 77
},
{
"epoch": 0.4992,
"grad_norm": 1.509546160697937,
"learning_rate": 0.0002147631528507739,
"loss": 6.7856,
"step": 78
},
{
"epoch": 0.5056,
"grad_norm": 1.1290587186813354,
"learning_rate": 0.00021251727907429355,
"loss": 6.8949,
"step": 79
},
{
"epoch": 0.512,
"grad_norm": 0.9573540687561035,
"learning_rate": 0.0002102543136979454,
"loss": 6.887,
"step": 80
},
{
"epoch": 0.5184,
"grad_norm": 1.0016741752624512,
"learning_rate": 0.0002079748753938678,
"loss": 6.8455,
"step": 81
},
{
"epoch": 0.5248,
"grad_norm": 0.7958142161369324,
"learning_rate": 0.0002056795873377331,
"loss": 6.7827,
"step": 82
},
{
"epoch": 0.5312,
"grad_norm": 0.6969197392463684,
"learning_rate": 0.00020336907703837748,
"loss": 6.7816,
"step": 83
},
{
"epoch": 0.5376,
"grad_norm": 0.8457598090171814,
"learning_rate": 0.00020104397616624645,
"loss": 6.7871,
"step": 84
},
{
"epoch": 0.544,
"grad_norm": 0.5099653005599976,
"learning_rate": 0.00019870492038070252,
"loss": 6.6791,
"step": 85
},
{
"epoch": 0.5504,
"grad_norm": 0.5534845590591431,
"learning_rate": 0.0001963525491562421,
"loss": 6.685,
"step": 86
},
{
"epoch": 0.5568,
"grad_norm": 0.5635961294174194,
"learning_rate": 0.0001939875056076697,
"loss": 6.694,
"step": 87
},
{
"epoch": 0.5632,
"grad_norm": 0.5438621640205383,
"learning_rate": 0.00019161043631427666,
"loss": 6.6969,
"step": 88
},
{
"epoch": 0.5696,
"grad_norm": 0.5128461122512817,
"learning_rate": 0.00018922199114307294,
"loss": 6.6693,
"step": 89
},
{
"epoch": 0.576,
"grad_norm": 0.6132482290267944,
"learning_rate": 0.00018682282307111987,
"loss": 6.6821,
"step": 90
},
{
"epoch": 0.5824,
"grad_norm": 0.5470501184463501,
"learning_rate": 0.00018441358800701273,
"loss": 6.6796,
"step": 91
},
{
"epoch": 0.5888,
"grad_norm": 0.5392007231712341,
"learning_rate": 0.00018199494461156203,
"loss": 6.6488,
"step": 92
},
{
"epoch": 0.5952,
"grad_norm": 0.6648130416870117,
"learning_rate": 0.000179567554117722,
"loss": 6.6436,
"step": 93
},
{
"epoch": 0.6016,
"grad_norm": 0.6511263847351074,
"learning_rate": 0.00017713208014981648,
"loss": 6.6311,
"step": 94
},
{
"epoch": 0.608,
"grad_norm": 0.6294538974761963,
"learning_rate": 0.00017468918854211007,
"loss": 6.627,
"step": 95
},
{
"epoch": 0.6144,
"grad_norm": 0.48442429304122925,
"learning_rate": 0.00017223954715677627,
"loss": 6.6497,
"step": 96
},
{
"epoch": 0.6208,
"grad_norm": 0.5661095976829529,
"learning_rate": 0.00016978382570131034,
"loss": 6.6348,
"step": 97
},
{
"epoch": 0.6272,
"grad_norm": 0.47108545899391174,
"learning_rate": 0.00016732269554543794,
"loss": 6.6478,
"step": 98
},
{
"epoch": 0.6336,
"grad_norm": 0.43120548129081726,
"learning_rate": 0.00016485682953756942,
"loss": 6.6553,
"step": 99
},
{
"epoch": 0.64,
"grad_norm": 0.47978243231773376,
"learning_rate": 0.00016238690182084986,
"loss": 6.6235,
"step": 100
},
{
"epoch": 0.64,
"eval_loss": 6.714487075805664,
"eval_runtime": 0.0452,
"eval_samples_per_second": 1106.805,
"eval_steps_per_second": 154.953,
"step": 100
},
{
"epoch": 0.6464,
"grad_norm": 0.442552387714386,
"learning_rate": 0.0001599135876488549,
"loss": 6.6315,
"step": 101
},
{
"epoch": 0.6528,
"grad_norm": 0.4487893283367157,
"learning_rate": 0.00015743756320098332,
"loss": 6.641,
"step": 102
},
{
"epoch": 0.6592,
"grad_norm": 0.4315354824066162,
"learning_rate": 0.0001549595053975962,
"loss": 6.6571,
"step": 103
},
{
"epoch": 0.6656,
"grad_norm": 0.44835346937179565,
"learning_rate": 0.00015248009171495378,
"loss": 6.6217,
"step": 104
},
{
"epoch": 0.672,
"grad_norm": 0.38457152247428894,
"learning_rate": 0.00015,
"loss": 6.6733,
"step": 105
},
{
"epoch": 0.6784,
"grad_norm": 0.44234663248062134,
"learning_rate": 0.00014751990828504622,
"loss": 6.6458,
"step": 106
},
{
"epoch": 0.6848,
"grad_norm": 0.5025720000267029,
"learning_rate": 0.00014504049460240375,
"loss": 6.6981,
"step": 107
},
{
"epoch": 0.6912,
"grad_norm": 0.467263400554657,
"learning_rate": 0.00014256243679901663,
"loss": 6.6535,
"step": 108
},
{
"epoch": 0.6976,
"grad_norm": 0.4463280141353607,
"learning_rate": 0.00014008641235114508,
"loss": 6.6794,
"step": 109
},
{
"epoch": 0.704,
"grad_norm": 0.5022115111351013,
"learning_rate": 0.00013761309817915014,
"loss": 6.6662,
"step": 110
},
{
"epoch": 0.7104,
"grad_norm": 0.47702935338020325,
"learning_rate": 0.00013514317046243058,
"loss": 6.6882,
"step": 111
},
{
"epoch": 0.7168,
"grad_norm": 0.6275339722633362,
"learning_rate": 0.00013267730445456208,
"loss": 6.7397,
"step": 112
},
{
"epoch": 0.7232,
"grad_norm": 0.8845940828323364,
"learning_rate": 0.00013021617429868963,
"loss": 6.7815,
"step": 113
},
{
"epoch": 0.7296,
"grad_norm": 0.8113049864768982,
"learning_rate": 0.00012776045284322368,
"loss": 6.7837,
"step": 114
},
{
"epoch": 0.736,
"grad_norm": 0.8991756439208984,
"learning_rate": 0.00012531081145788987,
"loss": 6.7817,
"step": 115
},
{
"epoch": 0.7424,
"grad_norm": 1.0973552465438843,
"learning_rate": 0.00012286791985018355,
"loss": 6.7701,
"step": 116
},
{
"epoch": 0.7488,
"grad_norm": 1.67232346534729,
"learning_rate": 0.00012043244588227796,
"loss": 6.7657,
"step": 117
},
{
"epoch": 0.7552,
"grad_norm": 1.071197748184204,
"learning_rate": 0.00011800505538843798,
"loss": 6.8816,
"step": 118
},
{
"epoch": 0.7616,
"grad_norm": 0.8870216608047485,
"learning_rate": 0.00011558641199298727,
"loss": 6.854,
"step": 119
},
{
"epoch": 0.768,
"grad_norm": 0.8856467604637146,
"learning_rate": 0.00011317717692888012,
"loss": 6.8337,
"step": 120
},
{
"epoch": 0.7744,
"grad_norm": 0.7022458910942078,
"learning_rate": 0.00011077800885692702,
"loss": 6.7082,
"step": 121
},
{
"epoch": 0.7808,
"grad_norm": 0.8062697052955627,
"learning_rate": 0.00010838956368572334,
"loss": 6.7377,
"step": 122
},
{
"epoch": 0.7872,
"grad_norm": 0.9013355374336243,
"learning_rate": 0.0001060124943923303,
"loss": 6.7669,
"step": 123
},
{
"epoch": 0.7936,
"grad_norm": 0.5724779367446899,
"learning_rate": 0.0001036474508437579,
"loss": 6.6174,
"step": 124
},
{
"epoch": 0.8,
"grad_norm": 0.5374643802642822,
"learning_rate": 0.00010129507961929748,
"loss": 6.6612,
"step": 125
},
{
"epoch": 0.8,
"eval_loss": 6.694909572601318,
"eval_runtime": 0.0438,
"eval_samples_per_second": 1142.364,
"eval_steps_per_second": 159.931,
"step": 125
},
{
"epoch": 0.8064,
"grad_norm": 0.5468215346336365,
"learning_rate": 9.895602383375353e-05,
"loss": 6.667,
"step": 126
},
{
"epoch": 0.8128,
"grad_norm": 0.4469563364982605,
"learning_rate": 9.663092296162251e-05,
"loss": 6.6637,
"step": 127
},
{
"epoch": 0.8192,
"grad_norm": 0.6111487150192261,
"learning_rate": 9.432041266226686e-05,
"loss": 6.6008,
"step": 128
},
{
"epoch": 0.8256,
"grad_norm": 0.46038818359375,
"learning_rate": 9.202512460613219e-05,
"loss": 6.6134,
"step": 129
},
{
"epoch": 0.832,
"grad_norm": 0.5608881115913391,
"learning_rate": 8.97456863020546e-05,
"loss": 6.5824,
"step": 130
},
{
"epoch": 0.8384,
"grad_norm": 0.5066159963607788,
"learning_rate": 8.748272092570646e-05,
"loss": 6.5936,
"step": 131
},
{
"epoch": 0.8448,
"grad_norm": 0.35695910453796387,
"learning_rate": 8.523684714922608e-05,
"loss": 6.6303,
"step": 132
},
{
"epoch": 0.8512,
"grad_norm": 0.5033596158027649,
"learning_rate": 8.300867897207903e-05,
"loss": 6.583,
"step": 133
},
{
"epoch": 0.8576,
"grad_norm": 0.417173832654953,
"learning_rate": 8.079882555319684e-05,
"loss": 6.6496,
"step": 134
},
{
"epoch": 0.864,
"grad_norm": 0.3751719892024994,
"learning_rate": 7.860789104443896e-05,
"loss": 6.6507,
"step": 135
},
{
"epoch": 0.8704,
"grad_norm": 0.3995749056339264,
"learning_rate": 7.643647442542382e-05,
"loss": 6.5948,
"step": 136
},
{
"epoch": 0.8768,
"grad_norm": 0.4013768136501312,
"learning_rate": 7.428516933977347e-05,
"loss": 6.6164,
"step": 137
},
{
"epoch": 0.8832,
"grad_norm": 0.3744851052761078,
"learning_rate": 7.215456393281776e-05,
"loss": 6.6189,
"step": 138
},
{
"epoch": 0.8896,
"grad_norm": 0.3870219886302948,
"learning_rate": 7.004524069080096e-05,
"loss": 6.6008,
"step": 139
},
{
"epoch": 0.896,
"grad_norm": 0.42525532841682434,
"learning_rate": 6.795777628163599e-05,
"loss": 6.6112,
"step": 140
},
{
"epoch": 0.9024,
"grad_norm": 0.3942902088165283,
"learning_rate": 6.58927413972491e-05,
"loss": 6.5992,
"step": 141
},
{
"epoch": 0.9088,
"grad_norm": 0.48578906059265137,
"learning_rate": 6.385070059755846e-05,
"loss": 6.5816,
"step": 142
},
{
"epoch": 0.9152,
"grad_norm": 0.36412331461906433,
"learning_rate": 6.183221215612904e-05,
"loss": 6.6061,
"step": 143
},
{
"epoch": 0.9216,
"grad_norm": 0.4250425398349762,
"learning_rate": 5.983782790754623e-05,
"loss": 6.6691,
"step": 144
},
{
"epoch": 0.928,
"grad_norm": 0.4022601842880249,
"learning_rate": 5.786809309654982e-05,
"loss": 6.6641,
"step": 145
},
{
"epoch": 0.9344,
"grad_norm": 0.5123599171638489,
"learning_rate": 5.592354622896944e-05,
"loss": 6.6763,
"step": 146
},
{
"epoch": 0.9408,
"grad_norm": 0.44589608907699585,
"learning_rate": 5.40047189245025e-05,
"loss": 6.6136,
"step": 147
},
{
"epoch": 0.9472,
"grad_norm": 0.4042072892189026,
"learning_rate": 5.211213577137469e-05,
"loss": 6.6563,
"step": 148
},
{
"epoch": 0.9536,
"grad_norm": 0.4555412232875824,
"learning_rate": 5.024631418292274e-05,
"loss": 6.661,
"step": 149
},
{
"epoch": 0.96,
"grad_norm": 0.3993721008300781,
"learning_rate": 4.840776425613886e-05,
"loss": 6.674,
"step": 150
},
{
"epoch": 0.96,
"eval_loss": 6.682973861694336,
"eval_runtime": 0.0454,
"eval_samples_per_second": 1101.37,
"eval_steps_per_second": 154.192,
"step": 150
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 1,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1878403276800.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}