|
{ |
|
"best_metric": 0.8823016285896301, |
|
"best_model_checkpoint": "output/headie-one/checkpoint-3900", |
|
"epoch": 52.0, |
|
"global_step": 3900, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0001357399755894565, |
|
"loss": 0.4785, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00013142205020853694, |
|
"loss": 0.55, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00012443002200473538, |
|
"loss": 0.5739, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00011506151581352585, |
|
"loss": 0.5448, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00010371531436754662, |
|
"loss": 0.5863, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.08743835874413e-05, |
|
"loss": 0.5127, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 7.708531450507727e-05, |
|
"loss": 0.5548, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 6.293505690059801e-05, |
|
"loss": 0.5604, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.902593501651139e-05, |
|
"loss": 0.5144, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.5950008837656755e-05, |
|
"loss": 0.6611, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.4263872282739445e-05, |
|
"loss": 0.5217, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.4464961055407408e-05, |
|
"loss": 0.544, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 6.97037864221025e-06, |
|
"loss": 0.5656, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.0991417565619363e-06, |
|
"loss": 0.4698, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 5.860097463115006e-08, |
|
"loss": 0.5957, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 5.471560001373291, |
|
"eval_runtime": 2.3923, |
|
"eval_samples_per_second": 42.637, |
|
"eval_steps_per_second": 5.434, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.35614586573241e-07, |
|
"loss": 0.5587, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.692851359682261e-06, |
|
"loss": 0.5065, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.1170379591190527e-05, |
|
"loss": 0.4748, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.009247481060283e-05, |
|
"loss": 0.408, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 3.1079356352801514e-05, |
|
"loss": 0.3892, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.3663353219321276e-05, |
|
"loss": 0.4958, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.7308811106741675e-05, |
|
"loss": 0.3676, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.143489323346855e-05, |
|
"loss": 0.4046, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.54403044178588e-05, |
|
"loss": 0.4215, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.87288859949296e-05, |
|
"loss": 0.552, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 0.00011073499209051121, |
|
"loss": 0.4987, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.00012094756707850673, |
|
"loss": 0.5333, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.00012893189933276512, |
|
"loss": 0.5114, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.00013434812529663614, |
|
"loss": 0.4435, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.00013696569622025754, |
|
"loss": 0.5415, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 5.518349647521973, |
|
"eval_runtime": 2.4007, |
|
"eval_samples_per_second": 42.487, |
|
"eval_steps_per_second": 5.415, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 0.00013667319176996132, |
|
"loss": 0.4488, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 0.00013348306278066356, |
|
"loss": 0.4539, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 0.00012753110126977412, |
|
"loss": 0.3929, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 0.00011907066027217684, |
|
"loss": 0.4031, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.00010846186953657296, |
|
"loss": 0.4346, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 9.615630613119375e-05, |
|
"loss": 0.5194, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 8.267777247462053e-05, |
|
"loss": 0.4893, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 6.860000000000001e-05, |
|
"loss": 0.4725, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.45222275253795e-05, |
|
"loss": 0.4897, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 4.1043693868806304e-05, |
|
"loss": 0.4427, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.873813046342713e-05, |
|
"loss": 0.4814, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.812933972782316e-05, |
|
"loss": 0.4612, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 9.668898730225861e-06, |
|
"loss": 0.4338, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.7169372193364486e-06, |
|
"loss": 0.4686, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 5.268082300387043e-07, |
|
"loss": 0.4222, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 5.570201873779297, |
|
"eval_runtime": 2.378, |
|
"eval_samples_per_second": 42.893, |
|
"eval_steps_per_second": 5.467, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.3430377974244087e-07, |
|
"loss": 0.3902, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.851874703363849e-06, |
|
"loss": 0.3359, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 8.268100667234827e-06, |
|
"loss": 0.3495, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.6252432921493247e-05, |
|
"loss": 0.4064, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.6465007909488756e-05, |
|
"loss": 0.4271, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.847111400507037e-05, |
|
"loss": 0.3916, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 5.175969558214116e-05, |
|
"loss": 0.356, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 6.576510676653141e-05, |
|
"loss": 0.3824, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 7.989118889325835e-05, |
|
"loss": 0.3627, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 9.353664678067862e-05, |
|
"loss": 0.4006, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 0.0001061206436471984, |
|
"loss": 0.382, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 0.0001171075251893971, |
|
"loss": 0.4378, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 0.0001260296204088095, |
|
"loss": 0.4331, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 0.00013250714864031774, |
|
"loss": 0.3896, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 0.00013626438541342674, |
|
"loss": 0.4182, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 5.707557678222656, |
|
"eval_runtime": 2.3844, |
|
"eval_samples_per_second": 42.778, |
|
"eval_steps_per_second": 5.452, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.3989, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 0.0001351008582434381, |
|
"loss": 0.3894, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 0.0001302296213577898, |
|
"loss": 0.3742, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 0.00012273503894459262, |
|
"loss": 0.3937, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 0.00011293612771726056, |
|
"loss": 0.4215, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 0.00010124999116234326, |
|
"loss": 0.3937, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 8.817406498348863e-05, |
|
"loss": 0.4438, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 7.426494309940199e-05, |
|
"loss": 0.4253, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 6.011468549492274e-05, |
|
"loss": 0.3437, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 4.6325616412558686e-05, |
|
"loss": 0.3752, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 3.3484685632453346e-05, |
|
"loss": 0.4207, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 2.2138484186474223e-05, |
|
"loss": 0.3517, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 1.2769977995264667e-05, |
|
"loss": 0.3324, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 5.777949791463086e-06, |
|
"loss": 0.3575, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 1.4600244105434905e-06, |
|
"loss": 0.3555, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.3544, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 5.685351371765137, |
|
"eval_runtime": 2.3666, |
|
"eval_samples_per_second": 43.1, |
|
"eval_steps_per_second": 5.493, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 1.4600244105434752e-06, |
|
"loss": 0.3192, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 5.777949791463048e-06, |
|
"loss": 0.3275, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 1.2769977995264689e-05, |
|
"loss": 0.282, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 2.2138484186474162e-05, |
|
"loss": 0.2882, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 3.3484685632453264e-05, |
|
"loss": 0.3381, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 4.632561641255849e-05, |
|
"loss": 0.3144, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 6.011468549492265e-05, |
|
"loss": 0.2941, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 7.426494309940202e-05, |
|
"loss": 0.2993, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 8.817406498348855e-05, |
|
"loss": 0.3661, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 0.00010124999116234329, |
|
"loss": 0.3125, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 0.0001129361277172605, |
|
"loss": 0.3826, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 0.00012273503894459265, |
|
"loss": 0.3616, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 0.00013022962135778973, |
|
"loss": 0.3224, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 0.0001351008582434381, |
|
"loss": 0.3485, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.3931, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 5.795953273773193, |
|
"eval_runtime": 2.3912, |
|
"eval_samples_per_second": 42.656, |
|
"eval_steps_per_second": 5.437, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 0.00013626438541342677, |
|
"loss": 0.4352, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 0.00013250714864031774, |
|
"loss": 0.3058, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 0.00012602962040880953, |
|
"loss": 0.3144, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 0.00011710752518939715, |
|
"loss": 0.2813, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 0.00010612064364719856, |
|
"loss": 0.366, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 9.35366467806787e-05, |
|
"loss": 0.357, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 7.989118889325843e-05, |
|
"loss": 0.3169, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 6.576510676653137e-05, |
|
"loss": 0.304, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 5.1759695582141247e-05, |
|
"loss": 0.3576, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 3.847111400507055e-05, |
|
"loss": 0.3524, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 2.6465007909488827e-05, |
|
"loss": 0.2983, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.6252432921493375e-05, |
|
"loss": 0.3261, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 8.268100667234864e-06, |
|
"loss": 0.3443, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 2.85187470336391e-06, |
|
"loss": 0.3208, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 2.343037797424485e-07, |
|
"loss": 0.3161, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 5.787503719329834, |
|
"eval_runtime": 2.3762, |
|
"eval_samples_per_second": 42.925, |
|
"eval_steps_per_second": 5.471, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 5.268082300386967e-07, |
|
"loss": 0.2751, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 3.716937219336464e-06, |
|
"loss": 0.2778, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 9.668898730225944e-06, |
|
"loss": 0.2676, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 1.812933972782302e-05, |
|
"loss": 0.2433, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 2.8738130463426962e-05, |
|
"loss": 0.2658, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 4.1043693868806114e-05, |
|
"loss": 0.2334, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 5.452222752537942e-05, |
|
"loss": 0.2809, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 6.859999999999982e-05, |
|
"loss": 0.283, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 8.267777247462046e-05, |
|
"loss": 0.2846, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 9.615630613119378e-05, |
|
"loss": 0.2926, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 0.00010846186953657294, |
|
"loss": 0.2787, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 0.0001190706602721769, |
|
"loss": 0.2776, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 0.000127531101269774, |
|
"loss": 0.2803, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 0.0001334830627806635, |
|
"loss": 0.3028, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 0.0001366731917699613, |
|
"loss": 0.3338, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 5.904983043670654, |
|
"eval_runtime": 2.3887, |
|
"eval_samples_per_second": 42.702, |
|
"eval_steps_per_second": 5.442, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 0.00013696569622025754, |
|
"loss": 0.3256, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 0.00013434812529663614, |
|
"loss": 0.2792, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 0.0001289318993327653, |
|
"loss": 0.2754, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 0.00012094756707850687, |
|
"loss": 0.3003, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 0.00011073499209051147, |
|
"loss": 0.288, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 9.872888599492977e-05, |
|
"loss": 0.2662, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 8.544030441785888e-05, |
|
"loss": 0.2787, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 7.143489323346877e-05, |
|
"loss": 0.3101, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 5.7308811106741695e-05, |
|
"loss": 0.2716, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 4.366335321932141e-05, |
|
"loss": 0.2699, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 3.1079356352801534e-05, |
|
"loss": 0.3072, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 2.0092474810602934e-05, |
|
"loss": 0.2632, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 1.1170379591190542e-05, |
|
"loss": 0.2761, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 4.692851359682307e-06, |
|
"loss": 0.2779, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 9.35614586573241e-07, |
|
"loss": 0.2747, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 5.883668422698975, |
|
"eval_runtime": 2.3812, |
|
"eval_samples_per_second": 42.835, |
|
"eval_steps_per_second": 5.459, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 5.860097463114244e-08, |
|
"loss": 0.2531, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"learning_rate": 2.099141756561944e-06, |
|
"loss": 0.2619, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 6.970378642210319e-06, |
|
"loss": 0.2333, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 1.4464961055407432e-05, |
|
"loss": 0.2043, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 2.426387228273959e-05, |
|
"loss": 0.2374, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 3.5950008837656816e-05, |
|
"loss": 0.228, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 4.90259350165111e-05, |
|
"loss": 0.2088, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 6.293505690059785e-05, |
|
"loss": 0.2676, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 7.708531450507698e-05, |
|
"loss": 0.2139, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 9.087438358744116e-05, |
|
"loss": 0.2626, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 0.00010371531436754642, |
|
"loss": 0.2233, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 0.00011506151581352575, |
|
"loss": 0.262, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 0.00012443002200473525, |
|
"loss": 0.2799, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 0.00013142205020853692, |
|
"loss": 0.224, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 0.00013573997558945648, |
|
"loss": 0.2749, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 0.0001372, |
|
"loss": 0.2684, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 5.94313907623291, |
|
"eval_runtime": 2.3788, |
|
"eval_samples_per_second": 42.88, |
|
"eval_steps_per_second": 5.465, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 0.0001357399755894565, |
|
"loss": 0.2072, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 10.13, |
|
"learning_rate": 0.00013142205020853697, |
|
"loss": 0.2447, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"learning_rate": 0.0001244300220047355, |
|
"loss": 0.2514, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 0.00011506151581352587, |
|
"loss": 0.2337, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 10.33, |
|
"learning_rate": 0.00010371531436754655, |
|
"loss": 0.2403, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 10.39, |
|
"learning_rate": 9.087438358744108e-05, |
|
"loss": 0.2396, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 10.46, |
|
"learning_rate": 7.708531450507739e-05, |
|
"loss": 0.2401, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 10.53, |
|
"learning_rate": 6.293505690059801e-05, |
|
"loss": 0.2772, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 10.59, |
|
"learning_rate": 4.9025935016511264e-05, |
|
"loss": 0.2206, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 3.595000883765696e-05, |
|
"loss": 0.2247, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 10.72, |
|
"learning_rate": 2.4263872282739537e-05, |
|
"loss": 0.2496, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 10.79, |
|
"learning_rate": 1.4464961055407682e-05, |
|
"loss": 0.2492, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 6.970378642210395e-06, |
|
"loss": 0.2406, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 10.92, |
|
"learning_rate": 2.099141756561982e-06, |
|
"loss": 0.2398, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 10.99, |
|
"learning_rate": 5.860097463115006e-08, |
|
"loss": 0.2381, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 5.973106384277344, |
|
"eval_runtime": 2.3796, |
|
"eval_samples_per_second": 42.864, |
|
"eval_steps_per_second": 5.463, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 11.05, |
|
"learning_rate": 9.356145865732563e-07, |
|
"loss": 0.1709, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 4.6928513596821544e-06, |
|
"loss": 0.1884, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 11.18, |
|
"learning_rate": 1.117037959119045e-05, |
|
"loss": 0.2089, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 2.0092474810602812e-05, |
|
"loss": 0.1931, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 11.32, |
|
"learning_rate": 3.1079356352801595e-05, |
|
"loss": 0.1894, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 11.38, |
|
"learning_rate": 4.366335321932103e-05, |
|
"loss": 0.2227, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 11.45, |
|
"learning_rate": 5.7308811106741526e-05, |
|
"loss": 0.1859, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 11.51, |
|
"learning_rate": 7.143489323346859e-05, |
|
"loss": 0.2015, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 11.58, |
|
"learning_rate": 8.544030441785894e-05, |
|
"loss": 0.194, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 11.64, |
|
"learning_rate": 9.872888599492985e-05, |
|
"loss": 0.2035, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 11.71, |
|
"learning_rate": 0.00011073499209051113, |
|
"loss": 0.2189, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 11.78, |
|
"learning_rate": 0.00012094756707850674, |
|
"loss": 0.2264, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 11.84, |
|
"learning_rate": 0.00012893189933276523, |
|
"loss": 0.2029, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 11.91, |
|
"learning_rate": 0.00013434812529663609, |
|
"loss": 0.2352, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 11.97, |
|
"learning_rate": 0.00013696569622025754, |
|
"loss": 0.2269, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 6.038574695587158, |
|
"eval_runtime": 2.3927, |
|
"eval_samples_per_second": 42.629, |
|
"eval_steps_per_second": 5.433, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 0.00013667319176996134, |
|
"loss": 0.2264, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 12.11, |
|
"learning_rate": 0.0001334830627806636, |
|
"loss": 0.1967, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 12.17, |
|
"learning_rate": 0.0001275311012697742, |
|
"loss": 0.2084, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 12.24, |
|
"learning_rate": 0.00011907066027217684, |
|
"loss": 0.214, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 12.3, |
|
"learning_rate": 0.00010846186953657327, |
|
"loss": 0.22, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 12.37, |
|
"learning_rate": 9.615630613119393e-05, |
|
"loss": 0.198, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 12.43, |
|
"learning_rate": 8.267777247462063e-05, |
|
"loss": 0.23, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 6.859999999999999e-05, |
|
"loss": 0.2105, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 5.452222752537934e-05, |
|
"loss": 0.2301, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 12.63, |
|
"learning_rate": 4.104369386880649e-05, |
|
"loss": 0.2116, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 12.7, |
|
"learning_rate": 2.8738130463427098e-05, |
|
"loss": 0.2089, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 12.76, |
|
"learning_rate": 1.8129339727823136e-05, |
|
"loss": 0.2142, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 12.83, |
|
"learning_rate": 9.668898730225785e-06, |
|
"loss": 0.2095, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 12.89, |
|
"learning_rate": 3.7169372193365172e-06, |
|
"loss": 0.2194, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 12.96, |
|
"learning_rate": 5.268082300387119e-07, |
|
"loss": 0.1903, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 6.056723594665527, |
|
"eval_runtime": 2.3764, |
|
"eval_samples_per_second": 42.923, |
|
"eval_steps_per_second": 5.471, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 2.343037797424561e-07, |
|
"loss": 0.1768, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 2.851874703363925e-06, |
|
"loss": 0.1757, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 13.16, |
|
"learning_rate": 8.268100667234788e-06, |
|
"loss": 0.1981, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 13.22, |
|
"learning_rate": 1.625243292149296e-05, |
|
"loss": 0.1515, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 13.29, |
|
"learning_rate": 2.6465007909488506e-05, |
|
"loss": 0.1837, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 13.36, |
|
"learning_rate": 3.847111400507018e-05, |
|
"loss": 0.1692, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 13.42, |
|
"learning_rate": 5.1759695582141084e-05, |
|
"loss": 0.1552, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 13.49, |
|
"learning_rate": 6.576510676653144e-05, |
|
"loss": 0.1596, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 13.55, |
|
"learning_rate": 7.989118889325803e-05, |
|
"loss": 0.1919, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 13.62, |
|
"learning_rate": 9.353664678067856e-05, |
|
"loss": 0.1759, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 13.68, |
|
"learning_rate": 0.00010612064364719844, |
|
"loss": 0.1885, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 13.75, |
|
"learning_rate": 0.00011710752518939722, |
|
"loss": 0.204, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 13.82, |
|
"learning_rate": 0.0001260296204088093, |
|
"loss": 0.196, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 13.88, |
|
"learning_rate": 0.0001325071486403177, |
|
"loss": 0.1722, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 13.95, |
|
"learning_rate": 0.00013626438541342674, |
|
"loss": 0.2164, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 6.018917083740234, |
|
"eval_runtime": 2.3894, |
|
"eval_samples_per_second": 42.688, |
|
"eval_steps_per_second": 5.441, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.1849, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 14.08, |
|
"learning_rate": 0.00013510085824343813, |
|
"loss": 0.1751, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 14.14, |
|
"learning_rate": 0.0001302296213577898, |
|
"loss": 0.185, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 14.21, |
|
"learning_rate": 0.0001227350389445926, |
|
"loss": 0.1819, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 14.28, |
|
"learning_rate": 0.00011293612771726044, |
|
"loss": 0.1962, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 14.34, |
|
"learning_rate": 0.00010124999116234302, |
|
"loss": 0.1865, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 14.41, |
|
"learning_rate": 8.81740649834887e-05, |
|
"loss": 0.188, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 14.47, |
|
"learning_rate": 7.426494309940244e-05, |
|
"loss": 0.2, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 14.54, |
|
"learning_rate": 6.011468549492306e-05, |
|
"loss": 0.1769, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 14.61, |
|
"learning_rate": 4.632561641255889e-05, |
|
"loss": 0.1978, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 14.67, |
|
"learning_rate": 3.3484685632453414e-05, |
|
"loss": 0.1777, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 14.74, |
|
"learning_rate": 2.2138484186474457e-05, |
|
"loss": 0.2118, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 14.8, |
|
"learning_rate": 1.2769977995264789e-05, |
|
"loss": 0.182, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 14.87, |
|
"learning_rate": 5.777949791463116e-06, |
|
"loss": 0.1877, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 14.93, |
|
"learning_rate": 1.4600244105434828e-06, |
|
"loss": 0.1746, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.1651, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 6.139342784881592, |
|
"eval_runtime": 2.3739, |
|
"eval_samples_per_second": 42.966, |
|
"eval_steps_per_second": 5.476, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 15.07, |
|
"learning_rate": 1.460024410543437e-06, |
|
"loss": 0.1611, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 15.13, |
|
"learning_rate": 5.777949791463017e-06, |
|
"loss": 0.1537, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 15.2, |
|
"learning_rate": 1.2769977995264636e-05, |
|
"loss": 0.1527, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 15.26, |
|
"learning_rate": 2.2138484186474274e-05, |
|
"loss": 0.1497, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 15.33, |
|
"learning_rate": 3.34846856324532e-05, |
|
"loss": 0.1454, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 15.39, |
|
"learning_rate": 4.6325616412558645e-05, |
|
"loss": 0.1285, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 15.46, |
|
"learning_rate": 6.0114685494922816e-05, |
|
"loss": 0.1492, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 15.53, |
|
"learning_rate": 7.42649430994022e-05, |
|
"loss": 0.1309, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 15.59, |
|
"learning_rate": 8.817406498348845e-05, |
|
"loss": 0.1646, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 15.66, |
|
"learning_rate": 0.00010124999116234279, |
|
"loss": 0.1558, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 15.72, |
|
"learning_rate": 0.00011293612771726026, |
|
"loss": 0.1872, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 15.79, |
|
"learning_rate": 0.00012273503894459243, |
|
"loss": 0.1446, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 15.86, |
|
"learning_rate": 0.0001302296213577897, |
|
"loss": 0.1797, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 15.92, |
|
"learning_rate": 0.00013510085824343807, |
|
"loss": 0.1899, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 15.99, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.1733, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 6.194965362548828, |
|
"eval_runtime": 2.3865, |
|
"eval_samples_per_second": 42.741, |
|
"eval_steps_per_second": 5.447, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 16.05, |
|
"learning_rate": 0.00013626438541342671, |
|
"loss": 0.1955, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 16.12, |
|
"learning_rate": 0.00013250714864031777, |
|
"loss": 0.1686, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 16.18, |
|
"learning_rate": 0.00012602962040880945, |
|
"loss": 0.1755, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 16.25, |
|
"learning_rate": 0.00011710752518939739, |
|
"loss": 0.1859, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 16.32, |
|
"learning_rate": 0.00010612064364719905, |
|
"loss": 0.1609, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 16.38, |
|
"learning_rate": 9.353664678067879e-05, |
|
"loss": 0.1691, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 16.45, |
|
"learning_rate": 7.989118889325875e-05, |
|
"loss": 0.1567, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 16.51, |
|
"learning_rate": 6.57651067665317e-05, |
|
"loss": 0.1615, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 16.58, |
|
"learning_rate": 5.17596955821418e-05, |
|
"loss": 0.1546, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 16.64, |
|
"learning_rate": 3.847111400507041e-05, |
|
"loss": 0.161, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 16.71, |
|
"learning_rate": 2.6465007909489085e-05, |
|
"loss": 0.1455, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 16.78, |
|
"learning_rate": 1.6252432921493118e-05, |
|
"loss": 0.1649, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 16.84, |
|
"learning_rate": 8.268100667234903e-06, |
|
"loss": 0.1703, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 16.91, |
|
"learning_rate": 2.851874703363864e-06, |
|
"loss": 0.1799, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 16.97, |
|
"learning_rate": 2.3430377974247896e-07, |
|
"loss": 0.1598, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 6.190419673919678, |
|
"eval_runtime": 2.3796, |
|
"eval_samples_per_second": 42.864, |
|
"eval_steps_per_second": 5.463, |
|
"step": 1292 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 5.268082300387423e-07, |
|
"loss": 0.1591, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 17.11, |
|
"learning_rate": 3.7169372193364334e-06, |
|
"loss": 0.1468, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 17.17, |
|
"learning_rate": 9.668898730225899e-06, |
|
"loss": 0.1398, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 17.24, |
|
"learning_rate": 1.8129339727822966e-05, |
|
"loss": 0.12, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 17.3, |
|
"learning_rate": 2.8738130463427288e-05, |
|
"loss": 0.1334, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 17.37, |
|
"learning_rate": 4.104369386880626e-05, |
|
"loss": 0.1346, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 17.43, |
|
"learning_rate": 5.452222752537957e-05, |
|
"loss": 0.144, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 6.859999999999973e-05, |
|
"loss": 0.1241, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 17.57, |
|
"learning_rate": 8.26777724746199e-05, |
|
"loss": 0.1374, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 17.63, |
|
"learning_rate": 9.61563061311937e-05, |
|
"loss": 0.1446, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 17.7, |
|
"learning_rate": 0.00010846186953657267, |
|
"loss": 0.1361, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 17.76, |
|
"learning_rate": 0.00011907066027217668, |
|
"loss": 0.1447, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 17.83, |
|
"learning_rate": 0.00012753110126977383, |
|
"loss": 0.1544, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 17.89, |
|
"learning_rate": 0.00013348306278066356, |
|
"loss": 0.1502, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 17.96, |
|
"learning_rate": 0.00013667319176996126, |
|
"loss": 0.1565, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 6.231907844543457, |
|
"eval_runtime": 2.391, |
|
"eval_samples_per_second": 42.66, |
|
"eval_steps_per_second": 5.437, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 18.03, |
|
"learning_rate": 0.00013696569622025757, |
|
"loss": 0.1556, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 18.09, |
|
"learning_rate": 0.0001343481252966363, |
|
"loss": 0.1436, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 18.16, |
|
"learning_rate": 0.00012893189933276512, |
|
"loss": 0.1651, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 18.22, |
|
"learning_rate": 0.00012094756707850692, |
|
"loss": 0.1512, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 18.29, |
|
"learning_rate": 0.00011073499209051094, |
|
"loss": 0.1571, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 18.36, |
|
"learning_rate": 9.872888599493006e-05, |
|
"loss": 0.1541, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 18.42, |
|
"learning_rate": 8.544030441785871e-05, |
|
"loss": 0.1489, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 18.49, |
|
"learning_rate": 7.143489323346883e-05, |
|
"loss": 0.1501, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 18.55, |
|
"learning_rate": 5.73088111067413e-05, |
|
"loss": 0.1598, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 18.62, |
|
"learning_rate": 4.366335321932126e-05, |
|
"loss": 0.1461, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 18.68, |
|
"learning_rate": 3.10793563528014e-05, |
|
"loss": 0.1573, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 18.75, |
|
"learning_rate": 2.0092474810602995e-05, |
|
"loss": 0.1337, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 18.82, |
|
"learning_rate": 1.1170379591190853e-05, |
|
"loss": 0.1445, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 18.88, |
|
"learning_rate": 4.692851359682246e-06, |
|
"loss": 0.1426, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 18.95, |
|
"learning_rate": 9.356145865732944e-07, |
|
"loss": 0.1351, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 6.250270843505859, |
|
"eval_runtime": 2.3794, |
|
"eval_samples_per_second": 42.868, |
|
"eval_steps_per_second": 5.464, |
|
"step": 1444 |
|
}, |
|
{ |
|
"epoch": 19.01, |
|
"learning_rate": 5.860097463114244e-08, |
|
"loss": 0.1946, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 19.08, |
|
"learning_rate": 2.099141756561799e-06, |
|
"loss": 0.1179, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 19.14, |
|
"learning_rate": 6.970378642210281e-06, |
|
"loss": 0.1372, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 19.21, |
|
"learning_rate": 1.4464961055407225e-05, |
|
"loss": 0.1174, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 19.28, |
|
"learning_rate": 2.426387228273934e-05, |
|
"loss": 0.1142, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 19.34, |
|
"learning_rate": 3.5950008837656314e-05, |
|
"loss": 0.1221, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 19.41, |
|
"learning_rate": 4.902593501651148e-05, |
|
"loss": 0.11, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 19.47, |
|
"learning_rate": 6.293505690059777e-05, |
|
"loss": 0.1134, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 19.54, |
|
"learning_rate": 7.708531450507714e-05, |
|
"loss": 0.1229, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 19.61, |
|
"learning_rate": 9.087438358744086e-05, |
|
"loss": 0.1244, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 19.67, |
|
"learning_rate": 0.00010371531436754677, |
|
"loss": 0.1288, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 19.74, |
|
"learning_rate": 0.0001150615158135257, |
|
"loss": 0.1384, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 19.8, |
|
"learning_rate": 0.00012443002200473535, |
|
"loss": 0.1294, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 19.87, |
|
"learning_rate": 0.00013142205020853678, |
|
"loss": 0.1234, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 19.93, |
|
"learning_rate": 0.00013573997558945656, |
|
"loss": 0.1569, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 0.0001372, |
|
"loss": 0.1293, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 6.285115718841553, |
|
"eval_runtime": 2.3913, |
|
"eval_samples_per_second": 42.655, |
|
"eval_steps_per_second": 5.436, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 20.07, |
|
"learning_rate": 0.00013573997558945661, |
|
"loss": 0.1331, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 20.13, |
|
"learning_rate": 0.00013142205020853692, |
|
"loss": 0.1233, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 20.2, |
|
"learning_rate": 0.00012443002200473552, |
|
"loss": 0.1277, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 20.26, |
|
"learning_rate": 0.00011506151581352594, |
|
"loss": 0.1432, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 20.33, |
|
"learning_rate": 0.00010371531436754704, |
|
"loss": 0.1511, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 20.39, |
|
"learning_rate": 9.087438358744162e-05, |
|
"loss": 0.1448, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 20.46, |
|
"learning_rate": 7.708531450507796e-05, |
|
"loss": 0.1498, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 20.53, |
|
"learning_rate": 6.29350569005981e-05, |
|
"loss": 0.1375, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 20.59, |
|
"learning_rate": 4.9025935016511806e-05, |
|
"loss": 0.1355, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 20.66, |
|
"learning_rate": 3.5950008837656606e-05, |
|
"loss": 0.132, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 20.72, |
|
"learning_rate": 2.4263872282739598e-05, |
|
"loss": 0.1243, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 20.79, |
|
"learning_rate": 1.4464961055407135e-05, |
|
"loss": 0.1311, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 20.86, |
|
"learning_rate": 6.97037864221022e-06, |
|
"loss": 0.1217, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 20.92, |
|
"learning_rate": 2.099141756562005e-06, |
|
"loss": 0.1347, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 20.99, |
|
"learning_rate": 5.860097463117291e-08, |
|
"loss": 0.1239, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 6.295395374298096, |
|
"eval_runtime": 2.3772, |
|
"eval_samples_per_second": 42.908, |
|
"eval_steps_per_second": 5.469, |
|
"step": 1596 |
|
}, |
|
{ |
|
"epoch": 21.05, |
|
"learning_rate": 9.35614586573241e-07, |
|
"loss": 0.1241, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 21.12, |
|
"learning_rate": 4.692851359682124e-06, |
|
"loss": 0.1257, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 21.18, |
|
"learning_rate": 1.117037959119067e-05, |
|
"loss": 0.1089, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 21.25, |
|
"learning_rate": 2.009247481060276e-05, |
|
"loss": 0.1013, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 21.32, |
|
"learning_rate": 3.107935635280112e-05, |
|
"loss": 0.1159, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 21.38, |
|
"learning_rate": 4.3663353219321405e-05, |
|
"loss": 0.1031, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 21.45, |
|
"learning_rate": 5.7308811106741445e-05, |
|
"loss": 0.1039, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 21.51, |
|
"learning_rate": 7.143489323346802e-05, |
|
"loss": 0.1228, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 21.58, |
|
"learning_rate": 8.544030441785793e-05, |
|
"loss": 0.1095, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 21.64, |
|
"learning_rate": 9.872888599492933e-05, |
|
"loss": 0.1061, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 21.71, |
|
"learning_rate": 0.00011073499209051068, |
|
"loss": 0.1078, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 21.78, |
|
"learning_rate": 0.0001209475670785067, |
|
"loss": 0.1246, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 21.84, |
|
"learning_rate": 0.00012893189933276496, |
|
"loss": 0.1134, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 21.91, |
|
"learning_rate": 0.0001343481252966362, |
|
"loss": 0.1251, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 21.97, |
|
"learning_rate": 0.00013696569622025754, |
|
"loss": 0.1355, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 6.330929756164551, |
|
"eval_runtime": 2.3944, |
|
"eval_samples_per_second": 42.599, |
|
"eval_steps_per_second": 5.429, |
|
"step": 1672 |
|
}, |
|
{ |
|
"epoch": 22.04, |
|
"learning_rate": 0.00013667319176996124, |
|
"loss": 0.1321, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 22.11, |
|
"learning_rate": 0.0001334830627806635, |
|
"loss": 0.1258, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 22.17, |
|
"learning_rate": 0.00012753110126977426, |
|
"loss": 0.1376, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 22.24, |
|
"learning_rate": 0.00011907066027217723, |
|
"loss": 0.1364, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 22.3, |
|
"learning_rate": 0.00010846186953657294, |
|
"loss": 0.1255, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 22.37, |
|
"learning_rate": 9.615630613119401e-05, |
|
"loss": 0.1304, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 22.43, |
|
"learning_rate": 8.267777247462023e-05, |
|
"loss": 0.1214, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"learning_rate": 6.860000000000005e-05, |
|
"loss": 0.1299, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 22.57, |
|
"learning_rate": 5.452222752537991e-05, |
|
"loss": 0.1184, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 22.63, |
|
"learning_rate": 4.104369386880612e-05, |
|
"loss": 0.1361, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 22.7, |
|
"learning_rate": 2.8738130463427166e-05, |
|
"loss": 0.1135, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 22.76, |
|
"learning_rate": 1.8129339727823525e-05, |
|
"loss": 0.1117, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 22.83, |
|
"learning_rate": 9.668898730226326e-06, |
|
"loss": 0.1102, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 22.89, |
|
"learning_rate": 3.7169372193365477e-06, |
|
"loss": 0.1149, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 22.96, |
|
"learning_rate": 5.26808230038788e-07, |
|
"loss": 0.0961, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 6.352334022521973, |
|
"eval_runtime": 2.3763, |
|
"eval_samples_per_second": 42.925, |
|
"eval_steps_per_second": 5.471, |
|
"step": 1748 |
|
}, |
|
{ |
|
"epoch": 23.03, |
|
"learning_rate": 2.343037797424485e-07, |
|
"loss": 0.0998, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 23.09, |
|
"learning_rate": 2.851874703363765e-06, |
|
"loss": 0.0979, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 23.16, |
|
"learning_rate": 8.26810066723498e-06, |
|
"loss": 0.0952, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 23.22, |
|
"learning_rate": 1.6252432921493216e-05, |
|
"loss": 0.103, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 23.29, |
|
"learning_rate": 2.6465007909489207e-05, |
|
"loss": 0.0907, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 23.36, |
|
"learning_rate": 3.847111400506967e-05, |
|
"loss": 0.0932, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 23.42, |
|
"learning_rate": 5.1759695582141e-05, |
|
"loss": 0.0952, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 23.49, |
|
"learning_rate": 6.576510676653087e-05, |
|
"loss": 0.0902, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 23.55, |
|
"learning_rate": 7.989118889325843e-05, |
|
"loss": 0.0934, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 23.62, |
|
"learning_rate": 9.353664678067846e-05, |
|
"loss": 0.1085, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 23.68, |
|
"learning_rate": 0.00010612064364719878, |
|
"loss": 0.1076, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 0.00011710752518939715, |
|
"loss": 0.1122, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 23.82, |
|
"learning_rate": 0.00012602962040880926, |
|
"loss": 0.1105, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 23.88, |
|
"learning_rate": 0.00013250714864031782, |
|
"loss": 0.1049, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 23.95, |
|
"learning_rate": 0.00013626438541342674, |
|
"loss": 0.1191, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 6.427064418792725, |
|
"eval_runtime": 2.3955, |
|
"eval_samples_per_second": 42.579, |
|
"eval_steps_per_second": 5.427, |
|
"step": 1824 |
|
}, |
|
{ |
|
"epoch": 24.01, |
|
"learning_rate": 0.00013714139902536887, |
|
"loss": 0.1256, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 24.08, |
|
"learning_rate": 0.0001351008582434383, |
|
"loss": 0.1109, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 24.14, |
|
"learning_rate": 0.00013022962135778984, |
|
"loss": 0.1139, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 24.21, |
|
"learning_rate": 0.00012273503894459295, |
|
"loss": 0.1074, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 24.28, |
|
"learning_rate": 0.0001129361277172605, |
|
"loss": 0.1146, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 24.34, |
|
"learning_rate": 0.00010124999116234352, |
|
"loss": 0.1217, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 24.41, |
|
"learning_rate": 8.817406498348832e-05, |
|
"loss": 0.1122, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 24.47, |
|
"learning_rate": 7.426494309940203e-05, |
|
"loss": 0.1164, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 24.54, |
|
"learning_rate": 6.011468549492314e-05, |
|
"loss": 0.1218, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 24.61, |
|
"learning_rate": 4.632561641255942e-05, |
|
"loss": 0.1065, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 24.67, |
|
"learning_rate": 3.348468563245349e-05, |
|
"loss": 0.1072, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 24.74, |
|
"learning_rate": 2.2138484186474528e-05, |
|
"loss": 0.1088, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 24.8, |
|
"learning_rate": 1.2769977995264552e-05, |
|
"loss": 0.1169, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 24.87, |
|
"learning_rate": 5.777949791463154e-06, |
|
"loss": 0.1083, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 24.93, |
|
"learning_rate": 1.460024410543399e-06, |
|
"loss": 0.0946, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.1189, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 6.392660617828369, |
|
"eval_runtime": 2.3742, |
|
"eval_samples_per_second": 42.962, |
|
"eval_steps_per_second": 5.476, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 25.07, |
|
"learning_rate": 1.4600244105434142e-06, |
|
"loss": 0.1014, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 25.13, |
|
"learning_rate": 5.777949791463177e-06, |
|
"loss": 0.0827, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 25.2, |
|
"learning_rate": 1.276997799526459e-05, |
|
"loss": 0.092, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 25.26, |
|
"learning_rate": 2.2138484186473857e-05, |
|
"loss": 0.0932, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 25.33, |
|
"learning_rate": 3.34846856324527e-05, |
|
"loss": 0.0958, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 25.39, |
|
"learning_rate": 4.6325616412558564e-05, |
|
"loss": 0.0902, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 25.46, |
|
"learning_rate": 6.011468549492225e-05, |
|
"loss": 0.0864, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 25.53, |
|
"learning_rate": 7.42649430994021e-05, |
|
"loss": 0.0765, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 25.59, |
|
"learning_rate": 8.817406498348838e-05, |
|
"loss": 0.0921, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 25.66, |
|
"learning_rate": 0.00010124999116234357, |
|
"loss": 0.0898, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 25.72, |
|
"learning_rate": 0.00011293612771726056, |
|
"loss": 0.0909, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 25.79, |
|
"learning_rate": 0.0001227350389445924, |
|
"loss": 0.0994, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 25.86, |
|
"learning_rate": 0.00013022962135778946, |
|
"loss": 0.1071, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 25.92, |
|
"learning_rate": 0.00013510085824343805, |
|
"loss": 0.1047, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 25.99, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.1126, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 6.413173198699951, |
|
"eval_runtime": 2.384, |
|
"eval_samples_per_second": 42.785, |
|
"eval_steps_per_second": 5.453, |
|
"step": 1976 |
|
}, |
|
{ |
|
"epoch": 26.05, |
|
"learning_rate": 0.00013626438541342674, |
|
"loss": 0.0971, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 26.12, |
|
"learning_rate": 0.0001325071486403178, |
|
"loss": 0.0972, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 26.18, |
|
"learning_rate": 0.00012602962040880923, |
|
"loss": 0.1153, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 26.25, |
|
"learning_rate": 0.0001171075251893971, |
|
"loss": 0.1152, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 26.32, |
|
"learning_rate": 0.00010612064364719871, |
|
"loss": 0.1087, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 26.38, |
|
"learning_rate": 9.353664678067931e-05, |
|
"loss": 0.112, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 26.45, |
|
"learning_rate": 7.989118889325931e-05, |
|
"loss": 0.1073, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 26.51, |
|
"learning_rate": 6.576510676653178e-05, |
|
"loss": 0.103, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 26.58, |
|
"learning_rate": 5.1759695582141884e-05, |
|
"loss": 0.0993, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 26.64, |
|
"learning_rate": 3.8471114005070485e-05, |
|
"loss": 0.1074, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 26.71, |
|
"learning_rate": 2.6465007909489153e-05, |
|
"loss": 0.0895, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 26.78, |
|
"learning_rate": 1.6252432921493172e-05, |
|
"loss": 0.1028, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 26.84, |
|
"learning_rate": 8.268100667234949e-06, |
|
"loss": 0.0976, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 26.91, |
|
"learning_rate": 2.851874703363742e-06, |
|
"loss": 0.0956, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 26.97, |
|
"learning_rate": 2.3430377974244087e-07, |
|
"loss": 0.0928, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 6.435521602630615, |
|
"eval_runtime": 2.3807, |
|
"eval_samples_per_second": 42.844, |
|
"eval_steps_per_second": 5.461, |
|
"step": 2052 |
|
}, |
|
{ |
|
"epoch": 27.04, |
|
"learning_rate": 5.268082300386738e-07, |
|
"loss": 0.0919, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 27.11, |
|
"learning_rate": 3.716937219336251e-06, |
|
"loss": 0.0829, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 27.17, |
|
"learning_rate": 9.668898730225861e-06, |
|
"loss": 0.0813, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 27.24, |
|
"learning_rate": 1.8129339727822905e-05, |
|
"loss": 0.0878, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 27.3, |
|
"learning_rate": 2.873813046342722e-05, |
|
"loss": 0.0754, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 27.37, |
|
"learning_rate": 4.104369386880618e-05, |
|
"loss": 0.0738, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 27.43, |
|
"learning_rate": 5.452222752537997e-05, |
|
"loss": 0.077, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"learning_rate": 6.860000000000014e-05, |
|
"loss": 0.0792, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 27.57, |
|
"learning_rate": 8.267777247462029e-05, |
|
"loss": 0.0824, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 27.63, |
|
"learning_rate": 9.615630613119318e-05, |
|
"loss": 0.08, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 27.7, |
|
"learning_rate": 0.00010846186953657221, |
|
"loss": 0.0833, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 27.76, |
|
"learning_rate": 0.00011907066027217662, |
|
"loss": 0.0927, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 27.83, |
|
"learning_rate": 0.0001275311012697738, |
|
"loss": 0.0996, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 27.89, |
|
"learning_rate": 0.00013348306278066353, |
|
"loss": 0.0852, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 27.96, |
|
"learning_rate": 0.00013667319176996126, |
|
"loss": 0.1059, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 6.499926567077637, |
|
"eval_runtime": 2.3706, |
|
"eval_samples_per_second": 43.027, |
|
"eval_steps_per_second": 5.484, |
|
"step": 2128 |
|
}, |
|
{ |
|
"epoch": 28.03, |
|
"learning_rate": 0.00013696569622025754, |
|
"loss": 0.09, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 28.09, |
|
"learning_rate": 0.00013434812529663617, |
|
"loss": 0.0978, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 28.16, |
|
"learning_rate": 0.0001289318993327654, |
|
"loss": 0.0895, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 28.22, |
|
"learning_rate": 0.0001209475670785073, |
|
"loss": 0.0973, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 28.29, |
|
"learning_rate": 0.0001107349920905114, |
|
"loss": 0.1049, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 28.36, |
|
"learning_rate": 9.872888599493015e-05, |
|
"loss": 0.0981, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 28.42, |
|
"learning_rate": 8.544030441785881e-05, |
|
"loss": 0.0997, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 28.49, |
|
"learning_rate": 7.143489323346893e-05, |
|
"loss": 0.0941, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 28.55, |
|
"learning_rate": 5.730881110674138e-05, |
|
"loss": 0.0977, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 28.62, |
|
"learning_rate": 4.366335321932134e-05, |
|
"loss": 0.0927, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 28.68, |
|
"learning_rate": 3.1079356352801066e-05, |
|
"loss": 0.0978, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 28.75, |
|
"learning_rate": 2.0092474810602707e-05, |
|
"loss": 0.0931, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 28.82, |
|
"learning_rate": 1.1170379591190633e-05, |
|
"loss": 0.0927, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 28.88, |
|
"learning_rate": 4.692851359682452e-06, |
|
"loss": 0.078, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 28.95, |
|
"learning_rate": 9.356145865733934e-07, |
|
"loss": 0.0822, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 6.48267936706543, |
|
"eval_runtime": 2.3919, |
|
"eval_samples_per_second": 42.644, |
|
"eval_steps_per_second": 5.435, |
|
"step": 2204 |
|
}, |
|
{ |
|
"epoch": 29.01, |
|
"learning_rate": 5.8600974631134825e-08, |
|
"loss": 0.0864, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 29.08, |
|
"learning_rate": 2.099141756561784e-06, |
|
"loss": 0.0784, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 29.14, |
|
"learning_rate": 6.97037864221025e-06, |
|
"loss": 0.0736, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 29.21, |
|
"learning_rate": 1.446496105540718e-05, |
|
"loss": 0.0811, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 29.28, |
|
"learning_rate": 2.4263872282739652e-05, |
|
"loss": 0.0735, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 29.34, |
|
"learning_rate": 3.595000883765667e-05, |
|
"loss": 0.0773, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 29.41, |
|
"learning_rate": 4.902593501651094e-05, |
|
"loss": 0.0769, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 29.47, |
|
"learning_rate": 6.29350569005972e-05, |
|
"loss": 0.0734, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 29.54, |
|
"learning_rate": 7.708531450507706e-05, |
|
"loss": 0.0796, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 29.61, |
|
"learning_rate": 9.087438358744077e-05, |
|
"loss": 0.0719, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 29.67, |
|
"learning_rate": 0.00010371531436754669, |
|
"loss": 0.072, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 29.74, |
|
"learning_rate": 0.00011506151581352563, |
|
"loss": 0.0799, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 29.8, |
|
"learning_rate": 0.00012443002200473557, |
|
"loss": 0.0839, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 29.87, |
|
"learning_rate": 0.00013142205020853694, |
|
"loss": 0.0816, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 29.93, |
|
"learning_rate": 0.00013573997558945664, |
|
"loss": 0.0891, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 0.0001372, |
|
"loss": 0.1024, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 6.566282749176025, |
|
"eval_runtime": 2.3662, |
|
"eval_samples_per_second": 43.107, |
|
"eval_steps_per_second": 5.494, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 30.07, |
|
"learning_rate": 0.00013573997558945675, |
|
"loss": 0.0863, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 30.13, |
|
"learning_rate": 0.00013142205020853713, |
|
"loss": 0.0869, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 30.2, |
|
"learning_rate": 0.00012443002200473587, |
|
"loss": 0.0866, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 30.26, |
|
"learning_rate": 0.000115061515813526, |
|
"loss": 0.087, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 30.33, |
|
"learning_rate": 0.00010371531436754714, |
|
"loss": 0.08, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 30.39, |
|
"learning_rate": 9.087438358744125e-05, |
|
"loss": 0.0955, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 30.46, |
|
"learning_rate": 7.708531450507755e-05, |
|
"loss": 0.0862, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 30.53, |
|
"learning_rate": 6.29350569005977e-05, |
|
"loss": 0.0974, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 30.59, |
|
"learning_rate": 4.902593501651142e-05, |
|
"loss": 0.0996, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 30.66, |
|
"learning_rate": 3.5950008837657114e-05, |
|
"loss": 0.082, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 30.72, |
|
"learning_rate": 2.4263872282740035e-05, |
|
"loss": 0.0779, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 30.79, |
|
"learning_rate": 1.4464961055407484e-05, |
|
"loss": 0.0959, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 30.86, |
|
"learning_rate": 6.970378642210471e-06, |
|
"loss": 0.0768, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 30.92, |
|
"learning_rate": 2.0991417565619058e-06, |
|
"loss": 0.0897, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 30.99, |
|
"learning_rate": 5.8600974631157675e-08, |
|
"loss": 0.0862, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 6.527225494384766, |
|
"eval_runtime": 2.3894, |
|
"eval_samples_per_second": 42.689, |
|
"eval_steps_per_second": 5.441, |
|
"step": 2356 |
|
}, |
|
{ |
|
"epoch": 31.05, |
|
"learning_rate": 9.356145865733097e-07, |
|
"loss": 0.0767, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 31.12, |
|
"learning_rate": 4.692851359682269e-06, |
|
"loss": 0.0717, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 31.18, |
|
"learning_rate": 1.1170379591190359e-05, |
|
"loss": 0.0769, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 2.0092474810602348e-05, |
|
"loss": 0.0754, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 31.32, |
|
"learning_rate": 3.107935635280064e-05, |
|
"loss": 0.0707, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 31.38, |
|
"learning_rate": 4.3663353219320876e-05, |
|
"loss": 0.0739, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 31.45, |
|
"learning_rate": 5.730881110674088e-05, |
|
"loss": 0.0702, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 31.51, |
|
"learning_rate": 7.143489323346843e-05, |
|
"loss": 0.0686, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 31.58, |
|
"learning_rate": 8.544030441785831e-05, |
|
"loss": 0.0702, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 31.64, |
|
"learning_rate": 9.872888599492969e-05, |
|
"loss": 0.0638, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 31.71, |
|
"learning_rate": 0.00011073499209051101, |
|
"loss": 0.0761, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 31.78, |
|
"learning_rate": 0.00012094756707850696, |
|
"loss": 0.0735, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 31.84, |
|
"learning_rate": 0.00012893189933276515, |
|
"loss": 0.0779, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 31.91, |
|
"learning_rate": 0.00013434812529663603, |
|
"loss": 0.0811, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 31.97, |
|
"learning_rate": 0.00013696569622025749, |
|
"loss": 0.0808, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 6.547709941864014, |
|
"eval_runtime": 2.3835, |
|
"eval_samples_per_second": 42.795, |
|
"eval_steps_per_second": 5.454, |
|
"step": 2432 |
|
}, |
|
{ |
|
"epoch": 32.04, |
|
"learning_rate": 0.00013667319176996132, |
|
"loss": 0.0711, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 32.11, |
|
"learning_rate": 0.00013348306278066337, |
|
"loss": 0.0734, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 32.17, |
|
"learning_rate": 0.00012753110126977453, |
|
"loss": 0.0796, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 32.24, |
|
"learning_rate": 0.00011907066027217696, |
|
"loss": 0.0897, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 32.3, |
|
"learning_rate": 0.0001084618695365726, |
|
"loss": 0.0848, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 32.37, |
|
"learning_rate": 9.615630613119364e-05, |
|
"loss": 0.0986, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 32.43, |
|
"learning_rate": 8.267777247462079e-05, |
|
"loss": 0.0795, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"learning_rate": 6.860000000000064e-05, |
|
"loss": 0.0877, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 32.57, |
|
"learning_rate": 5.452222752537951e-05, |
|
"loss": 0.0807, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 32.63, |
|
"learning_rate": 4.104369386880754e-05, |
|
"loss": 0.0848, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 32.7, |
|
"learning_rate": 2.8738130463427633e-05, |
|
"loss": 0.079, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 32.76, |
|
"learning_rate": 1.812933972782325e-05, |
|
"loss": 0.0741, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 32.83, |
|
"learning_rate": 9.668898730225618e-06, |
|
"loss": 0.084, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 32.89, |
|
"learning_rate": 3.7169372193367307e-06, |
|
"loss": 0.0808, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 32.96, |
|
"learning_rate": 5.268082300387347e-07, |
|
"loss": 0.0737, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 6.572361469268799, |
|
"eval_runtime": 2.376, |
|
"eval_samples_per_second": 42.929, |
|
"eval_steps_per_second": 5.471, |
|
"step": 2508 |
|
}, |
|
{ |
|
"epoch": 33.03, |
|
"learning_rate": 2.343037797424028e-07, |
|
"loss": 0.0638, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 33.09, |
|
"learning_rate": 2.8518747033638794e-06, |
|
"loss": 0.0645, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 33.16, |
|
"learning_rate": 8.26810066723424e-06, |
|
"loss": 0.0724, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 33.22, |
|
"learning_rate": 1.6252432921492843e-05, |
|
"loss": 0.0662, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 33.29, |
|
"learning_rate": 2.6465007909488756e-05, |
|
"loss": 0.0643, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 33.36, |
|
"learning_rate": 3.847111400507091e-05, |
|
"loss": 0.0656, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 33.42, |
|
"learning_rate": 5.175969558214045e-05, |
|
"loss": 0.0731, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 33.49, |
|
"learning_rate": 6.576510676653128e-05, |
|
"loss": 0.0711, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 33.55, |
|
"learning_rate": 7.989118889325882e-05, |
|
"loss": 0.0606, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 33.62, |
|
"learning_rate": 9.353664678067884e-05, |
|
"loss": 0.0641, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 33.68, |
|
"learning_rate": 0.00010612064364719829, |
|
"loss": 0.0674, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 33.75, |
|
"learning_rate": 0.00011710752518939675, |
|
"loss": 0.0743, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 33.82, |
|
"learning_rate": 0.00012602962040880947, |
|
"loss": 0.0674, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 33.88, |
|
"learning_rate": 0.00013250714864031726, |
|
"loss": 0.074, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 33.95, |
|
"learning_rate": 0.00013626438541342663, |
|
"loss": 0.0803, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 6.624580383300781, |
|
"eval_runtime": 2.3898, |
|
"eval_samples_per_second": 42.681, |
|
"eval_steps_per_second": 5.44, |
|
"step": 2584 |
|
}, |
|
{ |
|
"epoch": 34.01, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.0696, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 34.08, |
|
"learning_rate": 0.00013510085824343794, |
|
"loss": 0.0735, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 34.14, |
|
"learning_rate": 0.00013022962135779008, |
|
"loss": 0.0738, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 34.21, |
|
"learning_rate": 0.0001227350389445927, |
|
"loss": 0.074, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 34.28, |
|
"learning_rate": 0.00011293612771726094, |
|
"loss": 0.079, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 34.34, |
|
"learning_rate": 0.00010124999116234316, |
|
"loss": 0.0713, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 34.41, |
|
"learning_rate": 8.81740649834898e-05, |
|
"loss": 0.0775, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 34.47, |
|
"learning_rate": 7.42649430994026e-05, |
|
"loss": 0.0827, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 34.54, |
|
"learning_rate": 6.011468549492275e-05, |
|
"loss": 0.0746, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 34.61, |
|
"learning_rate": 4.632561641255812e-05, |
|
"loss": 0.0814, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 34.67, |
|
"learning_rate": 3.3484685632453976e-05, |
|
"loss": 0.0829, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 34.74, |
|
"learning_rate": 2.213848418647423e-05, |
|
"loss": 0.0812, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 34.8, |
|
"learning_rate": 1.276997799526488e-05, |
|
"loss": 0.0637, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 34.87, |
|
"learning_rate": 5.777949791462987e-06, |
|
"loss": 0.0834, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 34.93, |
|
"learning_rate": 1.460024410543719e-06, |
|
"loss": 0.0753, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.0667, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_loss": 6.608242034912109, |
|
"eval_runtime": 2.3714, |
|
"eval_samples_per_second": 43.012, |
|
"eval_steps_per_second": 5.482, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 35.07, |
|
"learning_rate": 1.460024410543498e-06, |
|
"loss": 0.0652, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 35.13, |
|
"learning_rate": 5.77794979146256e-06, |
|
"loss": 0.0662, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 35.2, |
|
"learning_rate": 1.2769977995264255e-05, |
|
"loss": 0.0526, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 35.26, |
|
"learning_rate": 2.2138484186474152e-05, |
|
"loss": 0.0655, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 35.33, |
|
"learning_rate": 3.3484685632453895e-05, |
|
"loss": 0.0666, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 35.39, |
|
"learning_rate": 4.632561641255803e-05, |
|
"loss": 0.0599, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 35.46, |
|
"learning_rate": 6.0114685494922647e-05, |
|
"loss": 0.0552, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 35.53, |
|
"learning_rate": 7.426494309940153e-05, |
|
"loss": 0.0632, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 35.59, |
|
"learning_rate": 8.817406498348876e-05, |
|
"loss": 0.064, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 35.66, |
|
"learning_rate": 0.00010124999116234222, |
|
"loss": 0.0597, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 35.72, |
|
"learning_rate": 0.00011293612771726013, |
|
"loss": 0.0753, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 35.79, |
|
"learning_rate": 0.00012273503894459265, |
|
"loss": 0.0719, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 35.86, |
|
"learning_rate": 0.00013022962135779005, |
|
"loss": 0.0701, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 35.92, |
|
"learning_rate": 0.0001351008582434379, |
|
"loss": 0.0655, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 35.99, |
|
"learning_rate": 0.00013714139902536884, |
|
"loss": 0.086, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_loss": 6.644789695739746, |
|
"eval_runtime": 2.3906, |
|
"eval_samples_per_second": 42.667, |
|
"eval_steps_per_second": 5.438, |
|
"step": 2736 |
|
}, |
|
{ |
|
"epoch": 36.05, |
|
"learning_rate": 0.00013626438541342682, |
|
"loss": 0.0695, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 36.12, |
|
"learning_rate": 0.00013250714864031766, |
|
"loss": 0.0606, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 36.18, |
|
"learning_rate": 0.00012602962040881007, |
|
"loss": 0.0733, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 36.25, |
|
"learning_rate": 0.00011710752518939751, |
|
"loss": 0.0775, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 36.32, |
|
"learning_rate": 0.00010612064364719839, |
|
"loss": 0.0782, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 36.38, |
|
"learning_rate": 9.353664678067986e-05, |
|
"loss": 0.0732, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 36.45, |
|
"learning_rate": 7.989118889325892e-05, |
|
"loss": 0.076, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 36.51, |
|
"learning_rate": 6.576510676653137e-05, |
|
"loss": 0.0736, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 36.58, |
|
"learning_rate": 5.175969558214054e-05, |
|
"loss": 0.0693, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 36.64, |
|
"learning_rate": 3.847111400507101e-05, |
|
"loss": 0.0771, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 36.71, |
|
"learning_rate": 2.6465007909489603e-05, |
|
"loss": 0.0622, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 36.78, |
|
"learning_rate": 1.6252432921493545e-05, |
|
"loss": 0.0668, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 36.84, |
|
"learning_rate": 8.26810066723475e-06, |
|
"loss": 0.0638, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 36.91, |
|
"learning_rate": 2.851874703364184e-06, |
|
"loss": 0.0742, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 36.97, |
|
"learning_rate": 2.3430377974248658e-07, |
|
"loss": 0.0651, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_loss": 6.649299144744873, |
|
"eval_runtime": 2.3884, |
|
"eval_samples_per_second": 42.707, |
|
"eval_steps_per_second": 5.443, |
|
"step": 2812 |
|
}, |
|
{ |
|
"epoch": 37.04, |
|
"learning_rate": 5.268082300387194e-07, |
|
"loss": 0.067, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 37.11, |
|
"learning_rate": 3.7169372193367e-06, |
|
"loss": 0.0551, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 37.17, |
|
"learning_rate": 9.668898730225563e-06, |
|
"loss": 0.0525, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 37.24, |
|
"learning_rate": 1.812933972782318e-05, |
|
"loss": 0.0596, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 37.3, |
|
"learning_rate": 2.8738130463426756e-05, |
|
"loss": 0.063, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 37.37, |
|
"learning_rate": 4.104369386880655e-05, |
|
"loss": 0.0566, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 37.43, |
|
"learning_rate": 5.452222752537846e-05, |
|
"loss": 0.0547, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"learning_rate": 6.859999999999957e-05, |
|
"loss": 0.0582, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 37.57, |
|
"learning_rate": 8.26777724746207e-05, |
|
"loss": 0.0583, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 37.63, |
|
"learning_rate": 9.615630613119265e-05, |
|
"loss": 0.0561, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 37.7, |
|
"learning_rate": 0.00010846186953657255, |
|
"loss": 0.0592, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 37.76, |
|
"learning_rate": 0.0001190706602721769, |
|
"loss": 0.0549, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 37.83, |
|
"learning_rate": 0.000127531101269774, |
|
"loss": 0.0686, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 37.89, |
|
"learning_rate": 0.00013348306278066334, |
|
"loss": 0.0682, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 37.96, |
|
"learning_rate": 0.00013667319176996116, |
|
"loss": 0.0665, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_loss": 6.660224437713623, |
|
"eval_runtime": 2.3808, |
|
"eval_samples_per_second": 42.844, |
|
"eval_steps_per_second": 5.46, |
|
"step": 2888 |
|
}, |
|
{ |
|
"epoch": 38.03, |
|
"learning_rate": 0.0001369656962202576, |
|
"loss": 0.0664, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 38.09, |
|
"learning_rate": 0.00013434812529663606, |
|
"loss": 0.0631, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 38.16, |
|
"learning_rate": 0.00012893189933276566, |
|
"loss": 0.0616, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 38.22, |
|
"learning_rate": 0.00012094756707850702, |
|
"loss": 0.0671, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 38.29, |
|
"learning_rate": 0.00011073499209051109, |
|
"loss": 0.0692, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 38.36, |
|
"learning_rate": 9.87288859949289e-05, |
|
"loss": 0.0676, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 38.42, |
|
"learning_rate": 8.544030441785936e-05, |
|
"loss": 0.0729, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 38.49, |
|
"learning_rate": 7.143489323346852e-05, |
|
"loss": 0.0826, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 38.55, |
|
"learning_rate": 5.730881110674194e-05, |
|
"loss": 0.0716, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 38.62, |
|
"learning_rate": 4.366335321932097e-05, |
|
"loss": 0.0631, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 38.68, |
|
"learning_rate": 3.107935635280236e-05, |
|
"loss": 0.0674, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 38.75, |
|
"learning_rate": 2.009247481060311e-05, |
|
"loss": 0.0674, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 38.82, |
|
"learning_rate": 1.1170379591190413e-05, |
|
"loss": 0.0611, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 38.88, |
|
"learning_rate": 4.692851359682665e-06, |
|
"loss": 0.068, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 38.95, |
|
"learning_rate": 9.356145865733248e-07, |
|
"loss": 0.0651, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_loss": 6.686155319213867, |
|
"eval_runtime": 2.3897, |
|
"eval_samples_per_second": 42.683, |
|
"eval_steps_per_second": 5.44, |
|
"step": 2964 |
|
}, |
|
{ |
|
"epoch": 39.01, |
|
"learning_rate": 5.8600974631157675e-08, |
|
"loss": 0.0541, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 39.08, |
|
"learning_rate": 2.099141756561883e-06, |
|
"loss": 0.0574, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 39.14, |
|
"learning_rate": 6.970378642209999e-06, |
|
"loss": 0.057, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 39.21, |
|
"learning_rate": 1.446496105540683e-05, |
|
"loss": 0.0571, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 39.28, |
|
"learning_rate": 2.4263872282739218e-05, |
|
"loss": 0.0495, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 39.34, |
|
"learning_rate": 3.595000883765702e-05, |
|
"loss": 0.0575, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 39.41, |
|
"learning_rate": 4.902593501651039e-05, |
|
"loss": 0.0554, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 39.47, |
|
"learning_rate": 6.29350569005976e-05, |
|
"loss": 0.0578, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 39.54, |
|
"learning_rate": 7.708531450507746e-05, |
|
"loss": 0.0509, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 39.61, |
|
"learning_rate": 9.087438358744115e-05, |
|
"loss": 0.0458, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 39.67, |
|
"learning_rate": 0.00010371531436754619, |
|
"loss": 0.05, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 39.74, |
|
"learning_rate": 0.00011506151581352521, |
|
"loss": 0.0553, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 39.8, |
|
"learning_rate": 0.00012443002200473525, |
|
"loss": 0.0575, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 39.87, |
|
"learning_rate": 0.00013142205020853708, |
|
"loss": 0.0554, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 39.93, |
|
"learning_rate": 0.00013573997558945632, |
|
"loss": 0.063, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 0.0001372, |
|
"loss": 0.0733, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 6.705848693847656, |
|
"eval_runtime": 2.382, |
|
"eval_samples_per_second": 42.821, |
|
"eval_steps_per_second": 5.458, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 40.07, |
|
"learning_rate": 0.00013573997558945645, |
|
"loss": 0.0557, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 40.13, |
|
"learning_rate": 0.00013142205020853738, |
|
"loss": 0.0663, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 40.2, |
|
"learning_rate": 0.00012443002200473563, |
|
"loss": 0.0627, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 40.26, |
|
"learning_rate": 0.0001150615158135257, |
|
"loss": 0.0658, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 40.33, |
|
"learning_rate": 0.00010371531436754677, |
|
"loss": 0.0668, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 40.39, |
|
"learning_rate": 9.087438358744178e-05, |
|
"loss": 0.0801, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 40.46, |
|
"learning_rate": 7.708531450507813e-05, |
|
"loss": 0.0727, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 40.53, |
|
"learning_rate": 6.293505690059827e-05, |
|
"loss": 0.0678, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 40.59, |
|
"learning_rate": 4.902593501651104e-05, |
|
"loss": 0.0593, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 40.66, |
|
"learning_rate": 3.5950008837657616e-05, |
|
"loss": 0.0635, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 40.72, |
|
"learning_rate": 2.426387228273973e-05, |
|
"loss": 0.0556, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 40.79, |
|
"learning_rate": 1.4464961055407835e-05, |
|
"loss": 0.0694, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 40.86, |
|
"learning_rate": 6.970378642209862e-06, |
|
"loss": 0.0597, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 40.92, |
|
"learning_rate": 2.0991417565622865e-06, |
|
"loss": 0.0636, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 40.99, |
|
"learning_rate": 5.860097463114244e-08, |
|
"loss": 0.0535, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_loss": 6.696343421936035, |
|
"eval_runtime": 2.3795, |
|
"eval_samples_per_second": 42.866, |
|
"eval_steps_per_second": 5.463, |
|
"step": 3116 |
|
}, |
|
{ |
|
"epoch": 41.05, |
|
"learning_rate": 9.356145865732106e-07, |
|
"loss": 0.0552, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 41.12, |
|
"learning_rate": 4.692851359682063e-06, |
|
"loss": 0.0537, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 41.18, |
|
"learning_rate": 1.1170379591190047e-05, |
|
"loss": 0.0546, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 41.25, |
|
"learning_rate": 2.0092474810601945e-05, |
|
"loss": 0.051, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 41.32, |
|
"learning_rate": 3.10793563528018e-05, |
|
"loss": 0.0501, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 41.38, |
|
"learning_rate": 4.366335321931943e-05, |
|
"loss": 0.0524, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 41.45, |
|
"learning_rate": 5.730881110674128e-05, |
|
"loss": 0.0483, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 41.51, |
|
"learning_rate": 7.143489323346784e-05, |
|
"loss": 0.0532, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 41.58, |
|
"learning_rate": 8.544030441785965e-05, |
|
"loss": 0.0542, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 41.64, |
|
"learning_rate": 9.87288859949283e-05, |
|
"loss": 0.0503, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 41.71, |
|
"learning_rate": 0.00011073499209051133, |
|
"loss": 0.0561, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 41.78, |
|
"learning_rate": 0.0001209475670785066, |
|
"loss": 0.0526, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 41.84, |
|
"learning_rate": 0.00012893189933276488, |
|
"loss": 0.0548, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 41.91, |
|
"learning_rate": 0.00013434812529663587, |
|
"loss": 0.0662, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 41.97, |
|
"learning_rate": 0.00013696569622025746, |
|
"loss": 0.0633, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_loss": 6.702749252319336, |
|
"eval_runtime": 2.3956, |
|
"eval_samples_per_second": 42.577, |
|
"eval_steps_per_second": 5.427, |
|
"step": 3192 |
|
}, |
|
{ |
|
"epoch": 42.04, |
|
"learning_rate": 0.00013667319176996126, |
|
"loss": 0.0521, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 42.11, |
|
"learning_rate": 0.00013348306278066356, |
|
"loss": 0.0563, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 42.17, |
|
"learning_rate": 0.00012753110126977434, |
|
"loss": 0.0619, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 42.24, |
|
"learning_rate": 0.00011907066027217736, |
|
"loss": 0.0651, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 42.3, |
|
"learning_rate": 0.00010846186953657228, |
|
"loss": 0.0628, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 42.37, |
|
"learning_rate": 9.615630613119328e-05, |
|
"loss": 0.0659, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 42.43, |
|
"learning_rate": 8.26777724746204e-05, |
|
"loss": 0.0655, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 6.860000000000023e-05, |
|
"loss": 0.0623, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 42.57, |
|
"learning_rate": 5.4522227525380065e-05, |
|
"loss": 0.064, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 42.63, |
|
"learning_rate": 4.1043693868807165e-05, |
|
"loss": 0.0627, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 42.7, |
|
"learning_rate": 2.8738130463428097e-05, |
|
"loss": 0.057, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 42.76, |
|
"learning_rate": 1.8129339727822976e-05, |
|
"loss": 0.056, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 42.83, |
|
"learning_rate": 9.668898730225914e-06, |
|
"loss": 0.0651, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 42.89, |
|
"learning_rate": 3.716937219336601e-06, |
|
"loss": 0.0567, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 42.96, |
|
"learning_rate": 5.268082300388033e-07, |
|
"loss": 0.0803, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_loss": 6.75062370300293, |
|
"eval_runtime": 2.3741, |
|
"eval_samples_per_second": 42.963, |
|
"eval_steps_per_second": 5.476, |
|
"step": 3268 |
|
}, |
|
{ |
|
"epoch": 43.03, |
|
"learning_rate": 2.343037797423571e-07, |
|
"loss": 0.0462, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 43.09, |
|
"learning_rate": 2.8518747033639933e-06, |
|
"loss": 0.0484, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 43.16, |
|
"learning_rate": 8.268100667233966e-06, |
|
"loss": 0.0574, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 43.22, |
|
"learning_rate": 1.625243292149311e-05, |
|
"loss": 0.0424, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 43.29, |
|
"learning_rate": 2.64650079094883e-05, |
|
"loss": 0.0483, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 43.36, |
|
"learning_rate": 3.847111400507128e-05, |
|
"loss": 0.0512, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 43.42, |
|
"learning_rate": 5.17596955821399e-05, |
|
"loss": 0.0528, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 43.49, |
|
"learning_rate": 6.576510676653168e-05, |
|
"loss": 0.0431, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 43.55, |
|
"learning_rate": 7.989118889325826e-05, |
|
"loss": 0.0531, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 43.62, |
|
"learning_rate": 9.353664678067831e-05, |
|
"loss": 0.0506, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 43.68, |
|
"learning_rate": 0.0001061206436471978, |
|
"loss": 0.0495, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 43.75, |
|
"learning_rate": 0.00011710752518939633, |
|
"loss": 0.059, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 43.82, |
|
"learning_rate": 0.0001260296204088097, |
|
"loss": 0.0589, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 43.88, |
|
"learning_rate": 0.00013250714864031707, |
|
"loss": 0.0594, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 43.95, |
|
"learning_rate": 0.00013626438541342671, |
|
"loss": 0.061, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_loss": 6.78729248046875, |
|
"eval_runtime": 2.3812, |
|
"eval_samples_per_second": 42.835, |
|
"eval_steps_per_second": 5.459, |
|
"step": 3344 |
|
}, |
|
{ |
|
"epoch": 44.01, |
|
"learning_rate": 0.00013714139902536887, |
|
"loss": 0.0634, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 44.08, |
|
"learning_rate": 0.00013510085824343783, |
|
"loss": 0.0513, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 44.14, |
|
"learning_rate": 0.00013022962135779035, |
|
"loss": 0.0597, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 44.21, |
|
"learning_rate": 0.00012273503894459246, |
|
"loss": 0.0558, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 44.28, |
|
"learning_rate": 0.00011293612771726063, |
|
"loss": 0.0677, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 44.34, |
|
"learning_rate": 0.00010124999116234365, |
|
"loss": 0.0662, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 44.41, |
|
"learning_rate": 8.817406498348941e-05, |
|
"loss": 0.0628, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 44.47, |
|
"learning_rate": 7.426494309940318e-05, |
|
"loss": 0.0616, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 44.54, |
|
"learning_rate": 6.011468549492234e-05, |
|
"loss": 0.0614, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 44.61, |
|
"learning_rate": 4.632561641255866e-05, |
|
"loss": 0.0577, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 44.67, |
|
"learning_rate": 3.348468563245363e-05, |
|
"loss": 0.0629, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 44.74, |
|
"learning_rate": 2.2138484186474647e-05, |
|
"loss": 0.0543, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 44.8, |
|
"learning_rate": 1.2769977995265214e-05, |
|
"loss": 0.0467, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 44.87, |
|
"learning_rate": 5.7779497914628265e-06, |
|
"loss": 0.0532, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 44.93, |
|
"learning_rate": 1.4600244105438407e-06, |
|
"loss": 0.0566, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.0512, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_loss": 6.75970458984375, |
|
"eval_runtime": 2.3797, |
|
"eval_samples_per_second": 42.863, |
|
"eval_steps_per_second": 5.463, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 45.07, |
|
"learning_rate": 1.4600244105433837e-06, |
|
"loss": 0.0475, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 45.13, |
|
"learning_rate": 5.77794979146272e-06, |
|
"loss": 0.0525, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 45.2, |
|
"learning_rate": 1.2769977995263928e-05, |
|
"loss": 0.0459, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 45.26, |
|
"learning_rate": 2.213848418647445e-05, |
|
"loss": 0.0464, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 45.33, |
|
"learning_rate": 3.348468563245339e-05, |
|
"loss": 0.0515, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 45.39, |
|
"learning_rate": 4.632561641255841e-05, |
|
"loss": 0.0465, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 45.46, |
|
"learning_rate": 6.011468549492208e-05, |
|
"loss": 0.0493, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 45.53, |
|
"learning_rate": 7.426494309940098e-05, |
|
"loss": 0.0489, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 45.59, |
|
"learning_rate": 8.817406498348916e-05, |
|
"loss": 0.0471, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 45.66, |
|
"learning_rate": 0.0001012499911623417, |
|
"loss": 0.0401, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 45.72, |
|
"learning_rate": 0.00011293612771726042, |
|
"loss": 0.0492, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 45.79, |
|
"learning_rate": 0.0001227350389445923, |
|
"loss": 0.0478, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 45.86, |
|
"learning_rate": 0.00013022962135779022, |
|
"loss": 0.0487, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 45.92, |
|
"learning_rate": 0.00013510085824343775, |
|
"loss": 0.0522, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 45.99, |
|
"learning_rate": 0.00013714139902536887, |
|
"loss": 0.0539, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_loss": 6.77425479888916, |
|
"eval_runtime": 2.3831, |
|
"eval_samples_per_second": 42.802, |
|
"eval_steps_per_second": 5.455, |
|
"step": 3496 |
|
}, |
|
{ |
|
"epoch": 46.05, |
|
"learning_rate": 0.00013626438541342674, |
|
"loss": 0.0588, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 46.12, |
|
"learning_rate": 0.00013250714864031785, |
|
"loss": 0.0596, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 46.18, |
|
"learning_rate": 0.00012602962040880985, |
|
"loss": 0.0537, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 46.25, |
|
"learning_rate": 0.00011710752518939791, |
|
"loss": 0.0534, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 46.32, |
|
"learning_rate": 0.00010612064364719805, |
|
"loss": 0.0602, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 46.38, |
|
"learning_rate": 9.353664678068038e-05, |
|
"loss": 0.0561, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 46.45, |
|
"learning_rate": 7.989118889325852e-05, |
|
"loss": 0.0507, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 46.51, |
|
"learning_rate": 6.576510676653195e-05, |
|
"loss": 0.0508, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 46.58, |
|
"learning_rate": 5.1759695582140156e-05, |
|
"loss": 0.0514, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 46.64, |
|
"learning_rate": 3.8471114005071515e-05, |
|
"loss": 0.0513, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 46.71, |
|
"learning_rate": 2.646500790949005e-05, |
|
"loss": 0.0556, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 46.78, |
|
"learning_rate": 1.6252432921493277e-05, |
|
"loss": 0.0532, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 46.84, |
|
"learning_rate": 8.268100667235025e-06, |
|
"loss": 0.0507, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 46.91, |
|
"learning_rate": 2.8518747033640696e-06, |
|
"loss": 0.052, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 46.97, |
|
"learning_rate": 2.343037797425399e-07, |
|
"loss": 0.0503, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_loss": 6.793975353240967, |
|
"eval_runtime": 2.3872, |
|
"eval_samples_per_second": 42.727, |
|
"eval_steps_per_second": 5.446, |
|
"step": 3572 |
|
}, |
|
{ |
|
"epoch": 47.04, |
|
"learning_rate": 5.268082300387728e-07, |
|
"loss": 0.0538, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 47.11, |
|
"learning_rate": 3.7169372193365096e-06, |
|
"loss": 0.0408, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 47.17, |
|
"learning_rate": 9.668898730225777e-06, |
|
"loss": 0.053, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 47.24, |
|
"learning_rate": 1.8129339727822793e-05, |
|
"loss": 0.047, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 47.3, |
|
"learning_rate": 2.873813046342629e-05, |
|
"loss": 0.0432, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 47.37, |
|
"learning_rate": 4.104369386880692e-05, |
|
"loss": 0.0432, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 47.43, |
|
"learning_rate": 5.4522227525377896e-05, |
|
"loss": 0.0438, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"learning_rate": 6.859999999999997e-05, |
|
"loss": 0.0433, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 47.57, |
|
"learning_rate": 8.267777247462013e-05, |
|
"loss": 0.0453, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 47.63, |
|
"learning_rate": 9.615630613119303e-05, |
|
"loss": 0.0466, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 47.7, |
|
"learning_rate": 0.00010846186953657208, |
|
"loss": 0.0471, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 47.76, |
|
"learning_rate": 0.00011907066027217717, |
|
"loss": 0.0438, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 47.83, |
|
"learning_rate": 0.00012753110126977418, |
|
"loss": 0.0497, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 47.89, |
|
"learning_rate": 0.00013348306278066348, |
|
"loss": 0.049, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 47.96, |
|
"learning_rate": 0.0001366731917699612, |
|
"loss": 0.0489, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_loss": 6.83149528503418, |
|
"eval_runtime": 2.3845, |
|
"eval_samples_per_second": 42.777, |
|
"eval_steps_per_second": 5.452, |
|
"step": 3648 |
|
}, |
|
{ |
|
"epoch": 48.03, |
|
"learning_rate": 0.00013696569622025762, |
|
"loss": 0.043, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 48.09, |
|
"learning_rate": 0.00013434812529663595, |
|
"loss": 0.0509, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 48.16, |
|
"learning_rate": 0.00012893189933276593, |
|
"loss": 0.0521, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 48.22, |
|
"learning_rate": 0.00012094756707850676, |
|
"loss": 0.051, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 48.29, |
|
"learning_rate": 0.00011073499209051154, |
|
"loss": 0.0583, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 48.36, |
|
"learning_rate": 9.872888599492854e-05, |
|
"loss": 0.0547, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 48.42, |
|
"learning_rate": 8.54403044178599e-05, |
|
"loss": 0.0645, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 48.49, |
|
"learning_rate": 7.143489323346813e-05, |
|
"loss": 0.0598, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 48.55, |
|
"learning_rate": 5.7308811106741546e-05, |
|
"loss": 0.0511, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 48.62, |
|
"learning_rate": 4.366335321932151e-05, |
|
"loss": 0.051, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 48.68, |
|
"learning_rate": 3.107935635280202e-05, |
|
"loss": 0.0494, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 48.75, |
|
"learning_rate": 2.0092474810603514e-05, |
|
"loss": 0.047, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 48.82, |
|
"learning_rate": 1.1170379591190191e-05, |
|
"loss": 0.0484, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 48.88, |
|
"learning_rate": 4.69285135968287e-06, |
|
"loss": 0.0499, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 48.95, |
|
"learning_rate": 9.356145865732563e-07, |
|
"loss": 0.0526, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_loss": 6.790609836578369, |
|
"eval_runtime": 2.386, |
|
"eval_samples_per_second": 42.749, |
|
"eval_steps_per_second": 5.448, |
|
"step": 3724 |
|
}, |
|
{ |
|
"epoch": 49.01, |
|
"learning_rate": 5.8600974631134825e-08, |
|
"loss": 0.0424, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 49.08, |
|
"learning_rate": 2.0991417565617385e-06, |
|
"loss": 0.0426, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 49.14, |
|
"learning_rate": 6.970378642209748e-06, |
|
"loss": 0.0428, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 49.21, |
|
"learning_rate": 1.446496105540648e-05, |
|
"loss": 0.0409, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 49.28, |
|
"learning_rate": 2.4263872282739523e-05, |
|
"loss": 0.0452, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 49.34, |
|
"learning_rate": 3.595000883765652e-05, |
|
"loss": 0.0406, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 49.41, |
|
"learning_rate": 4.902593501651078e-05, |
|
"loss": 0.0424, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 49.47, |
|
"learning_rate": 6.293505690059702e-05, |
|
"loss": 0.0368, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 49.54, |
|
"learning_rate": 7.708531450507785e-05, |
|
"loss": 0.0432, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 49.61, |
|
"learning_rate": 9.087438358744153e-05, |
|
"loss": 0.0429, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 49.67, |
|
"learning_rate": 0.00010371531436754655, |
|
"loss": 0.0429, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 49.74, |
|
"learning_rate": 0.00011506151581352551, |
|
"loss": 0.0473, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 49.8, |
|
"learning_rate": 0.0001244300220047349, |
|
"loss": 0.041, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 49.87, |
|
"learning_rate": 0.00013142205020853727, |
|
"loss": 0.051, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 49.93, |
|
"learning_rate": 0.0001357399755894562, |
|
"loss": 0.0513, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 0.0001372, |
|
"loss": 0.053, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_loss": 6.893409252166748, |
|
"eval_runtime": 2.3737, |
|
"eval_samples_per_second": 42.971, |
|
"eval_steps_per_second": 5.477, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 50.73, |
|
"learning_rate": 2.2697640403783063e-05, |
|
"loss": 0.9066, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 50.8, |
|
"learning_rate": 1.3101434185879145e-05, |
|
"loss": 0.6969, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 50.87, |
|
"learning_rate": 5.930781605717916e-06, |
|
"loss": 1.0504, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 50.93, |
|
"learning_rate": 1.4990745896610897e-06, |
|
"loss": 1.4716, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"learning_rate": 0.0, |
|
"loss": 1.2765, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"eval_loss": 1.2292253971099854, |
|
"eval_runtime": 2.6668, |
|
"eval_samples_per_second": 39.373, |
|
"eval_steps_per_second": 5.25, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 51.07, |
|
"learning_rate": 1.499074589660808e-06, |
|
"loss": 1.2797, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 51.13, |
|
"learning_rate": 5.9307816057173676e-06, |
|
"loss": 1.9281, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 51.2, |
|
"learning_rate": 1.3101434185878354e-05, |
|
"loss": 1.4825, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 51.27, |
|
"learning_rate": 2.2697640403782067e-05, |
|
"loss": 1.4677, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 51.33, |
|
"learning_rate": 3.429999999999976e-05, |
|
"loss": 1.0241, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 51.4, |
|
"learning_rate": 4.740143418587843e-05, |
|
"loss": 1.0786, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 51.47, |
|
"learning_rate": 6.142934741983887e-05, |
|
"loss": 1.4638, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 51.53, |
|
"learning_rate": 7.5770652580161e-05, |
|
"loss": 1.0569, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 51.6, |
|
"learning_rate": 8.979856581412144e-05, |
|
"loss": 1.0274, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 51.67, |
|
"learning_rate": 0.00010289999999999844, |
|
"loss": 0.9631, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 51.73, |
|
"learning_rate": 0.00011450235959621783, |
|
"loss": 0.7978, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 51.8, |
|
"learning_rate": 0.00012409856581412041, |
|
"loss": 0.838, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 51.87, |
|
"learning_rate": 0.00013126921839428258, |
|
"loss": 0.6216, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 51.93, |
|
"learning_rate": 0.00013570092541033876, |
|
"loss": 1.2543, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"learning_rate": 0.0001372, |
|
"loss": 1.0127, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"eval_loss": 0.8823016285896301, |
|
"eval_runtime": 2.4876, |
|
"eval_samples_per_second": 42.209, |
|
"eval_steps_per_second": 5.628, |
|
"step": 3900 |
|
} |
|
], |
|
"max_steps": 5250, |
|
"num_train_epochs": 70, |
|
"total_flos": 4029907009536000.0, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|