|
{ |
|
"best_metric": 1.387204885482788, |
|
"best_model_checkpoint": "output_dir/checkpoint-7500", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 8420, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.9406175771971496e-05, |
|
"loss": 3.5781, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.8812351543942996e-05, |
|
"loss": 3.0221, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.821852731591449e-05, |
|
"loss": 2.7656, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.7624703087885984e-05, |
|
"loss": 2.6432, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.7030878859857484e-05, |
|
"loss": 2.5316, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"eval_loss": 2.2591474056243896, |
|
"eval_runtime": 34.7222, |
|
"eval_samples_per_second": 129.226, |
|
"eval_steps_per_second": 8.093, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.643705463182898e-05, |
|
"loss": 2.4743, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.584323040380048e-05, |
|
"loss": 2.3371, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.524940617577197e-05, |
|
"loss": 2.3233, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.465558194774347e-05, |
|
"loss": 2.2412, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 4.406175771971497e-05, |
|
"loss": 2.2197, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"eval_loss": 1.979953408241272, |
|
"eval_runtime": 34.803, |
|
"eval_samples_per_second": 128.926, |
|
"eval_steps_per_second": 8.074, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.3467933491686466e-05, |
|
"loss": 2.1312, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 4.287410926365796e-05, |
|
"loss": 2.1109, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.228028503562946e-05, |
|
"loss": 2.0618, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.168646080760095e-05, |
|
"loss": 2.0827, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.109263657957245e-05, |
|
"loss": 2.0108, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"eval_loss": 1.830361008644104, |
|
"eval_runtime": 34.8169, |
|
"eval_samples_per_second": 128.874, |
|
"eval_steps_per_second": 8.071, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.049881235154395e-05, |
|
"loss": 1.9889, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.990498812351544e-05, |
|
"loss": 1.9597, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.9311163895486934e-05, |
|
"loss": 1.9155, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.8717339667458435e-05, |
|
"loss": 1.9138, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.812351543942993e-05, |
|
"loss": 1.855, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"eval_loss": 1.722902536392212, |
|
"eval_runtime": 34.0608, |
|
"eval_samples_per_second": 131.735, |
|
"eval_steps_per_second": 8.25, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.752969121140142e-05, |
|
"loss": 1.8592, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.693586698337292e-05, |
|
"loss": 1.8951, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 3.6342042755344416e-05, |
|
"loss": 1.8651, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.5748218527315916e-05, |
|
"loss": 1.808, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 3.515439429928741e-05, |
|
"loss": 1.7933, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"eval_loss": 1.6560965776443481, |
|
"eval_runtime": 34.0703, |
|
"eval_samples_per_second": 131.698, |
|
"eval_steps_per_second": 8.248, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 3.456057007125891e-05, |
|
"loss": 1.8124, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 3.396674584323041e-05, |
|
"loss": 1.7598, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 3.3372921615201904e-05, |
|
"loss": 1.7716, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 3.27790973871734e-05, |
|
"loss": 1.7557, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 3.21852731591449e-05, |
|
"loss": 1.7466, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"eval_loss": 1.6100554466247559, |
|
"eval_runtime": 34.1759, |
|
"eval_samples_per_second": 131.291, |
|
"eval_steps_per_second": 8.222, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 3.159144893111639e-05, |
|
"loss": 1.745, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 3.0997624703087885e-05, |
|
"loss": 1.729, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 3.0403800475059386e-05, |
|
"loss": 1.7112, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 2.980997624703088e-05, |
|
"loss": 1.6804, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 2.9216152019002373e-05, |
|
"loss": 1.6906, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"eval_loss": 1.5587605237960815, |
|
"eval_runtime": 34.0323, |
|
"eval_samples_per_second": 131.845, |
|
"eval_steps_per_second": 8.257, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 2.8622327790973873e-05, |
|
"loss": 1.6638, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 2.8028503562945367e-05, |
|
"loss": 1.6895, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 2.7434679334916867e-05, |
|
"loss": 1.6983, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 2.6840855106888364e-05, |
|
"loss": 1.6648, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.6247030878859858e-05, |
|
"loss": 1.6301, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"eval_loss": 1.5263967514038086, |
|
"eval_runtime": 34.0548, |
|
"eval_samples_per_second": 131.758, |
|
"eval_steps_per_second": 8.251, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 2.5653206650831358e-05, |
|
"loss": 1.6388, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 2.5059382422802852e-05, |
|
"loss": 1.6336, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 2.446555819477435e-05, |
|
"loss": 1.6427, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 2.3871733966745842e-05, |
|
"loss": 1.6219, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 2.327790973871734e-05, |
|
"loss": 1.6131, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"eval_loss": 1.489254355430603, |
|
"eval_runtime": 34.1712, |
|
"eval_samples_per_second": 131.309, |
|
"eval_steps_per_second": 8.223, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 2.2684085510688836e-05, |
|
"loss": 1.5886, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 2.2090261282660333e-05, |
|
"loss": 1.5906, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 2.149643705463183e-05, |
|
"loss": 1.5717, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 2.0902612826603327e-05, |
|
"loss": 1.5887, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 2.0308788598574824e-05, |
|
"loss": 1.5731, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"eval_loss": 1.4625672101974487, |
|
"eval_runtime": 34.0832, |
|
"eval_samples_per_second": 131.649, |
|
"eval_steps_per_second": 8.245, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 1.9714964370546318e-05, |
|
"loss": 1.5607, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 1.9121140142517815e-05, |
|
"loss": 1.5905, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 1.8527315914489312e-05, |
|
"loss": 1.6074, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 1.793349168646081e-05, |
|
"loss": 1.5401, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 1.7339667458432306e-05, |
|
"loss": 1.5248, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"eval_loss": 1.4445180892944336, |
|
"eval_runtime": 33.993, |
|
"eval_samples_per_second": 131.998, |
|
"eval_steps_per_second": 8.266, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.6745843230403803e-05, |
|
"loss": 1.5611, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 1.61520190023753e-05, |
|
"loss": 1.5312, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 1.5558194774346793e-05, |
|
"loss": 1.5318, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 1.496437054631829e-05, |
|
"loss": 1.5366, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 1.4370546318289787e-05, |
|
"loss": 1.509, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"eval_loss": 1.4273977279663086, |
|
"eval_runtime": 33.9883, |
|
"eval_samples_per_second": 132.016, |
|
"eval_steps_per_second": 8.268, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 1.3776722090261283e-05, |
|
"loss": 1.5425, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 1.318289786223278e-05, |
|
"loss": 1.5253, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 1.2589073634204277e-05, |
|
"loss": 1.5502, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 1.1995249406175772e-05, |
|
"loss": 1.5011, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 1.1401425178147269e-05, |
|
"loss": 1.5083, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"eval_loss": 1.410913348197937, |
|
"eval_runtime": 33.9769, |
|
"eval_samples_per_second": 132.06, |
|
"eval_steps_per_second": 8.27, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 1.0807600950118766e-05, |
|
"loss": 1.5303, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 1.0213776722090261e-05, |
|
"loss": 1.4877, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 9.619952494061758e-06, |
|
"loss": 1.5082, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 9.026128266033253e-06, |
|
"loss": 1.4751, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 8.432304038004752e-06, |
|
"loss": 1.5277, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"eval_loss": 1.4051495790481567, |
|
"eval_runtime": 33.9346, |
|
"eval_samples_per_second": 132.225, |
|
"eval_steps_per_second": 8.281, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 7.838479809976247e-06, |
|
"loss": 1.5207, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 7.244655581947744e-06, |
|
"loss": 1.4948, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 6.6508313539192404e-06, |
|
"loss": 1.4897, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 6.0570071258907366e-06, |
|
"loss": 1.4801, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 5.4631828978622335e-06, |
|
"loss": 1.4884, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"eval_loss": 1.387204885482788, |
|
"eval_runtime": 33.9831, |
|
"eval_samples_per_second": 132.036, |
|
"eval_steps_per_second": 8.269, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 4.86935866983373e-06, |
|
"loss": 1.5037, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 4.275534441805226e-06, |
|
"loss": 1.4884, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 3.681710213776722e-06, |
|
"loss": 1.4968, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 3.0878859857482185e-06, |
|
"loss": 1.4762, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 2.494061757719715e-06, |
|
"loss": 1.496, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"eval_loss": 1.3912383317947388, |
|
"eval_runtime": 33.9905, |
|
"eval_samples_per_second": 132.007, |
|
"eval_steps_per_second": 8.267, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 1.9002375296912116e-06, |
|
"loss": 1.4703, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 1.306413301662708e-06, |
|
"loss": 1.4722, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 7.125890736342043e-07, |
|
"loss": 1.4392, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 1.1876484560570073e-07, |
|
"loss": 1.4692, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 8420, |
|
"total_flos": 1.78429300862976e+16, |
|
"train_loss": 1.7687041851233982, |
|
"train_runtime": 3488.0926, |
|
"train_samples_per_second": 38.588, |
|
"train_steps_per_second": 2.414 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.3867350816726685, |
|
"eval_runtime": 34.1096, |
|
"eval_samples_per_second": 131.547, |
|
"eval_steps_per_second": 8.238, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.3883991241455078, |
|
"eval_runtime": 34.15, |
|
"eval_samples_per_second": 131.391, |
|
"eval_steps_per_second": 8.228, |
|
"step": 8420 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 8420, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 1.78429300862976e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|