Writing logs to ./outputs/2023-02-12-23-30-37-265125/train_log.txt. Wrote original training args to ./outputs/2023-02-12-23-30-37-265125/training_args.json. ***** Running training ***** Num examples = 160000 Num epochs = 3 Num clean epochs = 3 Instantaneous batch size per device = 64 Total train batch size (w. parallel, distributed & accumulation) = 64 Gradient accumulation steps = 1 Total optimization steps = 7500 ========================================================== Epoch 1 Running clean epoch 1/3 Train accuracy: 94.11% Eval accuracy: 96.77% Best score found. Saved model to ./outputs/2023-02-12-23-30-37-265125/best_model/ ========================================================== Epoch 2 Running clean epoch 2/3 Train accuracy: 97.52% Eval accuracy: 96.95% Best score found. Saved model to ./outputs/2023-02-12-23-30-37-265125/best_model/ ========================================================== Epoch 3 Running clean epoch 3/3 Train accuracy: 98.70% Eval accuracy: 97.07% Best score found. Saved model to ./outputs/2023-02-12-23-30-37-265125/best_model/ Wrote README to ./outputs/2023-02-12-23-30-37-265125/README.md.