|
{ |
|
"best_metric": 0.5536245703697205, |
|
"best_model_checkpoint": "./output/clip-finetuned-csu-p14-336-e3l57-l/checkpoint-11500", |
|
"epoch": 2.1186440677966103, |
|
"eval_steps": 500, |
|
"global_step": 11500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.09211495946941783, |
|
"grad_norm": 3.083087682723999, |
|
"learning_rate": 4.84647506755097e-07, |
|
"loss": 0.2708, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.09211495946941783, |
|
"eval_loss": 1.0631108283996582, |
|
"eval_runtime": 74.1964, |
|
"eval_samples_per_second": 16.268, |
|
"eval_steps_per_second": 2.035, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.18422991893883567, |
|
"grad_norm": 7.726487159729004, |
|
"learning_rate": 4.69295013510194e-07, |
|
"loss": 0.251, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.18422991893883567, |
|
"eval_loss": 0.942773699760437, |
|
"eval_runtime": 74.6047, |
|
"eval_samples_per_second": 16.179, |
|
"eval_steps_per_second": 2.024, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.2763448784082535, |
|
"grad_norm": 0.02718021161854267, |
|
"learning_rate": 4.5394252026529107e-07, |
|
"loss": 0.235, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.2763448784082535, |
|
"eval_loss": 0.8716810345649719, |
|
"eval_runtime": 75.03, |
|
"eval_samples_per_second": 16.087, |
|
"eval_steps_per_second": 2.013, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.36845983787767134, |
|
"grad_norm": 0.0001215290030813776, |
|
"learning_rate": 4.385900270203881e-07, |
|
"loss": 0.1529, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.36845983787767134, |
|
"eval_loss": 0.8318113088607788, |
|
"eval_runtime": 76.6917, |
|
"eval_samples_per_second": 15.738, |
|
"eval_steps_per_second": 1.969, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.46057479734708917, |
|
"grad_norm": 419.545166015625, |
|
"learning_rate": 4.232375337754851e-07, |
|
"loss": 0.1781, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.46057479734708917, |
|
"eval_loss": 0.7549223303794861, |
|
"eval_runtime": 75.2036, |
|
"eval_samples_per_second": 16.05, |
|
"eval_steps_per_second": 2.008, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.552689756816507, |
|
"grad_norm": 368.7484436035156, |
|
"learning_rate": 4.0788504053058217e-07, |
|
"loss": 0.1681, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.552689756816507, |
|
"eval_loss": 0.7217888236045837, |
|
"eval_runtime": 76.0606, |
|
"eval_samples_per_second": 15.869, |
|
"eval_steps_per_second": 1.985, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.6448047162859248, |
|
"grad_norm": 20.03949737548828, |
|
"learning_rate": 3.925325472856792e-07, |
|
"loss": 0.1064, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.6448047162859248, |
|
"eval_loss": 0.7048410177230835, |
|
"eval_runtime": 76.0578, |
|
"eval_samples_per_second": 15.87, |
|
"eval_steps_per_second": 1.985, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.7369196757553427, |
|
"grad_norm": 363.6124267578125, |
|
"learning_rate": 3.771800540407762e-07, |
|
"loss": 0.1357, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.7369196757553427, |
|
"eval_loss": 0.6961681246757507, |
|
"eval_runtime": 76.3574, |
|
"eval_samples_per_second": 15.807, |
|
"eval_steps_per_second": 1.978, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.8290346352247605, |
|
"grad_norm": 494.03570556640625, |
|
"learning_rate": 3.618275607958732e-07, |
|
"loss": 0.1098, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.8290346352247605, |
|
"eval_loss": 0.6778352856636047, |
|
"eval_runtime": 76.4844, |
|
"eval_samples_per_second": 15.781, |
|
"eval_steps_per_second": 1.974, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.9211495946941783, |
|
"grad_norm": 4.2969353671651334e-05, |
|
"learning_rate": 3.464750675509703e-07, |
|
"loss": 0.1142, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.9211495946941783, |
|
"eval_loss": 0.6656659245491028, |
|
"eval_runtime": 76.6475, |
|
"eval_samples_per_second": 15.747, |
|
"eval_steps_per_second": 1.97, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.013264554163596, |
|
"grad_norm": 0.0010161151876673102, |
|
"learning_rate": 3.311225743060673e-07, |
|
"loss": 0.1113, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.013264554163596, |
|
"eval_loss": 0.6430702209472656, |
|
"eval_runtime": 76.4915, |
|
"eval_samples_per_second": 15.78, |
|
"eval_steps_per_second": 1.974, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.105379513633014, |
|
"grad_norm": 0.05515381693840027, |
|
"learning_rate": 3.157700810611643e-07, |
|
"loss": 0.0572, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.105379513633014, |
|
"eval_loss": 0.636746883392334, |
|
"eval_runtime": 76.0169, |
|
"eval_samples_per_second": 15.878, |
|
"eval_steps_per_second": 1.986, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.1974944731024317, |
|
"grad_norm": 0.0038496837951242924, |
|
"learning_rate": 3.0041758781626137e-07, |
|
"loss": 0.0746, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.1974944731024317, |
|
"eval_loss": 0.6261336207389832, |
|
"eval_runtime": 76.1916, |
|
"eval_samples_per_second": 15.842, |
|
"eval_steps_per_second": 1.982, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.2896094325718497, |
|
"grad_norm": 2.594869386030041e-07, |
|
"learning_rate": 2.8506509457135833e-07, |
|
"loss": 0.0494, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.2896094325718497, |
|
"eval_loss": 0.6245251893997192, |
|
"eval_runtime": 75.9712, |
|
"eval_samples_per_second": 15.888, |
|
"eval_steps_per_second": 1.988, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.3817243920412676, |
|
"grad_norm": 25.23366928100586, |
|
"learning_rate": 2.697126013264554e-07, |
|
"loss": 0.0788, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.3817243920412676, |
|
"eval_loss": 0.611993670463562, |
|
"eval_runtime": 75.1885, |
|
"eval_samples_per_second": 16.053, |
|
"eval_steps_per_second": 2.008, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.4738393515106853, |
|
"grad_norm": 0.0002172550739487633, |
|
"learning_rate": 2.5436010808155247e-07, |
|
"loss": 0.0808, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.4738393515106853, |
|
"eval_loss": 0.6011174917221069, |
|
"eval_runtime": 76.2572, |
|
"eval_samples_per_second": 15.828, |
|
"eval_steps_per_second": 1.98, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.565954310980103, |
|
"grad_norm": 0.032347481697797775, |
|
"learning_rate": 2.390076148366495e-07, |
|
"loss": 0.0536, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.565954310980103, |
|
"eval_loss": 0.5893377661705017, |
|
"eval_runtime": 75.429, |
|
"eval_samples_per_second": 16.002, |
|
"eval_steps_per_second": 2.002, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.658069270449521, |
|
"grad_norm": 2.7040863415095373e-07, |
|
"learning_rate": 2.236551215917465e-07, |
|
"loss": 0.0869, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.658069270449521, |
|
"eval_loss": 0.59163498878479, |
|
"eval_runtime": 75.4027, |
|
"eval_samples_per_second": 16.007, |
|
"eval_steps_per_second": 2.003, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.750184229918939, |
|
"grad_norm": 3.4501523166197146e-11, |
|
"learning_rate": 2.083026283468435e-07, |
|
"loss": 0.0752, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.750184229918939, |
|
"eval_loss": 0.5707471370697021, |
|
"eval_runtime": 76.0154, |
|
"eval_samples_per_second": 15.878, |
|
"eval_steps_per_second": 1.986, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.8422991893883567, |
|
"grad_norm": 0.00039183301851153374, |
|
"learning_rate": 1.9295013510194055e-07, |
|
"loss": 0.0577, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.8422991893883567, |
|
"eval_loss": 0.5678250789642334, |
|
"eval_runtime": 75.8084, |
|
"eval_samples_per_second": 15.922, |
|
"eval_steps_per_second": 1.992, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.9344141488577744, |
|
"grad_norm": 1.0438248487787405e-08, |
|
"learning_rate": 1.7759764185703757e-07, |
|
"loss": 0.0891, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.9344141488577744, |
|
"eval_loss": 0.563089907169342, |
|
"eval_runtime": 75.3868, |
|
"eval_samples_per_second": 16.011, |
|
"eval_steps_per_second": 2.003, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 2.026529108327192, |
|
"grad_norm": 1.162303306045942e-06, |
|
"learning_rate": 1.6224514861213458e-07, |
|
"loss": 0.0559, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 2.026529108327192, |
|
"eval_loss": 0.5547806620597839, |
|
"eval_runtime": 76.4427, |
|
"eval_samples_per_second": 15.79, |
|
"eval_steps_per_second": 1.975, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 2.1186440677966103, |
|
"grad_norm": 123.83905029296875, |
|
"learning_rate": 1.4689265536723165e-07, |
|
"loss": 0.0385, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 2.1186440677966103, |
|
"eval_loss": 0.5536245703697205, |
|
"eval_runtime": 75.6886, |
|
"eval_samples_per_second": 15.947, |
|
"eval_steps_per_second": 1.995, |
|
"step": 11500 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 16284, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4137940556192520.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|