|
{ |
|
"best_metric": 0.5478764772415161, |
|
"best_model_checkpoint": "./output/clip-finetuned-csu-p14-336-e4l57-l/checkpoint-12000", |
|
"epoch": 1.2798634812286689, |
|
"eval_steps": 500, |
|
"global_step": 12000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05332764505119454, |
|
"grad_norm": 414.2552795410156, |
|
"learning_rate": 4.911120591581342e-07, |
|
"loss": 0.3809, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05332764505119454, |
|
"eval_loss": 1.1109352111816406, |
|
"eval_runtime": 63.6734, |
|
"eval_samples_per_second": 15.501, |
|
"eval_steps_per_second": 1.947, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.10665529010238908, |
|
"grad_norm": 45.880027770996094, |
|
"learning_rate": 4.822241183162685e-07, |
|
"loss": 0.2649, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.10665529010238908, |
|
"eval_loss": 0.9588962197303772, |
|
"eval_runtime": 62.6825, |
|
"eval_samples_per_second": 15.746, |
|
"eval_steps_per_second": 1.978, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.1599829351535836, |
|
"grad_norm": 432.9915466308594, |
|
"learning_rate": 4.733361774744027e-07, |
|
"loss": 0.2104, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.1599829351535836, |
|
"eval_loss": 0.90110844373703, |
|
"eval_runtime": 62.6757, |
|
"eval_samples_per_second": 15.748, |
|
"eval_steps_per_second": 1.978, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.21331058020477817, |
|
"grad_norm": 0.014208819717168808, |
|
"learning_rate": 4.6444823663253695e-07, |
|
"loss": 0.1849, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.21331058020477817, |
|
"eval_loss": 0.8570474982261658, |
|
"eval_runtime": 63.7296, |
|
"eval_samples_per_second": 15.487, |
|
"eval_steps_per_second": 1.946, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.2666382252559727, |
|
"grad_norm": 174.73324584960938, |
|
"learning_rate": 4.5556029579067116e-07, |
|
"loss": 0.2056, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.2666382252559727, |
|
"eval_loss": 0.7895939350128174, |
|
"eval_runtime": 63.8387, |
|
"eval_samples_per_second": 15.461, |
|
"eval_steps_per_second": 1.942, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.3199658703071672, |
|
"grad_norm": 9.852073823424234e-09, |
|
"learning_rate": 4.4667235494880547e-07, |
|
"loss": 0.159, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.3199658703071672, |
|
"eval_loss": 0.7647623419761658, |
|
"eval_runtime": 63.9155, |
|
"eval_samples_per_second": 15.442, |
|
"eval_steps_per_second": 1.94, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.37329351535836175, |
|
"grad_norm": 0.0025215104687958956, |
|
"learning_rate": 4.377844141069397e-07, |
|
"loss": 0.1696, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.37329351535836175, |
|
"eval_loss": 0.7638036012649536, |
|
"eval_runtime": 63.8003, |
|
"eval_samples_per_second": 15.47, |
|
"eval_steps_per_second": 1.944, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.42662116040955633, |
|
"grad_norm": 2.4611830711364746, |
|
"learning_rate": 4.2889647326507393e-07, |
|
"loss": 0.1125, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.42662116040955633, |
|
"eval_loss": 0.7486374378204346, |
|
"eval_runtime": 63.9847, |
|
"eval_samples_per_second": 15.426, |
|
"eval_steps_per_second": 1.938, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.47994880546075086, |
|
"grad_norm": 5.230295658111572, |
|
"learning_rate": 4.2000853242320814e-07, |
|
"loss": 0.1322, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.47994880546075086, |
|
"eval_loss": 0.7251659035682678, |
|
"eval_runtime": 64.4782, |
|
"eval_samples_per_second": 15.308, |
|
"eval_steps_per_second": 1.923, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.5332764505119454, |
|
"grad_norm": 3.821002974291332e-05, |
|
"learning_rate": 4.1112059158134245e-07, |
|
"loss": 0.1633, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.5332764505119454, |
|
"eval_loss": 0.7035414576530457, |
|
"eval_runtime": 62.6186, |
|
"eval_samples_per_second": 15.762, |
|
"eval_steps_per_second": 1.98, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.58660409556314, |
|
"grad_norm": 5.916939244343666e-06, |
|
"learning_rate": 4.0223265073947665e-07, |
|
"loss": 0.125, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.58660409556314, |
|
"eval_loss": 0.6792380809783936, |
|
"eval_runtime": 65.0551, |
|
"eval_samples_per_second": 15.172, |
|
"eval_steps_per_second": 1.906, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.6399317406143344, |
|
"grad_norm": 92.8386459350586, |
|
"learning_rate": 3.933447098976109e-07, |
|
"loss": 0.1297, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.6399317406143344, |
|
"eval_loss": 0.6549482345581055, |
|
"eval_runtime": 64.0827, |
|
"eval_samples_per_second": 15.402, |
|
"eval_steps_per_second": 1.935, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.693259385665529, |
|
"grad_norm": 0.00014717792510055006, |
|
"learning_rate": 3.8445676905574517e-07, |
|
"loss": 0.1114, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.693259385665529, |
|
"eval_loss": 0.6223254799842834, |
|
"eval_runtime": 64.89, |
|
"eval_samples_per_second": 15.21, |
|
"eval_steps_per_second": 1.911, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.7465870307167235, |
|
"grad_norm": 4.789559397977428e-07, |
|
"learning_rate": 3.755688282138794e-07, |
|
"loss": 0.144, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.7465870307167235, |
|
"eval_loss": 0.6204637289047241, |
|
"eval_runtime": 62.7674, |
|
"eval_samples_per_second": 15.725, |
|
"eval_steps_per_second": 1.976, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.7999146757679181, |
|
"grad_norm": 1.7078508138656616, |
|
"learning_rate": 3.6668088737201363e-07, |
|
"loss": 0.1399, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.7999146757679181, |
|
"eval_loss": 0.6074371933937073, |
|
"eval_runtime": 63.4694, |
|
"eval_samples_per_second": 15.551, |
|
"eval_steps_per_second": 1.954, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.8532423208191127, |
|
"grad_norm": 870.1183471679688, |
|
"learning_rate": 3.5779294653014783e-07, |
|
"loss": 0.141, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.8532423208191127, |
|
"eval_loss": 0.5950366854667664, |
|
"eval_runtime": 62.915, |
|
"eval_samples_per_second": 15.688, |
|
"eval_steps_per_second": 1.971, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.9065699658703071, |
|
"grad_norm": 5.781071013188921e-05, |
|
"learning_rate": 3.4890500568828214e-07, |
|
"loss": 0.1291, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.9065699658703071, |
|
"eval_loss": 0.5800224542617798, |
|
"eval_runtime": 62.3336, |
|
"eval_samples_per_second": 15.834, |
|
"eval_steps_per_second": 1.989, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.9598976109215017, |
|
"grad_norm": 1.3669992685317993, |
|
"learning_rate": 3.4001706484641635e-07, |
|
"loss": 0.1246, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.9598976109215017, |
|
"eval_loss": 0.5702486038208008, |
|
"eval_runtime": 64.055, |
|
"eval_samples_per_second": 15.409, |
|
"eval_steps_per_second": 1.936, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.0132252559726962, |
|
"grad_norm": 644.68115234375, |
|
"learning_rate": 3.311291240045506e-07, |
|
"loss": 0.1257, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.0132252559726962, |
|
"eval_loss": 0.5807780623435974, |
|
"eval_runtime": 62.5312, |
|
"eval_samples_per_second": 15.784, |
|
"eval_steps_per_second": 1.983, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.0665529010238908, |
|
"grad_norm": 27.854568481445312, |
|
"learning_rate": 3.2224118316268486e-07, |
|
"loss": 0.0336, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.0665529010238908, |
|
"eval_loss": 0.5713164210319519, |
|
"eval_runtime": 62.5622, |
|
"eval_samples_per_second": 15.776, |
|
"eval_steps_per_second": 1.982, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.1198805460750854, |
|
"grad_norm": 0.00045933027286082506, |
|
"learning_rate": 3.133532423208191e-07, |
|
"loss": 0.0663, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.1198805460750854, |
|
"eval_loss": 0.5601483583450317, |
|
"eval_runtime": 63.4675, |
|
"eval_samples_per_second": 15.551, |
|
"eval_steps_per_second": 1.954, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.17320819112628, |
|
"grad_norm": 0.14330387115478516, |
|
"learning_rate": 3.044653014789533e-07, |
|
"loss": 0.078, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.17320819112628, |
|
"eval_loss": 0.5580261945724487, |
|
"eval_runtime": 63.3927, |
|
"eval_samples_per_second": 15.57, |
|
"eval_steps_per_second": 1.956, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.2265358361774745, |
|
"grad_norm": 4.066005518388316e-11, |
|
"learning_rate": 2.955773606370876e-07, |
|
"loss": 0.1118, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.2265358361774745, |
|
"eval_loss": 0.5491370558738708, |
|
"eval_runtime": 63.7293, |
|
"eval_samples_per_second": 15.487, |
|
"eval_steps_per_second": 1.946, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.2798634812286689, |
|
"grad_norm": 7.519358769059181e-05, |
|
"learning_rate": 2.8668941979522184e-07, |
|
"loss": 0.0534, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.2798634812286689, |
|
"eval_loss": 0.5478764772415161, |
|
"eval_runtime": 62.7547, |
|
"eval_samples_per_second": 15.728, |
|
"eval_steps_per_second": 1.976, |
|
"step": 12000 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 28128, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4317819731822520.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|