{ "best_metric": 10.136499555844708, "best_model_checkpoint": "./whisper-large-v3-ta/checkpoint-500", "epoch": 0.014250292130988685, "eval_steps": 500, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0007125146065494343, "grad_norm": 3.9452075958251953, "learning_rate": 5e-06, "loss": 0.5658, "step": 25 }, { "epoch": 0.0014250292130988686, "grad_norm": 6.297654151916504, "learning_rate": 1e-05, "loss": 0.2991, "step": 50 }, { "epoch": 0.002137543819648303, "grad_norm": 5.006739139556885, "learning_rate": 1e-05, "loss": 0.3457, "step": 75 }, { "epoch": 0.002850058426197737, "grad_norm": 3.2383642196655273, "learning_rate": 1e-05, "loss": 0.209, "step": 100 }, { "epoch": 0.003562573032747171, "grad_norm": 3.5915024280548096, "learning_rate": 1e-05, "loss": 0.2085, "step": 125 }, { "epoch": 0.004275087639296606, "grad_norm": 3.4095346927642822, "learning_rate": 1e-05, "loss": 0.1911, "step": 150 }, { "epoch": 0.0049876022458460395, "grad_norm": 2.908203363418579, "learning_rate": 1e-05, "loss": 0.1881, "step": 175 }, { "epoch": 0.005700116852395474, "grad_norm": 2.7387442588806152, "learning_rate": 1e-05, "loss": 0.1907, "step": 200 }, { "epoch": 0.006412631458944908, "grad_norm": 3.590416669845581, "learning_rate": 1e-05, "loss": 0.1678, "step": 225 }, { "epoch": 0.007125146065494342, "grad_norm": 2.9250645637512207, "learning_rate": 1e-05, "loss": 0.1907, "step": 250 }, { "epoch": 0.007837660672043776, "grad_norm": 3.180162191390991, "learning_rate": 1e-05, "loss": 0.178, "step": 275 }, { "epoch": 0.008550175278593211, "grad_norm": 3.205806255340576, "learning_rate": 1e-05, "loss": 0.1698, "step": 300 }, { "epoch": 0.009262689885142646, "grad_norm": 3.011608123779297, "learning_rate": 1e-05, "loss": 0.177, "step": 325 }, { "epoch": 0.009975204491692079, "grad_norm": 2.0001132488250732, "learning_rate": 1e-05, "loss": 0.1689, "step": 350 }, { "epoch": 0.010687719098241514, "grad_norm": 2.6594409942626953, "learning_rate": 1e-05, "loss": 0.1726, "step": 375 }, { "epoch": 0.011400233704790949, "grad_norm": 2.706954002380371, "learning_rate": 1e-05, "loss": 0.1781, "step": 400 }, { "epoch": 0.012112748311340382, "grad_norm": 2.7729318141937256, "learning_rate": 1e-05, "loss": 0.1596, "step": 425 }, { "epoch": 0.012825262917889817, "grad_norm": 2.474515914916992, "learning_rate": 1e-05, "loss": 0.1608, "step": 450 }, { "epoch": 0.013537777524439252, "grad_norm": 2.822539806365967, "learning_rate": 1e-05, "loss": 0.1775, "step": 475 }, { "epoch": 0.014250292130988685, "grad_norm": 1.9778774976730347, "learning_rate": 1e-05, "loss": 0.1489, "step": 500 }, { "epoch": 0.014250292130988685, "eval_loss": 0.15865246951580048, "eval_runtime": 75329.7882, "eval_samples_per_second": 0.414, "eval_steps_per_second": 0.052, "eval_wer": 10.136499555844708, "eval_wer_ortho": 36.29684933552337, "step": 500 } ], "logging_steps": 25, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.358999322624e+19, "train_batch_size": 8, "trial_name": null, "trial_params": null }