Quentin Meeus
finetune small model on NER task (slu_weight=1.)
cfee215
raw
history blame
49.8 kB
/usr/data/condor/execute/dir_32133/whisper_slu/scripts/run_ner_trainer_peft.py:143: UserWarning: DeprecationWarning: use_peft is always True
warnings.warn("DeprecationWarning: use_peft is always True")
Process rank: 0, device: cuda:0, n_gpu: 1distributed training: True, 16-bits training: True
/users/spraak/qmeeus/micromamba/envs/torch-cu121/lib/python3.10/site-packages/transformers/configuration_utils.py:508: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
warnings.warn(
[INFO|configuration_utils.py:737] 2024-01-04 16:15:59,208 >> loading configuration file configs/whisper_small_ner.json
[WARNING|configuration_utils.py:617] 2024-01-04 16:15:59,208 >> You are using a model of type whisper to instantiate a model of type whisper_for_slu. This is not supported for all configurations of models and can yield errors.
[INFO|configuration_utils.py:802] 2024-01-04 16:15:59,210 >> Model config WhisperSLUConfig {
"_name_or_path": "openai/whisper-small",
"activation_dropout": 0.0,
"activation_function": "gelu",
"adaptor_activation": "relu",
"adaptor_init": "constant",
"adaptor_layernorm": true,
"apply_spec_augment": false,
"architectures": [
"WhisperForConditionalGeneration"
],
"attention_dropout": 0.0,
"begin_suppress_tokens": [
220,
50257
],
"bos_token_id": 50257,
"classifier_proj_size": 256,
"crf_transition_matrix": null,
"d_model": 768,
"decoder_attention_heads": 12,
"decoder_ffn_dim": 3072,
"decoder_layerdrop": 0.0,
"decoder_layers": 12,
"decoder_start_token_id": 50258,
"dropout": 0.0,
"encoder_attention_heads": 12,
"encoder_ffn_dim": 3072,
"encoder_layerdrop": 0.0,
"encoder_layers": 12,
"eos_token_id": 50257,
"forced_decoder_ids": [
[
1,
50259
],
[
2,
50359
],
[
3,
50363
]
],
"init_std": 0.02,
"is_encoder_decoder": true,
"mask_feature_length": 10,
"mask_feature_min_masks": 0,
"mask_feature_prob": 0.0,
"mask_time_length": 10,
"mask_time_min_masks": 2,
"mask_time_prob": 0.05,
"max_length": 448,
"max_source_positions": 1500,
"max_target_positions": 448,
"median_filter_width": 7,
"model_type": "whisper_for_slu",
"num_hidden_layers": 12,
"num_mel_bins": 80,
"pad_token_id": 50257,
"scale_embedding": false,
"slu_attention_heads": 12,
"slu_dropout": 0.3,
"slu_embed_dim": 768,
"slu_ffn_dim": 2048,
"slu_focus": 1.0,
"slu_input_from": "decoder",
"slu_input_layers": [
11
],
"slu_labels": null,
"slu_layers": 2,
"slu_max_positions": null,
"slu_output_dim": 37,
"slu_pad_token_id": 1,
"slu_start_token_id": 36,
"slu_task": "named_entity_recognition",
"slu_weight": 1.0,
"suppress_tokens": [
1,
2,
7,
8,
9,
10,
14,
25,
26,
27,
28,
29,
31,
58,
59,
60,
61,
62,
63,
90,
91,
92,
93,
359,
503,
522,
542,
873,
893,
902,
918,
922,
931,
1350,
1853,
1982,
2460,
2627,
3246,
3253,
3268,
3536,
3846,
3961,
4183,
4667,
6585,
6647,
7273,
9061,
9383,
10428,
10929,
11938,
12033,
12331,
12562,
13793,
14157,
14635,
15265,
15618,
16553,
16604,
18362,
18956,
20075,
21675,
22520,
26130,
26161,
26435,
28279,
29464,
31650,
32302,
32470,
36865,
42863,
47425,
49870,
50254,
50258,
50360,
50361,
50362
],
"task": "token_classification",
"teacher": null,
"torch_dtype": "float32",
"transformers_version": "4.37.0.dev0",
"use_cache": true,
"use_crf": false,
"use_weighted_layer_sum": false,
"vocab_size": 51865
}
/users/spraak/qmeeus/micromamba/envs/torch-cu121/lib/python3.10/site-packages/transformers/models/auto/feature_extraction_auto.py:328: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
warnings.warn(
[INFO|feature_extraction_utils.py:535] 2024-01-04 16:15:59,215 >> loading configuration file /esat/audioslave/qmeeus/exp/whisper_slu/train/whisper-small-spoken-ner/preprocessor_config.json
[INFO|feature_extraction_utils.py:579] 2024-01-04 16:15:59,220 >> Feature extractor WhisperFeatureExtractor {
"chunk_length": 30,
"feature_extractor_type": "WhisperFeatureExtractor",
"feature_size": 80,
"hop_length": 160,
"n_fft": 400,
"n_samples": 480000,
"nb_max_frames": 3000,
"padding_side": "right",
"padding_value": 0.0,
"processor_class": "WhisperProcessor",
"return_attention_mask": false,
"sampling_rate": 16000
}
/users/spraak/qmeeus/micromamba/envs/torch-cu121/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:691: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
warnings.warn(
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,234 >> loading file vocab.json
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,234 >> loading file tokenizer.json
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,234 >> loading file merges.txt
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,234 >> loading file normalizer.json
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,234 >> loading file added_tokens.json
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,234 >> loading file special_tokens_map.json
[INFO|tokenization_utils_base.py:2024] 2024-01-04 16:15:59,235 >> loading file tokenizer_config.json
[WARNING|logging.py:314] 2024-01-04 16:15:59,977 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
/users/spraak/qmeeus/micromamba/envs/torch-cu121/lib/python3.10/site-packages/transformers/modeling_utils.py:2790: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
warnings.warn(
[INFO|modeling_utils.py:3373] 2024-01-04 16:15:59,982 >> loading weights file /esat/audioslave/qmeeus/exp/whisper_slu/train/whisper-small-spoken-ner/model.safetensors
[INFO|configuration_utils.py:826] 2024-01-04 16:16:01,058 >> Generate config GenerationConfig {
"begin_suppress_tokens": [
220,
50257
],
"bos_token_id": 50257,
"decoder_start_token_id": 50258,
"eos_token_id": 50257,
"forced_decoder_ids": [
[
1,
50259
],
[
2,
50359
],
[
3,
50363
]
],
"max_length": 448,
"pad_token_id": 50257
}
[INFO|modeling_utils.py:4227] 2024-01-04 16:16:01,653 >> All model checkpoint weights were used when initializing WhisperSLU.
[INFO|modeling_utils.py:4235] 2024-01-04 16:16:01,653 >> All the weights of WhisperSLU were initialized from the model checkpoint at /esat/audioslave/qmeeus/exp/whisper_slu/train/whisper-small-spoken-ner.
If your task is similar to the task the model of the checkpoint was trained on, you can already use WhisperSLU for predictions without further training.
[INFO|configuration_utils.py:779] 2024-01-04 16:16:01,656 >> loading configuration file /esat/audioslave/qmeeus/exp/whisper_slu/train/whisper-small-spoken-ner/generation_config.json
[INFO|configuration_utils.py:826] 2024-01-04 16:16:01,657 >> Generate config GenerationConfig {
"alignment_heads": [
[
5,
3
],
[
5,
9
],
[
8,
0
],
[
8,
4
],
[
8,
7
],
[
8,
8
],
[
9,
0
],
[
9,
7
],
[
9,
9
],
[
10,
5
]
],
"begin_suppress_tokens": [
220,
50257
],
"bos_token_id": 50257,
"decoder_start_token_id": 50258,
"eos_token_id": 50257,
"forced_decoder_ids": [
[
1,
null
],
[
2,
50359
]
],
"is_multilingual": true,
"lang_to_id": {
"<|af|>": 50327,
"<|am|>": 50334,
"<|ar|>": 50272,
"<|as|>": 50350,
"<|az|>": 50304,
"<|ba|>": 50355,
"<|be|>": 50330,
"<|bg|>": 50292,
"<|bn|>": 50302,
"<|bo|>": 50347,
"<|br|>": 50309,
"<|bs|>": 50315,
"<|ca|>": 50270,
"<|cs|>": 50283,
"<|cy|>": 50297,
"<|da|>": 50285,
"<|de|>": 50261,
"<|el|>": 50281,
"<|en|>": 50259,
"<|es|>": 50262,
"<|et|>": 50307,
"<|eu|>": 50310,
"<|fa|>": 50300,
"<|fi|>": 50277,
"<|fo|>": 50338,
"<|fr|>": 50265,
"<|gl|>": 50319,
"<|gu|>": 50333,
"<|haw|>": 50352,
"<|ha|>": 50354,
"<|he|>": 50279,
"<|hi|>": 50276,
"<|hr|>": 50291,
"<|ht|>": 50339,
"<|hu|>": 50286,
"<|hy|>": 50312,
"<|id|>": 50275,
"<|is|>": 50311,
"<|it|>": 50274,
"<|ja|>": 50266,
"<|jw|>": 50356,
"<|ka|>": 50329,
"<|kk|>": 50316,
"<|km|>": 50323,
"<|kn|>": 50306,
"<|ko|>": 50264,
"<|la|>": 50294,
"<|lb|>": 50345,
"<|ln|>": 50353,
"<|lo|>": 50336,
"<|lt|>": 50293,
"<|lv|>": 50301,
"<|mg|>": 50349,
"<|mi|>": 50295,
"<|mk|>": 50308,
"<|ml|>": 50296,
"<|mn|>": 50314,
"<|mr|>": 50320,
"<|ms|>": 50282,
"<|mt|>": 50343,
"<|my|>": 50346,
"<|ne|>": 50313,
"<|nl|>": 50271,
"<|nn|>": 50342,
"<|no|>": 50288,
"<|oc|>": 50328,
"<|pa|>": 50321,
"<|pl|>": 50269,
"<|ps|>": 50340,
"<|pt|>": 50267,
"<|ro|>": 50284,
"<|ru|>": 50263,
"<|sa|>": 50344,
"<|sd|>": 50332,
"<|si|>": 50322,
"<|sk|>": 50298,
"<|sl|>": 50305,
"<|sn|>": 50324,
"<|so|>": 50326,
"<|sq|>": 50317,
"<|sr|>": 50303,
"<|su|>": 50357,
"<|sv|>": 50273,
"<|sw|>": 50318,
"<|ta|>": 50287,
"<|te|>": 50299,
"<|tg|>": 50331,
"<|th|>": 50289,
"<|tk|>": 50341,
"<|tl|>": 50348,
"<|tr|>": 50268,
"<|tt|>": 50351,
"<|uk|>": 50280,
"<|ur|>": 50290,
"<|uz|>": 50337,
"<|vi|>": 50278,
"<|yi|>": 50335,
"<|yo|>": 50325,
"<|zh|>": 50260
},
"max_initial_timestamp_index": 1,
"max_length": 448,
"no_timestamps_token_id": 50363,
"pad_token_id": 50257,
"return_timestamps": false,
"suppress_tokens": [
1,
2,
7,
8,
9,
10,
14,
25,
26,
27,
28,
29,
31,
58,
59,
60,
61,
62,
63,
90,
91,
92,
93,
359,
503,
522,
542,
873,
893,
902,
918,
922,
931,
1350,
1853,
1982,
2460,
2627,
3246,
3253,
3268,
3536,
3846,
3961,
4183,
4667,
6585,
6647,
7273,
9061,
9383,
10428,
10929,
11938,
12033,
12331,
12562,
13793,
14157,
14635,
15265,
15618,
16553,
16604,
18362,
18956,
20075,
21675,
22520,
26130,
26161,
26435,
28279,
29464,
31650,
32302,
32470,
36865,
42863,
47425,
49870,
50254,
50258,
50358,
50359,
50360,
50361,
50362
],
"task_to_id": {
"transcribe": 50359,
"translate": 50358
}
}
[INFO|feature_extraction_utils.py:425] 2024-01-04 16:16:02,487 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/preprocessor_config.json
[INFO|tokenization_utils_base.py:2432] 2024-01-04 16:16:02,697 >> tokenizer config file saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tokenizer_config.json
[INFO|tokenization_utils_base.py:2441] 2024-01-04 16:16:02,698 >> Special tokens file saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/special_tokens_map.json
[INFO|configuration_utils.py:483] 2024-01-04 16:16:02,770 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/config.json
trainable params: 11,990,309 || all params: 253,728,185 || trainable%: 4.725651192436504
[INFO|trainer.py:522] 2024-01-04 16:16:03,058 >> max_steps is given, it will override any value given in num_train_epochs
[INFO|trainer.py:571] 2024-01-04 16:16:03,058 >> Using auto half precision backend
wandb: Currently logged in as: qmeeus. Use `wandb login --relogin` to force relogin
wandb: wandb version 0.16.1 is available! To upgrade, please run:
wandb: $ pip install wandb --upgrade
wandb: Tracking run with wandb version 0.15.12
wandb: Run data is saved locally in /usr/data/condor/execute/dir_32133/whisper_slu/wandb/run-20240104_161604-szfugz25
wandb: Run `wandb offline` to turn off syncing.
wandb: Syncing run run-2024-01-04_16-16-03
wandb: ⭐️ View project at https://wandb.ai/qmeeus/Whisper%20PEFT%20Fine-Tuning
wandb: πŸš€ View run at https://wandb.ai/qmeeus/Whisper%20PEFT%20Fine-Tuning/runs/szfugz25
[INFO|trainer.py:718] 2024-01-04 16:16:05,367 >> The following columns in the training set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:1712] 2024-01-04 16:16:05,412 >> ***** Running training *****
[INFO|trainer.py:1713] 2024-01-04 16:16:05,412 >> Num examples = 71,615
[INFO|trainer.py:1714] 2024-01-04 16:16:05,412 >> Num Epochs = 9
[INFO|trainer.py:1715] 2024-01-04 16:16:05,412 >> Instantaneous batch size per device = 4
[INFO|trainer.py:1718] 2024-01-04 16:16:05,412 >> Total train batch size (w. parallel, distributed & accumulation) = 128
[INFO|trainer.py:1719] 2024-01-04 16:16:05,412 >> Gradient Accumulation steps = 32
[INFO|trainer.py:1720] 2024-01-04 16:16:05,413 >> Total optimization steps = 5,000
[INFO|trainer.py:1721] 2024-01-04 16:16:05,414 >> Number of trainable parameters = 11,990,309
[INFO|integration_utils.py:722] 2024-01-04 16:16:05,416 >> Automatic Weights & Biases logging enabled, to disable set os.environ["WANDB_DISABLED"] = "true"
[WARNING|logging.py:314] 2024-01-04 16:16:05,440 >> You're using a WhisperTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
[INFO|trainer.py:718] 2024-01-04 16:39:41,300 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 16:47:18,571 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-200
[INFO|configuration_utils.py:483] 2024-01-04 16:47:18,573 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-200/config.json
[INFO|configuration_utils.py:594] 2024-01-04 16:47:18,575 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-200/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 16:47:21,264 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-200/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 16:47:21,268 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-200/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 17:10:58,917 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 17:18:36,642 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-400
[INFO|configuration_utils.py:483] 2024-01-04 17:18:36,645 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-400/config.json
[INFO|configuration_utils.py:594] 2024-01-04 17:18:36,646 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-400/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 17:18:39,103 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-400/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 17:18:39,105 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-400/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 17:42:19,026 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 17:49:57,271 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-600
[INFO|configuration_utils.py:483] 2024-01-04 17:49:57,273 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-600/config.json
[INFO|configuration_utils.py:594] 2024-01-04 17:49:57,275 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-600/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 17:50:00,352 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-600/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 17:50:00,355 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-600/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 18:13:32,213 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 18:21:08,995 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-800
[INFO|configuration_utils.py:483] 2024-01-04 18:21:08,998 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-800/config.json
[INFO|configuration_utils.py:594] 2024-01-04 18:21:09,000 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-800/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 18:21:11,464 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-800/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 18:21:11,466 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-800/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 18:44:51,227 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 18:52:34,548 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1000
[INFO|configuration_utils.py:483] 2024-01-04 18:52:34,551 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1000/config.json
[INFO|configuration_utils.py:594] 2024-01-04 18:52:34,553 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1000/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 18:52:37,349 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1000/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 18:52:37,353 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1000/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 19:16:14,157 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 19:23:53,038 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1200
[INFO|configuration_utils.py:483] 2024-01-04 19:23:53,040 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1200/config.json
[INFO|configuration_utils.py:594] 2024-01-04 19:23:53,042 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1200/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 19:23:55,385 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1200/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 19:23:55,388 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1200/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 19:47:32,898 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 19:55:11,424 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1400
[INFO|configuration_utils.py:483] 2024-01-04 19:55:11,426 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1400/config.json
[INFO|configuration_utils.py:594] 2024-01-04 19:55:11,428 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1400/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 19:55:13,704 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1400/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 19:55:13,706 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1400/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 20:18:49,683 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 20:26:30,082 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1600
[INFO|configuration_utils.py:483] 2024-01-04 20:26:30,085 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1600/config.json
[INFO|configuration_utils.py:594] 2024-01-04 20:26:30,087 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1600/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 20:26:32,387 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1600/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 20:26:32,390 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1600/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 20:50:08,100 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 20:57:48,948 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1800
[INFO|configuration_utils.py:483] 2024-01-04 20:57:48,951 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1800/config.json
[INFO|configuration_utils.py:594] 2024-01-04 20:57:48,953 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1800/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 20:57:51,514 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1800/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 20:57:51,517 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-1800/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 21:21:29,551 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 21:29:10,404 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2000
[INFO|configuration_utils.py:483] 2024-01-04 21:29:10,406 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2000/config.json
[INFO|configuration_utils.py:594] 2024-01-04 21:29:10,408 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2000/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 21:29:12,916 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2000/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 21:29:12,918 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2000/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 21:52:50,352 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 22:00:26,819 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2200
[INFO|configuration_utils.py:483] 2024-01-04 22:00:26,822 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2200/config.json
[INFO|configuration_utils.py:594] 2024-01-04 22:00:26,824 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2200/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 22:00:29,097 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2200/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 22:00:29,100 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2200/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 22:24:09,821 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 22:31:47,017 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2400
[INFO|configuration_utils.py:483] 2024-01-04 22:31:47,020 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2400/config.json
[INFO|configuration_utils.py:594] 2024-01-04 22:31:47,021 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2400/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 22:31:49,857 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2400/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 22:31:49,860 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2400/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 22:55:31,344 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 23:03:07,885 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2600
[INFO|configuration_utils.py:483] 2024-01-04 23:03:07,887 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2600/config.json
[INFO|configuration_utils.py:594] 2024-01-04 23:03:07,889 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2600/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 23:03:10,261 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2600/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 23:03:10,264 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2600/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 23:26:47,554 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-04 23:34:26,704 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2800
[INFO|configuration_utils.py:483] 2024-01-04 23:34:26,708 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2800/config.json
[INFO|configuration_utils.py:594] 2024-01-04 23:34:26,710 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2800/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-04 23:34:29,264 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2800/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-04 23:34:29,266 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-2800/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-04 23:58:10,328 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 00:05:48,535 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3000
[INFO|configuration_utils.py:483] 2024-01-05 00:05:48,538 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3000/config.json
[INFO|configuration_utils.py:594] 2024-01-05 00:05:48,562 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3000/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 00:05:50,957 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3000/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 00:05:50,960 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3000/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 00:29:27,582 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 00:37:04,888 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3200
[INFO|configuration_utils.py:483] 2024-01-05 00:37:04,891 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3200/config.json
[INFO|configuration_utils.py:594] 2024-01-05 00:37:04,892 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3200/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 00:37:07,573 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3200/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 00:37:07,575 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3200/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 01:00:45,990 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 01:08:25,718 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3400
[INFO|configuration_utils.py:483] 2024-01-05 01:08:25,721 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3400/config.json
[INFO|configuration_utils.py:594] 2024-01-05 01:08:25,722 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3400/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 01:08:28,043 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3400/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 01:08:28,046 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3400/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 01:32:02,935 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 01:39:42,260 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3600
[INFO|configuration_utils.py:483] 2024-01-05 01:39:42,262 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3600/config.json
[INFO|configuration_utils.py:594] 2024-01-05 01:39:42,264 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3600/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 01:39:45,037 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3600/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 01:39:45,040 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3600/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 02:03:18,499 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 02:10:54,524 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3800
[INFO|configuration_utils.py:483] 2024-01-05 02:10:54,528 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3800/config.json
[INFO|configuration_utils.py:594] 2024-01-05 02:10:54,530 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3800/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 02:10:57,233 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3800/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 02:10:57,235 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-3800/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 02:34:36,237 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 02:42:13,902 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4000
[INFO|configuration_utils.py:483] 2024-01-05 02:42:13,904 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4000/config.json
[INFO|configuration_utils.py:594] 2024-01-05 02:42:13,906 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4000/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 02:42:16,664 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4000/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 02:42:16,667 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4000/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 03:05:53,707 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 03:13:29,092 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4200
[INFO|configuration_utils.py:483] 2024-01-05 03:13:29,095 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4200/config.json
[INFO|configuration_utils.py:594] 2024-01-05 03:13:29,096 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4200/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 03:13:31,472 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4200/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 03:13:31,475 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4200/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 03:37:07,138 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 03:44:45,151 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4400
[INFO|configuration_utils.py:483] 2024-01-05 03:44:45,154 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4400/config.json
[INFO|configuration_utils.py:594] 2024-01-05 03:44:45,155 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4400/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 03:44:47,624 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4400/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 03:44:47,627 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4400/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 04:08:24,477 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 04:16:02,899 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4600
[INFO|configuration_utils.py:483] 2024-01-05 04:16:02,902 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4600/config.json
[INFO|configuration_utils.py:594] 2024-01-05 04:16:02,904 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4600/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 04:16:05,855 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4600/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 04:16:05,858 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4600/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 04:39:43,515 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 04:47:18,754 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4800
[INFO|configuration_utils.py:483] 2024-01-05 04:47:18,758 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4800/config.json
[INFO|configuration_utils.py:594] 2024-01-05 04:47:18,760 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4800/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 04:47:21,424 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4800/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 04:47:21,426 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-4800/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 05:10:56,324 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
[INFO|trainer.py:2895] 2024-01-05 05:18:32,407 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-5000
[INFO|configuration_utils.py:483] 2024-01-05 05:18:32,420 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-5000/config.json
[INFO|configuration_utils.py:594] 2024-01-05 05:18:32,422 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-5000/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 05:18:34,954 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-5000/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 05:18:34,957 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/tmp-checkpoint-5000/preprocessor_config.json
[INFO|trainer.py:1953] 2024-01-05 05:18:35,242 >>
Training completed. Do not forget to share your model on huggingface.co/models =)
[INFO|trainer.py:2895] 2024-01-05 05:18:35,248 >> Saving model checkpoint to /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner
[INFO|configuration_utils.py:483] 2024-01-05 05:18:35,250 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/config.json
[INFO|configuration_utils.py:594] 2024-01-05 05:18:35,252 >> Configuration saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/generation_config.json
[INFO|modeling_utils.py:2413] 2024-01-05 05:18:37,732 >> Model weights saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/model.safetensors
[INFO|feature_extraction_utils.py:425] 2024-01-05 05:18:37,734 >> Feature extractor saved in /esat/audioslave/qmeeus/exp/whisper_slu/finetune/whisper-small-spoken-ner/preprocessor_config.json
[INFO|trainer.py:718] 2024-01-05 05:18:37,740 >> The following columns in the evaluation set don't have a corresponding argument in `WhisperSLU.forward` and have been ignored: input_length. If input_length are not expected by `WhisperSLU.forward`, you can safely ignore this message.
wandb: Waiting for W&B process to finish... (success).
wandb:
wandb: Run history:
wandb: eval/f1_score β–ˆβ–†β–ƒβ–„β–„β–‚β–ƒβ–‚β–‚β–„β–β–β–ƒβ–‚β–‚β–‚β–β–‚β–‚β–‚β–ƒβ–‚β–‚β–‚β–‚β–‚
wandb: eval/label_f1 β–β–ƒβ–β–„β–„β–„β–…β–„β–†β–ˆβ–ˆβ–†β–ˆβ–†β–‡β–†β–†β–…β–†β–…β–†β–…β–„β–…β–„β–„
wandb: eval/loss β–ˆβ–†β–†β–…β–„β–„β–ƒβ–ƒβ–ƒβ–ƒβ–‚β–‚β–‚β–‚β–β–‚β–‚β–β–β–β–β–β–β–β–β–
wandb: eval/runtime β–ƒβ–ƒβ–„β–‚β–ˆβ–„β–„β–…β–†β–†β–‚β–ƒβ–‚β–„β–„β–ƒβ–…β–…β–‚β–ƒβ–β–ƒβ–„β–β–‚β–ƒ
wandb: eval/samples_per_second β–†β–†β–…β–‡β–β–…β–…β–„β–ƒβ–ƒβ–‡β–†β–‡β–…β–…β–†β–„β–„β–‡β–†β–ˆβ–…β–…β–ˆβ–‡β–†
wandb: eval/steps_per_second β–†β–†β–†β–†β–β–…β–…β–ƒβ–ƒβ–ƒβ–‡β–†β–‡β–„β–†β–†β–„β–„β–‡β–†β–ˆβ–†β–…β–ˆβ–‡β–†
wandb: eval/wer β–β–‚β–„β–…β–…β–†β–†β–‡β–†β–‡β–‡β–‡β–‡β–‡β–ˆβ–‡β–‡β–†β–‡β–‡β–‡β–‡β–‡β–‡β–‡β–‡
wandb: train/epoch β–β–β–β–β–‚β–‚β–‚β–‚β–‚β–ƒβ–ƒβ–ƒβ–ƒβ–ƒβ–„β–„β–„β–„β–„β–…β–…β–…β–…β–…β–…β–†β–†β–†β–†β–†β–‡β–‡β–‡β–‡β–‡β–‡β–ˆβ–ˆβ–ˆβ–ˆ
wandb: train/global_step β–β–β–β–β–‚β–‚β–‚β–‚β–‚β–ƒβ–ƒβ–ƒβ–ƒβ–ƒβ–„β–„β–„β–„β–„β–…β–…β–…β–…β–…β–…β–†β–†β–†β–†β–†β–‡β–‡β–‡β–‡β–‡β–‡β–ˆβ–ˆβ–ˆβ–ˆ
wandb: train/learning_rate β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‡β–‡β–‡β–‡β–‡β–†β–†β–†β–…β–…β–…β–…β–„β–„β–„β–„β–ƒβ–ƒβ–ƒβ–ƒβ–‚β–‚β–‚β–‚β–‚β–β–β–β–β–β–β–
wandb: train/loss β–ˆβ–‡β–…β–…β–…β–„β–„β–„β–„β–ƒβ–ƒβ–„β–ƒβ–ƒβ–ƒβ–ƒβ–ƒβ–ƒβ–ƒβ–‚β–‚β–‚β–‚β–β–‚β–‚β–‚β–β–β–β–β–‚β–‚β–β–β–β–β–‚β–β–
wandb: train/total_flos ▁
wandb: train/train_loss ▁
wandb: train/train_runtime ▁
wandb: train/train_samples_per_second ▁
wandb: train/train_steps_per_second ▁
wandb:
wandb: Run summary:
wandb: eval/f1_score 0.60982
wandb: eval/label_f1 0.83694
wandb: eval/loss 0.04444
wandb: eval/runtime 457.7019
wandb: eval/samples_per_second 2.185
wandb: eval/steps_per_second 0.546
wandb: eval/wer 0.10856
wandb: train/epoch 8.94
wandb: train/global_step 5000
wandb: train/learning_rate 0.0
wandb: train/loss 0.0297
wandb: train/total_flos 1.9542812433710973e+20
wandb: train/train_loss 0.03341
wandb: train/train_runtime 46949.8286
wandb: train/train_samples_per_second 13.632
wandb: train/train_steps_per_second 0.106
wandb:
wandb: πŸš€ View run run-2024-01-04_16-16-03 at: https://wandb.ai/qmeeus/Whisper%20PEFT%20Fine-Tuning/runs/szfugz25
wandb: ️⚑ View job at https://wandb.ai/qmeeus/Whisper%20PEFT%20Fine-Tuning/jobs/QXJ0aWZhY3RDb2xsZWN0aW9uOjEyNzQzMjk1OA==/version_details/v3
wandb: Synced 5 W&B file(s), 0 media file(s), 2 artifact file(s) and 0 other file(s)
wandb: Find logs at: ./wandb/run-20240104_161604-szfugz25/logs