DewiBrynJones commited on
Commit
2509f6c
·
verified ·
1 Parent(s): 795ae3b

Model save

Browse files
Files changed (2) hide show
  1. README.md +14 -13
  2. generation_config.json +1 -1
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: openai/whisper-large-v3
4
  tags:
@@ -15,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # whisper-large-v3-ft-btb-ca-cy
17
 
18
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the DewiBrynJones/banc-trawsgrifiadau-bangor-clean train main, cymen-arfor/15awr train+dev+test main dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.4037
21
- - Wer: 0.2791
22
 
23
  ## Model description
24
 
@@ -43,7 +44,7 @@ The following hyperparameters were used during training:
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
  - total_train_batch_size: 32
46
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_steps: 500
49
  - training_steps: 5000
@@ -53,16 +54,16 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Wer |
55
  |:-------------:|:------:|:----:|:---------------:|:------:|
56
- | 0.4739 | 0.6388 | 1000 | 0.4840 | 0.3518 |
57
- | 0.3335 | 1.2775 | 2000 | 0.4168 | 0.3130 |
58
- | 0.3281 | 1.9163 | 3000 | 0.3878 | 0.2965 |
59
- | 0.1901 | 2.5551 | 4000 | 0.3930 | 0.2849 |
60
- | 0.1306 | 3.1939 | 5000 | 0.4037 | 0.2791 |
61
 
62
 
63
  ### Framework versions
64
 
65
- - Transformers 4.44.0
66
- - Pytorch 2.4.0+cu121
67
- - Datasets 2.20.0
68
- - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: openai/whisper-large-v3
5
  tags:
 
16
 
17
  # whisper-large-v3-ft-btb-ca-cy
18
 
19
+ This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.4080
22
+ - Wer: 0.2782
23
 
24
  ## Model description
25
 
 
44
  - seed: 42
45
  - gradient_accumulation_steps: 2
46
  - total_train_batch_size: 32
47
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 500
50
  - training_steps: 5000
 
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Wer |
56
  |:-------------:|:------:|:----:|:---------------:|:------:|
57
+ | 0.9864 | 0.6555 | 1000 | 0.4853 | 0.3736 |
58
+ | 0.6533 | 1.3110 | 2000 | 0.4187 | 0.3088 |
59
+ | 0.5981 | 1.9666 | 3000 | 0.3845 | 0.2868 |
60
+ | 0.3903 | 2.6221 | 4000 | 0.3909 | 0.2854 |
61
+ | 0.2479 | 3.2776 | 5000 | 0.4080 | 0.2782 |
62
 
63
 
64
  ### Framework versions
65
 
66
+ - Transformers 4.46.1
67
+ - Pytorch 2.5.1+cu124
68
+ - Datasets 3.1.0
69
+ - Tokenizers 0.20.1
generation_config.json CHANGED
@@ -253,5 +253,5 @@
253
  "transcribe": 50360,
254
  "translate": 50359
255
  },
256
- "transformers_version": "4.44.0"
257
  }
 
253
  "transcribe": 50360,
254
  "translate": 50359
255
  },
256
+ "transformers_version": "4.46.1"
257
  }