doof-ferb commited on
Commit
49a2658
1 Parent(s): 4deb04e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -0
README.md CHANGED
@@ -1,3 +1,66 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - doof-ferb/vlsp2020_vinai_100h
5
+ - doof-ferb/fpt_fosd
6
+ - doof-ferb/infore1_25hours
7
+ - doof-ferb/infore2_audiobooks
8
+ - quocanh34/viet_vlsp
9
+ - linhtran92/final_dataset_500hrs_wer0
10
+ - linhtran92/viet_youtube_asr_corpus_v2
11
+ - google/fleurs
12
+ - mozilla-foundation/common_voice_16_1
13
+ - vivos
14
+ language: ["vi"]
15
+ metrics: ["wer"]
16
+ library_name: transformers
17
+ base_model: openai/whisper-tiny
18
+ pipeline_tag: automatic-speech-recognition
19
+ model-index:
20
+ - name: doof-ferb/whisper-tiny-vi
21
+ results:
22
+ - task:
23
+ type: automatic-speech-recognition
24
+ dataset:
25
+ type: mozilla-foundation/common_voice_16_1
26
+ name: Mozilla CommonVoice (Vietnamese) v16.1
27
+ config: vi
28
+ split: test
29
+ metrics:
30
+ - type: wer
31
+ value: 26.6
32
+ verified: false
33
+ - task:
34
+ type: automatic-speech-recognition
35
+ dataset:
36
+ type: google/fleurs
37
+ name: Google FLEURS (Vietnamese)
38
+ config: vi_vn
39
+ split: test
40
+ metrics:
41
+ - type: wer
42
+ value: 37.1
43
+ verified: false
44
+ - task:
45
+ type: automatic-speech-recognition
46
+ dataset:
47
+ type: vivos
48
+ name: ĐHQG TPHCM VIVOS
49
+ split: test
50
+ metrics:
51
+ - type: wer
52
+ value: 18.7
53
+ verified: false
54
  ---
55
+
56
+ whisper tiny fine-tuned on a very big collection of vietnamese speech datasets
57
+
58
+ TODO:
59
+ - [x] training then publish checkpoint (*no ETA*)
60
+ - [x] evaluate WER on Common Voice & FLEURS
61
+ - [ ] convert to `openai-whisper`, `whisper.cpp`, `faster-whisper`
62
+ - [ ] convert to ONNX: to try `k2-fsa/sherpa-onnx` & `zhuzilin/whisper-openvino`
63
+
64
+ 21k steps, warm-up 5%, batch size 16×2 (kaggle free T4×2)
65
+
66
+ all training + evaluation scripts are on my repo: https://github.com/phineas-pta/fine-tune-whisper-vi