appledora commited on
Commit
589f132
1 Parent(s): 13d73b2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -1
README.md CHANGED
@@ -40,10 +40,16 @@ if not os.path.exists("<RESAMPLED AUDIO FILE PATH>"):
40
  tfm.build(input_filepath= "<AUDIO FILE PATH>", output_filepath= "<RESAMPLED AUDIO FILE PATH>")
41
  ```
42
  ## Training
43
- We used the official [NeMo documentation on training an ASR model](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/asr/examples/kinyarwanda_asr.html) to prepare our transcript manifest and train our model. However, we did not train any custom tokenizer and instead downloaded the tokenizer from [banglaBERT-large](https://huggingface.co/csebuetnlp/banglabert_large/) for better vocabulary coverage. For validation, we have used `29589` samples separated from the training data and processed accordingly. The final validation score was `22.4% WER` , at epoch `164`.
 
 
 
 
 
44
  Training script : [training.sh](training.sh)
45
  ## Evaluation
46
  `14,016` test samples have been used to evaluate the dataset. The generated output file contains both ground truth and predicted strings. The final result is the Word Error Rate (WER) and Character Error Rate (CER) for the model.
 
47
  Evaluation script: [evaluation.sh](evaluation.sh)
48
 
49
  **Test Dataset WER/CER 69.25%/42.13%**
 
40
  tfm.build(input_filepath= "<AUDIO FILE PATH>", output_filepath= "<RESAMPLED AUDIO FILE PATH>")
41
  ```
42
  ## Training
43
+ We used the official [NeMo documentation on training an ASR model](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/asr/examples/kinyarwanda_asr.html)
44
+ to prepare our transcript manifest and train our model. However, we did not train any custom tokenizer and instead downloaded the tokenizer
45
+ from [banglaBERT-large](https://huggingface.co/csebuetnlp/banglabert_large/) for better vocabulary coverage.
46
+ For validation, we have used `29589` samples separated from the training data and processed accordingly.
47
+ **The final validation score was `22.4% WER` , at epoch `164`.**
48
+
49
  Training script : [training.sh](training.sh)
50
  ## Evaluation
51
  `14,016` test samples have been used to evaluate the dataset. The generated output file contains both ground truth and predicted strings. The final result is the Word Error Rate (WER) and Character Error Rate (CER) for the model.
52
+
53
  Evaluation script: [evaluation.sh](evaluation.sh)
54
 
55
  **Test Dataset WER/CER 69.25%/42.13%**