distil-wav2vec2 / README.md
OthmaneJ's picture
Update README.md
57f7010
|
raw
history blame
610 Bytes
metadata
language: en
datasets:
  - librispeech_asr
tags:
  - speech
  - audio
  - automatic-speech-recognition
license: apache-2.0

Distil-wav2vec2

This model is a distilled version of the wav2vec2 model (https://arxiv.org/pdf/2006.11477.pdf). This model is 4 times smaller and 3 times faster than the original wav2vec2 large model.

Evaluation results

When used with a light tri-gram language model head, this model achieves the following results :

Dataset WER
Librispeech-clean 0.127

Usage

notebook (google colab) at https://github.com/OthmaneJ/distil-wav2vec2