wanchichen commited on
Commit
cceed7c
1 Parent(s): 11e9781

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -154,12 +154,15 @@ language:
154
 
155
  [XEUS - A Cross-lingual Encoder for Universal Speech]()
156
 
157
- XEUS is a large-scale multilingual speech encoder by Carnegie Mellon University's WAVLab that covers over **4000** languages. It is pre-trained on over 1 million hours of publicly available speech datasets. It can be requires fine-tuning to be used in downstream tasks such as Speech Recognition or Translation. XEUS uses the [E-Branchformer]() architecture and is trained using [HuBERT]()-style masked prediction of discrete speech tokens. During training, the input speech is also augmented with acoustic noise and reverberation, making XEUS more robust.
158
 
159
  XEUS tops the [ML-SUPERB]() multilingual speech recognition benchmark, outperforming [MMS](), [w2v-BERT 2.0](), and [XLS-R](). XEUS also sets a new state-of-the-art on 4 tasks in the monolingual [SUPERB]() benchmark.
160
 
161
-
162
 
163
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630438615c70c21d0eae6613/RCAWBxSuDLXJ5zdj-OBdn.png)
164
 
 
 
 
165
 
 
154
 
155
  [XEUS - A Cross-lingual Encoder for Universal Speech]()
156
 
157
+ XEUS is a large-scale multilingual speech encoder by Carnegie Mellon University's WAVLab that covers over **4000** languages. It is pre-trained on over 1 million hours of publicly available speech datasets. It can be requires fine-tuning to be used in downstream tasks such as Speech Recognition or Translation. XEUS uses the [E-Branchformer]() architecture and is trained using [HuBERT]()-style masked prediction of discrete speech tokens. During training, the input speech is also augmented with acoustic noise and reverberation, making XEUS more robust. The total model size is 577M parameters.
158
 
159
  XEUS tops the [ML-SUPERB]() multilingual speech recognition benchmark, outperforming [MMS](), [w2v-BERT 2.0](), and [XLS-R](). XEUS also sets a new state-of-the-art on 4 tasks in the monolingual [SUPERB]() benchmark.
160
 
161
+ ## Results
162
 
163
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630438615c70c21d0eae6613/RCAWBxSuDLXJ5zdj-OBdn.png)
164
 
165
+
166
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630438615c70c21d0eae6613/B3J2yL7C7XnE6-WxQbmRD.png)
167
+
168