smajumdar94 commited on
Commit
e00282d
1 Parent(s): fafc2f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -10
README.md CHANGED
@@ -64,21 +64,21 @@ img {
64
  | [![Model size](https://img.shields.io/badge/Params-114M-lightgrey#model-badge)](#model-architecture)
65
  | [![Language](https://img.shields.io/badge/Language-en--US-lightgrey#model-badge)](#datasets)
66
 
67
- This collection contains large size versions of cache-aware FastConformer-Hybrid (around 114M parameters) with multiple look-ahead support, trained on a large scale english speech.
68
- These models are trained for streaming ASR which be used for streaming applications with a variety of latencies (0ms, 80ms, 480s, 1040ms).
69
 
70
 
71
  ## Model Architecture
72
 
73
- These models are cache-aware versions of Hybrid FastConfomer which are trianed for streaming ASR. You may find more info on cache-aware models here: [Cache-aware Streaming Conformer](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#cache-aware-streaming-conformer).
74
  The models are trained with multiple look-aheads which makes the model to be able to support different latencies.
75
- To learn on how to switch between different look-aheads, you may read the documentations on the cache-aware models.
76
 
77
  FastConformer [4] is an optimized version of the Conformer model [1], and
78
  you may find more information on the details of FastConformer here: [Fast-Conformer Model](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#fast-conformer).
79
 
80
  The model is trained in a multitask setup with joint Transducer and CTC decoder loss. You can find more about Hybrid Transducer-CTC training here: [Hybrid Transducer-CTC](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#hybrid-transducer-ctc).
81
- You may also find more on how to switch between the Transducer and CTC decoders in the documentations.
82
 
83
 
84
 
@@ -130,15 +130,11 @@ You will need to install [NVIDIA NeMo](https://github.com/NVIDIA/NeMo). We recom
130
  ```
131
  pip install nemo_toolkit['all']
132
  '''
133
- '''
134
- (if it causes an error):
135
- pip install nemo_toolkit[all]
136
- ```
137
 
138
  ### Simulate Streaming ASR
139
 
140
  You may use this script to simulate streaming ASR with these models: [cache-aware streaming simulation](https://github.com/NVIDIA/NeMo/blob/main/examples/asr/asr_cache_aware_streaming/speech_to_text_cache_aware_streaming_infer.py).
141
- You may use --att_context_size to set the context size otherwise the default which is the first context size in the list (1040ms) is going to be used.
142
 
143
  ### Transcribing using Python
144
  Cache-aware models are designed in a way that the model's predictions are the same in both offline and streaming mode.
 
64
  | [![Model size](https://img.shields.io/badge/Params-114M-lightgrey#model-badge)](#model-architecture)
65
  | [![Language](https://img.shields.io/badge/Language-en--US-lightgrey#model-badge)](#datasets)
66
 
67
+ This collection contains large-size versions of cache-aware FastConformer-Hybrid (around 114M parameters) with multiple look-ahead support, trained on a large scale english speech.
68
+ These models are trained for streaming ASR, which be used for streaming applications with a variety of latencies (0ms, 80ms, 480s, 1040ms).
69
 
70
 
71
  ## Model Architecture
72
 
73
+ These models are cache-aware versions of Hybrid FastConfomer which are trained for streaming ASR. You may find more info on cache-aware models here: [Cache-aware Streaming Conformer](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#cache-aware-streaming-conformer).
74
  The models are trained with multiple look-aheads which makes the model to be able to support different latencies.
75
+ To learn on how to switch between different look-ahead, you may read the documentation on the cache-aware models.
76
 
77
  FastConformer [4] is an optimized version of the Conformer model [1], and
78
  you may find more information on the details of FastConformer here: [Fast-Conformer Model](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#fast-conformer).
79
 
80
  The model is trained in a multitask setup with joint Transducer and CTC decoder loss. You can find more about Hybrid Transducer-CTC training here: [Hybrid Transducer-CTC](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#hybrid-transducer-ctc).
81
+ You may also find more on how to switch between the Transducer and CTC decoders in the documentation.
82
 
83
 
84
 
 
130
  ```
131
  pip install nemo_toolkit['all']
132
  '''
 
 
 
 
133
 
134
  ### Simulate Streaming ASR
135
 
136
  You may use this script to simulate streaming ASR with these models: [cache-aware streaming simulation](https://github.com/NVIDIA/NeMo/blob/main/examples/asr/asr_cache_aware_streaming/speech_to_text_cache_aware_streaming_infer.py).
137
+ You may use --att_context_size to set the context size otherwise, the default, which is the first context size in the list (1040ms), is going to be used.
138
 
139
  ### Transcribing using Python
140
  Cache-aware models are designed in a way that the model's predictions are the same in both offline and streaming mode.