lijialudew commited on
Commit
974c69d
1 Parent(s): 5490fe0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -31,15 +31,15 @@ For more information regarding this model, please checkout our paper
31
  - **Paper [optional]:** [More Information Needed]
32
 
33
  # Uses
34
- We develop fine-tuning recipe using SpeechBrain toolkit available at
35
 
36
  - **Repository:** https://github.com/jialuli3/speechbrain/tree/infant-voc-classification/recipes/wav2vec_kic
37
 
38
 
39
- ## Quick Start [optional]
40
 
41
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
42
- If you wish to use fairseq framework, the following code snippet can be used to load our pretrained model.
43
 
44
  <pre><code>
45
  import torch
@@ -136,6 +136,8 @@ We test 4 unlabeled datasets on unsupervised pretrained W2V2-base models:
136
  We then fine-tune pretrained models on 11.7h of LB labeled home recordings, the f1 scores across three tasks are
137
 
138
  ![results](results.png)
 
 
139
  For more details of experiments and results, please refer to our paper.
140
 
141
  # Citation
 
31
  - **Paper [optional]:** [More Information Needed]
32
 
33
  # Uses
34
+ We develop our complete fine-tuning recipe using SpeechBrain toolkit available at
35
 
36
  - **Repository:** https://github.com/jialuli3/speechbrain/tree/infant-voc-classification/recipes/wav2vec_kic
37
 
38
 
39
+ ## Quick Start
40
 
41
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
42
+ If you wish to use fairseq framework, the following code snippet provides two functions of loading our pretrained model and extracting W2V2 features.
43
 
44
  <pre><code>
45
  import torch
 
136
  We then fine-tune pretrained models on 11.7h of LB labeled home recordings, the f1 scores across three tasks are
137
 
138
  ![results](results.png)
139
+
140
+ Additionally, we improve our model performances through incorporating relevant labeled home recordings and data augmentation techniques of SpecAug and noise/reverberation corruption.
141
  For more details of experiments and results, please refer to our paper.
142
 
143
  # Citation