jungjee commited on
Commit
099d455
1 Parent(s): 0a66e0f

Update model

Browse files
Files changed (23) hide show
  1. README.md +285 -0
  2. meta.yaml +8 -0
  3. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/8epoch.pth +3 -0
  4. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/RESULTS.md +17 -0
  5. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/config.yaml +194 -0
  6. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/backward_time.png +0 -0
  7. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/clip.png +0 -0
  8. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/eer.png +0 -0
  9. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/forward_time.png +0 -0
  10. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/gpu_max_cached_mem_GB.png +0 -0
  11. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/grad_norm.png +0 -0
  12. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/iter_time.png +0 -0
  13. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/loss.png +0 -0
  14. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/loss_scale.png +0 -0
  15. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/mindcf.png +0 -0
  16. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/n_trials.png +0 -0
  17. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/nontrg_mean.png +0 -0
  18. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/nontrg_std.png +0 -0
  19. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/optim0_lr0.png +0 -0
  20. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/optim_step_time.png +0 -0
  21. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/train_time.png +0 -0
  22. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/trg_mean.png +0 -0
  23. save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/trg_std.png +0 -0
README.md ADDED
@@ -0,0 +1,285 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - espnet
4
+ - audio
5
+ - speaker-recognition
6
+ language: multilingual
7
+ datasets:
8
+ - voxceleb
9
+ license: cc-by-4.0
10
+ ---
11
+
12
+ ## ESPnet2 SPK model
13
+
14
+ ### `espnet/voxcelebs12_ecapa_frozen`
15
+
16
+ This model was trained by Jungjee using voxceleb recipe in [espnet](https://github.com/espnet/espnet/).
17
+
18
+ ### Demo: How to use in ESPnet2
19
+
20
+ Follow the [ESPnet installation instructions](https://espnet.github.io/espnet/installation.html)
21
+ if you haven't done that already.
22
+
23
+ ```bash
24
+ cd espnet
25
+ git checkout d9646a75807a30afff85a83155247a81cc7fe389
26
+ pip install -e .
27
+ cd egs2/voxceleb/spk1
28
+ ./run.sh --skip_data_prep false --skip_train true --download_model espnet/voxcelebs12_ecapa_frozen
29
+ ```
30
+
31
+ <!-- Generated by scripts/utils/show_spk_result.py -->
32
+ # RESULTS
33
+ ## Environments
34
+ date: 2024-01-02 18:13:10.597501
35
+
36
+ - python version: 3.9.16 (main, Mar 8 2023, 14:00:05) [GCC 11.2.0]
37
+ - espnet version: 202310
38
+ - pytorch version: 2.0.1
39
+
40
+ | | Mean | Std |
41
+ |---|---|---|
42
+ | Target | 8.0224 | 2.7891 |
43
+ | Non-target | 1.9364 | 1.9364 |
44
+
45
+ | Model name | EER(%) | minDCF |
46
+ |---|---|---|
47
+ | conf/tuning/train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm | 0.638 | 0.04994 |
48
+
49
+ ## SPK config
50
+
51
+ <details><summary>expand</summary>
52
+
53
+ ```
54
+ config: conf/tuning/train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm.yaml
55
+ print_config: false
56
+ log_level: INFO
57
+ drop_last_iter: true
58
+ dry_run: false
59
+ iterator_type: category
60
+ valid_iterator_type: sequence
61
+ output_dir: exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp
62
+ ngpu: 1
63
+ seed: 0
64
+ num_workers: 8
65
+ num_att_plot: 0
66
+ dist_backend: nccl
67
+ dist_init_method: env://
68
+ dist_world_size: 4
69
+ dist_rank: 0
70
+ local_rank: 0
71
+ dist_master_addr: localhost
72
+ dist_master_port: 37387
73
+ dist_launcher: null
74
+ multiprocessing_distributed: true
75
+ unused_parameters: false
76
+ sharded_ddp: false
77
+ cudnn_enabled: true
78
+ cudnn_benchmark: true
79
+ cudnn_deterministic: false
80
+ collect_stats: false
81
+ write_collected_feats: false
82
+ max_epoch: 40
83
+ patience: null
84
+ val_scheduler_criterion:
85
+ - valid
86
+ - loss
87
+ early_stopping_criterion:
88
+ - valid
89
+ - loss
90
+ - min
91
+ best_model_criterion:
92
+ - - valid
93
+ - eer
94
+ - min
95
+ keep_nbest_models: 3
96
+ nbest_averaging_interval: 0
97
+ grad_clip: 9999
98
+ grad_clip_type: 2.0
99
+ grad_noise: false
100
+ accum_grad: 1
101
+ no_forward_run: false
102
+ resume: true
103
+ train_dtype: float32
104
+ use_amp: true
105
+ log_interval: 100
106
+ use_matplotlib: true
107
+ use_tensorboard: true
108
+ create_graph_in_tensorboard: false
109
+ use_wandb: false
110
+ wandb_project: null
111
+ wandb_id: null
112
+ wandb_entity: null
113
+ wandb_name: null
114
+ wandb_model_log_interval: -1
115
+ detect_anomaly: false
116
+ pretrain_path: null
117
+ init_param: []
118
+ ignore_init_mismatch: false
119
+ freeze_param:
120
+ - frontend.upstream
121
+ num_iters_per_epoch: null
122
+ batch_size: 512
123
+ valid_batch_size: 40
124
+ batch_bins: 1000000
125
+ valid_batch_bins: null
126
+ train_shape_file:
127
+ - exp/spk_stats_16k_sp/train/speech_shape
128
+ valid_shape_file:
129
+ - exp/spk_stats_16k_sp/valid/speech_shape
130
+ batch_type: folded
131
+ valid_batch_type: null
132
+ fold_length:
133
+ - 120000
134
+ sort_in_batch: descending
135
+ shuffle_within_batch: false
136
+ sort_batch: descending
137
+ multiple_iterator: false
138
+ chunk_length: 500
139
+ chunk_shift_ratio: 0.5
140
+ num_cache_chunks: 1024
141
+ chunk_excluded_key_prefixes: []
142
+ train_data_path_and_name_and_type:
143
+ - - dump/raw/voxceleb12_devs_sp/wav.scp
144
+ - speech
145
+ - sound
146
+ - - dump/raw/voxceleb12_devs_sp/utt2spk
147
+ - spk_labels
148
+ - text
149
+ valid_data_path_and_name_and_type:
150
+ - - dump/raw/voxceleb1_test/trial.scp
151
+ - speech
152
+ - sound
153
+ - - dump/raw/voxceleb1_test/trial2.scp
154
+ - speech2
155
+ - sound
156
+ - - dump/raw/voxceleb1_test/trial_label
157
+ - spk_labels
158
+ - text
159
+ allow_variable_data_keys: false
160
+ max_cache_size: 0.0
161
+ max_cache_fd: 32
162
+ valid_max_cache_size: null
163
+ exclude_weight_decay: false
164
+ exclude_weight_decay_conf: {}
165
+ optim: adam
166
+ optim_conf:
167
+ lr: 0.001
168
+ weight_decay: 5.0e-05
169
+ amsgrad: false
170
+ scheduler: cosineannealingwarmuprestarts
171
+ scheduler_conf:
172
+ first_cycle_steps: 71280
173
+ cycle_mult: 1.0
174
+ max_lr: 0.001
175
+ min_lr: 5.0e-06
176
+ warmup_steps: 1000
177
+ gamma: 0.75
178
+ init: null
179
+ use_preprocessor: true
180
+ input_size: null
181
+ target_duration: 3.0
182
+ spk2utt: dump/raw/voxceleb12_devs_sp/spk2utt
183
+ spk_num: 21615
184
+ sample_rate: 16000
185
+ num_eval: 10
186
+ rir_scp: ''
187
+ model_conf:
188
+ extract_feats_in_collect_stats: false
189
+ frontend: s3prl
190
+ frontend_conf:
191
+ frontend_conf:
192
+ upstream: wavlm_large
193
+ download_dir: ./hub
194
+ multilayer_feature: true
195
+ specaug: null
196
+ specaug_conf: {}
197
+ normalize: utterance_mvn
198
+ normalize_conf:
199
+ norm_vars: false
200
+ encoder: ecapa_tdnn
201
+ encoder_conf:
202
+ model_scale: 8
203
+ ndim: 1024
204
+ output_size: 1536
205
+ pooling: chn_attn_stat
206
+ pooling_conf: {}
207
+ projector: rawnet3
208
+ projector_conf:
209
+ output_size: 192
210
+ preprocessor: spk
211
+ preprocessor_conf:
212
+ target_duration: 3.0
213
+ sample_rate: 16000
214
+ num_eval: 5
215
+ noise_apply_prob: 0.5
216
+ noise_info:
217
+ - - 1.0
218
+ - dump/raw/musan_speech.scp
219
+ - - 4
220
+ - 7
221
+ - - 13
222
+ - 20
223
+ - - 1.0
224
+ - dump/raw/musan_noise.scp
225
+ - - 1
226
+ - 1
227
+ - - 0
228
+ - 15
229
+ - - 1.0
230
+ - dump/raw/musan_music.scp
231
+ - - 1
232
+ - 1
233
+ - - 5
234
+ - 15
235
+ rir_apply_prob: 0.5
236
+ rir_scp: dump/raw/rirs.scp
237
+ loss: aamsoftmax_sc_topk
238
+ loss_conf:
239
+ margin: 0.3
240
+ scale: 30
241
+ K: 3
242
+ mp: 0.06
243
+ k_top: 5
244
+ required:
245
+ - output_dir
246
+ version: '202308'
247
+ distributed: true
248
+ ```
249
+
250
+ </details>
251
+
252
+
253
+
254
+ ### Citing ESPnet
255
+
256
+ ```BibTex
257
+ @inproceedings{watanabe2018espnet,
258
+ author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Yalta and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai},
259
+ title={{ESPnet}: End-to-End Speech Processing Toolkit},
260
+ year={2018},
261
+ booktitle={Proceedings of Interspeech},
262
+ pages={2207--2211},
263
+ doi={10.21437/Interspeech.2018-1456},
264
+ url={http://dx.doi.org/10.21437/Interspeech.2018-1456}
265
+ }
266
+
267
+
268
+
269
+
270
+
271
+
272
+ ```
273
+
274
+ or arXiv:
275
+
276
+ ```bibtex
277
+ @misc{watanabe2018espnet,
278
+ title={ESPnet: End-to-End Speech Processing Toolkit},
279
+ author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Yalta and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai},
280
+ year={2018},
281
+ eprint={1804.00015},
282
+ archivePrefix={arXiv},
283
+ primaryClass={cs.CL}
284
+ }
285
+ ```
meta.yaml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ espnet: '202310'
2
+ files:
3
+ model_file: save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/8epoch.pth
4
+ python: "3.9.16 (main, Mar 8 2023, 14:00:05) \n[GCC 11.2.0]"
5
+ timestamp: 1704237231.660156
6
+ torch: 2.0.1
7
+ yaml_files:
8
+ train_config: save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/config.yaml
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/8epoch.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a48f08c93fab652aa6c034aa7a41eaa601c8dbf88e9dce88734e0881f5ea5aa1
3
+ size 1389974254
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/RESULTS.md ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- Generated by scripts/utils/show_spk_result.py -->
2
+ # RESULTS
3
+ ## Environments
4
+ date: 2024-01-02 18:13:10.597501
5
+
6
+ - python version: 3.9.16 (main, Mar 8 2023, 14:00:05) [GCC 11.2.0]
7
+ - espnet version: 202310
8
+ - pytorch version: 2.0.1
9
+
10
+ | | Mean | Std |
11
+ |---|---|---|
12
+ | Target | 8.0224 | 2.7891 |
13
+ | Non-target | 1.9364 | 1.9364 |
14
+
15
+ | Model name | EER(%) | minDCF |
16
+ |---|---|---|
17
+ | conf/tuning/train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm | 0.638 | 0.04994 |
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/config.yaml ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config: conf/tuning/train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm.yaml
2
+ print_config: false
3
+ log_level: INFO
4
+ drop_last_iter: true
5
+ dry_run: false
6
+ iterator_type: category
7
+ valid_iterator_type: sequence
8
+ output_dir: exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp
9
+ ngpu: 1
10
+ seed: 0
11
+ num_workers: 8
12
+ num_att_plot: 0
13
+ dist_backend: nccl
14
+ dist_init_method: env://
15
+ dist_world_size: 4
16
+ dist_rank: 0
17
+ local_rank: 0
18
+ dist_master_addr: localhost
19
+ dist_master_port: 37387
20
+ dist_launcher: null
21
+ multiprocessing_distributed: true
22
+ unused_parameters: false
23
+ sharded_ddp: false
24
+ cudnn_enabled: true
25
+ cudnn_benchmark: true
26
+ cudnn_deterministic: false
27
+ collect_stats: false
28
+ write_collected_feats: false
29
+ max_epoch: 40
30
+ patience: null
31
+ val_scheduler_criterion:
32
+ - valid
33
+ - loss
34
+ early_stopping_criterion:
35
+ - valid
36
+ - loss
37
+ - min
38
+ best_model_criterion:
39
+ - - valid
40
+ - eer
41
+ - min
42
+ keep_nbest_models: 3
43
+ nbest_averaging_interval: 0
44
+ grad_clip: 9999
45
+ grad_clip_type: 2.0
46
+ grad_noise: false
47
+ accum_grad: 1
48
+ no_forward_run: false
49
+ resume: true
50
+ train_dtype: float32
51
+ use_amp: true
52
+ log_interval: 100
53
+ use_matplotlib: true
54
+ use_tensorboard: true
55
+ create_graph_in_tensorboard: false
56
+ use_wandb: false
57
+ wandb_project: null
58
+ wandb_id: null
59
+ wandb_entity: null
60
+ wandb_name: null
61
+ wandb_model_log_interval: -1
62
+ detect_anomaly: false
63
+ pretrain_path: null
64
+ init_param: []
65
+ ignore_init_mismatch: false
66
+ freeze_param:
67
+ - frontend.upstream
68
+ num_iters_per_epoch: null
69
+ batch_size: 512
70
+ valid_batch_size: 40
71
+ batch_bins: 1000000
72
+ valid_batch_bins: null
73
+ train_shape_file:
74
+ - exp/spk_stats_16k_sp/train/speech_shape
75
+ valid_shape_file:
76
+ - exp/spk_stats_16k_sp/valid/speech_shape
77
+ batch_type: folded
78
+ valid_batch_type: null
79
+ fold_length:
80
+ - 120000
81
+ sort_in_batch: descending
82
+ shuffle_within_batch: false
83
+ sort_batch: descending
84
+ multiple_iterator: false
85
+ chunk_length: 500
86
+ chunk_shift_ratio: 0.5
87
+ num_cache_chunks: 1024
88
+ chunk_excluded_key_prefixes: []
89
+ train_data_path_and_name_and_type:
90
+ - - dump/raw/voxceleb12_devs_sp/wav.scp
91
+ - speech
92
+ - sound
93
+ - - dump/raw/voxceleb12_devs_sp/utt2spk
94
+ - spk_labels
95
+ - text
96
+ valid_data_path_and_name_and_type:
97
+ - - dump/raw/voxceleb1_test/trial.scp
98
+ - speech
99
+ - sound
100
+ - - dump/raw/voxceleb1_test/trial2.scp
101
+ - speech2
102
+ - sound
103
+ - - dump/raw/voxceleb1_test/trial_label
104
+ - spk_labels
105
+ - text
106
+ allow_variable_data_keys: false
107
+ max_cache_size: 0.0
108
+ max_cache_fd: 32
109
+ valid_max_cache_size: null
110
+ exclude_weight_decay: false
111
+ exclude_weight_decay_conf: {}
112
+ optim: adam
113
+ optim_conf:
114
+ lr: 0.001
115
+ weight_decay: 5.0e-05
116
+ amsgrad: false
117
+ scheduler: cosineannealingwarmuprestarts
118
+ scheduler_conf:
119
+ first_cycle_steps: 71280
120
+ cycle_mult: 1.0
121
+ max_lr: 0.001
122
+ min_lr: 5.0e-06
123
+ warmup_steps: 1000
124
+ gamma: 0.75
125
+ init: null
126
+ use_preprocessor: true
127
+ input_size: null
128
+ target_duration: 3.0
129
+ spk2utt: dump/raw/voxceleb12_devs_sp/spk2utt
130
+ spk_num: 21615
131
+ sample_rate: 16000
132
+ num_eval: 10
133
+ rir_scp: ''
134
+ model_conf:
135
+ extract_feats_in_collect_stats: false
136
+ frontend: s3prl
137
+ frontend_conf:
138
+ frontend_conf:
139
+ upstream: wavlm_large
140
+ download_dir: ./hub
141
+ multilayer_feature: true
142
+ specaug: null
143
+ specaug_conf: {}
144
+ normalize: utterance_mvn
145
+ normalize_conf:
146
+ norm_vars: false
147
+ encoder: ecapa_tdnn
148
+ encoder_conf:
149
+ model_scale: 8
150
+ ndim: 1024
151
+ output_size: 1536
152
+ pooling: chn_attn_stat
153
+ pooling_conf: {}
154
+ projector: rawnet3
155
+ projector_conf:
156
+ output_size: 192
157
+ preprocessor: spk
158
+ preprocessor_conf:
159
+ target_duration: 3.0
160
+ sample_rate: 16000
161
+ num_eval: 5
162
+ noise_apply_prob: 0.5
163
+ noise_info:
164
+ - - 1.0
165
+ - dump/raw/musan_speech.scp
166
+ - - 4
167
+ - 7
168
+ - - 13
169
+ - 20
170
+ - - 1.0
171
+ - dump/raw/musan_noise.scp
172
+ - - 1
173
+ - 1
174
+ - - 0
175
+ - 15
176
+ - - 1.0
177
+ - dump/raw/musan_music.scp
178
+ - - 1
179
+ - 1
180
+ - - 5
181
+ - 15
182
+ rir_apply_prob: 0.5
183
+ rir_scp: dump/raw/rirs.scp
184
+ loss: aamsoftmax_sc_topk
185
+ loss_conf:
186
+ margin: 0.3
187
+ scale: 30
188
+ K: 3
189
+ mp: 0.06
190
+ k_top: 5
191
+ required:
192
+ - output_dir
193
+ version: '202308'
194
+ distributed: true
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/backward_time.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/clip.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/eer.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/forward_time.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/gpu_max_cached_mem_GB.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/grad_norm.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/iter_time.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/loss.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/loss_scale.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/mindcf.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/n_trials.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/nontrg_mean.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/nontrg_std.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/optim0_lr0.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/optim_step_time.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/train_time.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/trg_mean.png ADDED
save_exp/spk_train_ecapa_Vox12_emb192_torchmelspec_subcentertopk_wavlm_raw_sp/images/trg_std.png ADDED