Refactor
Browse filesThis view is limited to 50 files because it contains too many changes. Β
See raw diff
- exp/asr_train_asr_raw_en_word/126epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/137epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/154epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/174epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/192epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/195epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/197epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/198epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/199epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/200epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/66epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/67epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/68epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/70epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/71epoch.pth +0 -3
- exp/asr_train_asr_raw_en_word/train.1.log +0 -0
- exp/asr_train_asr_raw_en_word/train.2.log +0 -557
- exp/asr_train_asr_raw_en_word/train.acc.ave.pth +0 -1
- exp/asr_train_asr_raw_en_word/train.acc.ave_1best.pth +0 -1
- exp/asr_train_asr_raw_en_word/train.acc.ave_5best.pth +0 -3
- exp/asr_train_asr_raw_en_word/train.acc.best.pth +0 -1
- exp/asr_train_asr_raw_en_word/train.log +0 -0
- exp/asr_train_asr_raw_en_word/train.loss.ave.pth +0 -1
- exp/asr_train_asr_raw_en_word/train.loss.ave_1best.pth +0 -1
- exp/asr_train_asr_raw_en_word/train.loss.ave_5best.pth +0 -3
- exp/asr_train_asr_raw_en_word/train.loss.best.pth +0 -1
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/RESULTS.md +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.100ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.101ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.102ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.103ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.104ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.105ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.106ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.107ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.108ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.109ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.10ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.110ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.111ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.112ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.113ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.114ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.115ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.116ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.117ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.118ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.119ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.11ep.png +0 -0
- exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.120ep.png +0 -0
exp/asr_train_asr_raw_en_word/126epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:524df69c03273cdd0593cc5b62f2c82b2b18b86853a9390252f255f5713b4fba
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/137epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:9c89eb2c81fd8181534ddff02232ee375ccd26f28ab7a7831344e45c8cbfbc7a
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/154epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:99cfe5a5d6abdd10797e678903097cbb28a7d7734d047c44d1ebc1d06b14fb87
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/174epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:deee35c14d5235e2b506147b84b0889071fb45ee13c8aa185a24fd98ce4d5b25
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/192epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:be20c8f3c0d747b7597ea0f4dee8f4aa1ef30329179ecec20cbdce4b24ee865e
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/195epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:1ba689f98c1511c7dbbbf79841b918d7b52831197ee439176c4526344eb97bff
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/197epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:a94a5974c14fe9f71f5be435473bac0059da139d1901f42a11e41fa6166f8485
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/198epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3bf19a35ffb6767da69e0f7fdd562d362fd2bbd99de0171b5c00c4de0f49bff4
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/199epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:7398df97ab7cd040dd143e7d863502a4617ccfe6a914fef59d78ca7887fb8745
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/200epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:e05edb01164041f1cc97611a5fcc69a192e4cf26eec351f06ba21da40b702133
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/66epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:d01c3e6b22c6a3ea3a69d1cbf4a9fcd2f331bc52c2a43ab857326f1dd8213c74
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/67epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:cb468260a5db3ed6bb528e94fc27b67e750d99d37482c5d01d25a5242cdd2607
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/68epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:0b41e2bc1ba70e18bb8982e54332df7ce3b25f2ae74aed3238ea7ac22a4bf0bc
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/70epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:839f0de480a2fb929682d232346c2f1b996991979a260c4abc03443224dd11d7
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/71epoch.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:7e986ca9e9d613f4a2ab154f573a8d229dee39923b28ee1e5bda676d4c1366fb
|
3 |
-
size 110488525
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/train.1.log
DELETED
The diff for this file is too large to render.
See raw diff
|
|
exp/asr_train_asr_raw_en_word/train.2.log
DELETED
@@ -1,557 +0,0 @@
|
|
1 |
-
# Running on v030.ib.bridges2.psc.edu
|
2 |
-
# Started at Mon Jan 31 01:55:52 EST 2022
|
3 |
-
# SLURMD_NODENAME=v030
|
4 |
-
# SLURM_CLUSTER_NAME=bridges2
|
5 |
-
# SLURM_CONF=/var/spool/slurm/d/conf-cache/slurm.conf
|
6 |
-
# SLURM_CPUS_ON_NODE=5
|
7 |
-
# SLURM_CPUS_PER_TASK=1
|
8 |
-
# SLURM_EXPORT_ENV=PATH
|
9 |
-
# SLURM_GET_USER_ENV=1
|
10 |
-
# SLURM_GTIDS=0
|
11 |
-
# SLURM_JOBID=6473234
|
12 |
-
# SLURM_JOB_ACCOUNT=cis210027p
|
13 |
-
# SLURM_JOB_CPUS_PER_NODE=5
|
14 |
-
# SLURM_JOB_GID=24886
|
15 |
-
# SLURM_JOB_GPUS=5
|
16 |
-
# SLURM_JOB_ID=6473234
|
17 |
-
# SLURM_JOB_NAME=exp/asr_train_asr_raw_en_word/train.log
|
18 |
-
# SLURM_JOB_NODELIST=v030
|
19 |
-
# SLURM_JOB_NUM_NODES=1
|
20 |
-
# SLURM_JOB_PARTITION=GPU-shared
|
21 |
-
# SLURM_JOB_QOS=gpu
|
22 |
-
# SLURM_JOB_UID=82326
|
23 |
-
# SLURM_JOB_USER=ganesank
|
24 |
-
# SLURM_LOCALID=0
|
25 |
-
# SLURM_NNODES=1
|
26 |
-
# SLURM_NODEID=0
|
27 |
-
# SLURM_NODELIST=v030
|
28 |
-
# SLURM_NODE_ALIASES='(null)'
|
29 |
-
# SLURM_OPEN_MODE=a
|
30 |
-
# SLURM_PRIO_PROCESS=0
|
31 |
-
# SLURM_PROCID=0
|
32 |
-
# SLURM_SUBMIT_DIR=/ocean/projects/cis210027p/ganesank/karthik_new/espnet/egs2/dstc2/asr2
|
33 |
-
# SLURM_SUBMIT_HOST=br012.ib.bridges2.psc.edu
|
34 |
-
# SLURM_TASKS_PER_NODE=5
|
35 |
-
# SLURM_TASK_PID=54262
|
36 |
-
# SLURM_TOPOLOGY_ADDR=v030
|
37 |
-
# SLURM_TOPOLOGY_ADDR_PATTERN=node
|
38 |
-
# SLURM_WORKING_CLUSTER=bridges2:br003:6814:9216:109
|
39 |
-
# python3 -m espnet2.bin.asr_train --use_preprocessor true --bpemodel none --token_type word --token_list data/en_token_list/word/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/valid/wav.scp,speech,sound --valid_data_path_and_name_and_type dump/raw/valid/text,text,text --valid_shape_file exp/asr_stats_raw_en_word/valid/speech_shape --valid_shape_file exp/asr_stats_raw_en_word/valid/text_shape.word --resume true --fold_length 80000 --fold_length 150 --output_dir exp/asr_train_asr_raw_en_word --config conf/train_asr.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp/asr_stats_raw_en_word/train/feats_stats.npz --train_data_path_and_name_and_type dump/raw/train/wav.scp,speech,sound --train_data_path_and_name_and_type dump/raw/train/text,text,text --train_shape_file exp/asr_stats_raw_en_word/train/speech_shape --train_shape_file exp/asr_stats_raw_en_word/train/text_shape.word --ngpu 1 --multiprocessing_distributed True
|
40 |
-
/ocean/projects/cis210027p/ganesank/karthik_new/espnet/tools/venv/bin/python3 /ocean/projects/cis210027p/ganesank/karthik_new/espnet/espnet2/bin/asr_train.py --use_preprocessor true --bpemodel none --token_type word --token_list data/en_token_list/word/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/valid/wav.scp,speech,sound --valid_data_path_and_name_and_type dump/raw/valid/text,text,text --valid_shape_file exp/asr_stats_raw_en_word/valid/speech_shape --valid_shape_file exp/asr_stats_raw_en_word/valid/text_shape.word --resume true --fold_length 80000 --fold_length 150 --output_dir exp/asr_train_asr_raw_en_word --config conf/train_asr.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp/asr_stats_raw_en_word/train/feats_stats.npz --train_data_path_and_name_and_type dump/raw/train/wav.scp,speech,sound --train_data_path_and_name_and_type dump/raw/train/text,text,text --train_shape_file exp/asr_stats_raw_en_word/train/speech_shape --train_shape_file exp/asr_stats_raw_en_word/train/text_shape.word --ngpu 1 --multiprocessing_distributed True
|
41 |
-
[v030] 2022-01-31 01:56:49,667 (asr:382) INFO: Vocabulary size: 613
|
42 |
-
[v030] 2022-01-31 01:57:10,900 (abs_task:1132) INFO: pytorch.version=1.8.1+cu102, cuda.available=True, cudnn.version=7605, cudnn.benchmark=False, cudnn.deterministic=True
|
43 |
-
[v030] 2022-01-31 01:57:11,231 (abs_task:1133) INFO: Model structure:
|
44 |
-
ESPnetASRModel(
|
45 |
-
(frontend): DefaultFrontend(
|
46 |
-
(stft): Stft(n_fft=512, win_length=512, hop_length=128, center=True, normalized=False, onesided=True)
|
47 |
-
(frontend): Frontend()
|
48 |
-
(logmel): LogMel(sr=16000, n_fft=512, n_mels=80, fmin=0, fmax=8000.0, htk=False)
|
49 |
-
)
|
50 |
-
(specaug): SpecAug(
|
51 |
-
(time_warp): TimeWarp(window=5, mode=bicubic)
|
52 |
-
(freq_mask): MaskAlongAxis(mask_width_range=[0, 30], num_mask=2, axis=freq)
|
53 |
-
(time_mask): MaskAlongAxis(mask_width_range=[0, 40], num_mask=2, axis=time)
|
54 |
-
)
|
55 |
-
(normalize): GlobalMVN(stats_file=exp/asr_stats_raw_en_word/train/feats_stats.npz, norm_means=True, norm_vars=True)
|
56 |
-
(encoder): TransformerEncoder(
|
57 |
-
(embed): Conv2dSubsampling(
|
58 |
-
(conv): Sequential(
|
59 |
-
(0): Conv2d(1, 256, kernel_size=(3, 3), stride=(2, 2))
|
60 |
-
(1): ReLU()
|
61 |
-
(2): Conv2d(256, 256, kernel_size=(3, 3), stride=(2, 2))
|
62 |
-
(3): ReLU()
|
63 |
-
)
|
64 |
-
(out): Sequential(
|
65 |
-
(0): Linear(in_features=4864, out_features=256, bias=True)
|
66 |
-
(1): PositionalEncoding(
|
67 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
68 |
-
)
|
69 |
-
)
|
70 |
-
)
|
71 |
-
(encoders): MultiSequential(
|
72 |
-
(0): EncoderLayer(
|
73 |
-
(self_attn): MultiHeadedAttention(
|
74 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
75 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
76 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
77 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
78 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
79 |
-
)
|
80 |
-
(feed_forward): PositionwiseFeedForward(
|
81 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
82 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
83 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
84 |
-
(activation): ReLU()
|
85 |
-
)
|
86 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
87 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
88 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
89 |
-
)
|
90 |
-
(1): EncoderLayer(
|
91 |
-
(self_attn): MultiHeadedAttention(
|
92 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
93 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
94 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
95 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
96 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
97 |
-
)
|
98 |
-
(feed_forward): PositionwiseFeedForward(
|
99 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
100 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
101 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
102 |
-
(activation): ReLU()
|
103 |
-
)
|
104 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
105 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
106 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
107 |
-
)
|
108 |
-
(2): EncoderLayer(
|
109 |
-
(self_attn): MultiHeadedAttention(
|
110 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
111 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
112 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
113 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
114 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
115 |
-
)
|
116 |
-
(feed_forward): PositionwiseFeedForward(
|
117 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
118 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
119 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
120 |
-
(activation): ReLU()
|
121 |
-
)
|
122 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
123 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
124 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
125 |
-
)
|
126 |
-
(3): EncoderLayer(
|
127 |
-
(self_attn): MultiHeadedAttention(
|
128 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
129 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
130 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
131 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
132 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
133 |
-
)
|
134 |
-
(feed_forward): PositionwiseFeedForward(
|
135 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
136 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
137 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
138 |
-
(activation): ReLU()
|
139 |
-
)
|
140 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
141 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
142 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
143 |
-
)
|
144 |
-
(4): EncoderLayer(
|
145 |
-
(self_attn): MultiHeadedAttention(
|
146 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
147 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
148 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
149 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
150 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
151 |
-
)
|
152 |
-
(feed_forward): PositionwiseFeedForward(
|
153 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
154 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
155 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
156 |
-
(activation): ReLU()
|
157 |
-
)
|
158 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
159 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
160 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
161 |
-
)
|
162 |
-
(5): EncoderLayer(
|
163 |
-
(self_attn): MultiHeadedAttention(
|
164 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
165 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
166 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
167 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
168 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
169 |
-
)
|
170 |
-
(feed_forward): PositionwiseFeedForward(
|
171 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
172 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
173 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
174 |
-
(activation): ReLU()
|
175 |
-
)
|
176 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
177 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
178 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
179 |
-
)
|
180 |
-
(6): EncoderLayer(
|
181 |
-
(self_attn): MultiHeadedAttention(
|
182 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
183 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
184 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
185 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
186 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
187 |
-
)
|
188 |
-
(feed_forward): PositionwiseFeedForward(
|
189 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
190 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
191 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
192 |
-
(activation): ReLU()
|
193 |
-
)
|
194 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
195 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
196 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
197 |
-
)
|
198 |
-
(7): EncoderLayer(
|
199 |
-
(self_attn): MultiHeadedAttention(
|
200 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
201 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
202 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
203 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
204 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
205 |
-
)
|
206 |
-
(feed_forward): PositionwiseFeedForward(
|
207 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
208 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
209 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
210 |
-
(activation): ReLU()
|
211 |
-
)
|
212 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
213 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
214 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
215 |
-
)
|
216 |
-
(8): EncoderLayer(
|
217 |
-
(self_attn): MultiHeadedAttention(
|
218 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
219 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
220 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
221 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
222 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
223 |
-
)
|
224 |
-
(feed_forward): PositionwiseFeedForward(
|
225 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
226 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
227 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
228 |
-
(activation): ReLU()
|
229 |
-
)
|
230 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
231 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
232 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
233 |
-
)
|
234 |
-
(9): EncoderLayer(
|
235 |
-
(self_attn): MultiHeadedAttention(
|
236 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
237 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
238 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
239 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
240 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
241 |
-
)
|
242 |
-
(feed_forward): PositionwiseFeedForward(
|
243 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
244 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
245 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
246 |
-
(activation): ReLU()
|
247 |
-
)
|
248 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
249 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
250 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
251 |
-
)
|
252 |
-
(10): EncoderLayer(
|
253 |
-
(self_attn): MultiHeadedAttention(
|
254 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
255 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
256 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
257 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
258 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
259 |
-
)
|
260 |
-
(feed_forward): PositionwiseFeedForward(
|
261 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
262 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
263 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
264 |
-
(activation): ReLU()
|
265 |
-
)
|
266 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
267 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
268 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
269 |
-
)
|
270 |
-
(11): EncoderLayer(
|
271 |
-
(self_attn): MultiHeadedAttention(
|
272 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
273 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
274 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
275 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
276 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
277 |
-
)
|
278 |
-
(feed_forward): PositionwiseFeedForward(
|
279 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
280 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
281 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
282 |
-
(activation): ReLU()
|
283 |
-
)
|
284 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
285 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
286 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
287 |
-
)
|
288 |
-
)
|
289 |
-
(after_norm): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
290 |
-
)
|
291 |
-
(decoder): TransformerDecoder(
|
292 |
-
(embed): Sequential(
|
293 |
-
(0): Embedding(613, 256)
|
294 |
-
(1): PositionalEncoding(
|
295 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
296 |
-
)
|
297 |
-
)
|
298 |
-
(after_norm): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
299 |
-
(output_layer): Linear(in_features=256, out_features=613, bias=True)
|
300 |
-
(decoders): MultiSequential(
|
301 |
-
(0): DecoderLayer(
|
302 |
-
(self_attn): MultiHeadedAttention(
|
303 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
304 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
305 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
306 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
307 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
308 |
-
)
|
309 |
-
(src_attn): MultiHeadedAttention(
|
310 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
311 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
312 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
313 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
314 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
315 |
-
)
|
316 |
-
(feed_forward): PositionwiseFeedForward(
|
317 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
318 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
319 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
320 |
-
(activation): ReLU()
|
321 |
-
)
|
322 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
323 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
324 |
-
(norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
325 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
326 |
-
)
|
327 |
-
(1): DecoderLayer(
|
328 |
-
(self_attn): MultiHeadedAttention(
|
329 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
330 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
331 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
332 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
333 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
334 |
-
)
|
335 |
-
(src_attn): MultiHeadedAttention(
|
336 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
337 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
338 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
339 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
340 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
341 |
-
)
|
342 |
-
(feed_forward): PositionwiseFeedForward(
|
343 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
344 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
345 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
346 |
-
(activation): ReLU()
|
347 |
-
)
|
348 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
349 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
350 |
-
(norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
351 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
352 |
-
)
|
353 |
-
(2): DecoderLayer(
|
354 |
-
(self_attn): MultiHeadedAttention(
|
355 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
356 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
357 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
358 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
359 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
360 |
-
)
|
361 |
-
(src_attn): MultiHeadedAttention(
|
362 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
363 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
364 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
365 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
366 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
367 |
-
)
|
368 |
-
(feed_forward): PositionwiseFeedForward(
|
369 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
370 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
371 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
372 |
-
(activation): ReLU()
|
373 |
-
)
|
374 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
375 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
376 |
-
(norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
377 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
378 |
-
)
|
379 |
-
(3): DecoderLayer(
|
380 |
-
(self_attn): MultiHeadedAttention(
|
381 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
382 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
383 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
384 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
385 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
386 |
-
)
|
387 |
-
(src_attn): MultiHeadedAttention(
|
388 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
389 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
390 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
391 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
392 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
393 |
-
)
|
394 |
-
(feed_forward): PositionwiseFeedForward(
|
395 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
396 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
397 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
398 |
-
(activation): ReLU()
|
399 |
-
)
|
400 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
401 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
402 |
-
(norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
403 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
404 |
-
)
|
405 |
-
(4): DecoderLayer(
|
406 |
-
(self_attn): MultiHeadedAttention(
|
407 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
408 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
409 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
410 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
411 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
412 |
-
)
|
413 |
-
(src_attn): MultiHeadedAttention(
|
414 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
415 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
416 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
417 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
418 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
419 |
-
)
|
420 |
-
(feed_forward): PositionwiseFeedForward(
|
421 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
422 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
423 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
424 |
-
(activation): ReLU()
|
425 |
-
)
|
426 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
427 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
428 |
-
(norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
429 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
430 |
-
)
|
431 |
-
(5): DecoderLayer(
|
432 |
-
(self_attn): MultiHeadedAttention(
|
433 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
434 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
435 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
436 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
437 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
438 |
-
)
|
439 |
-
(src_attn): MultiHeadedAttention(
|
440 |
-
(linear_q): Linear(in_features=256, out_features=256, bias=True)
|
441 |
-
(linear_k): Linear(in_features=256, out_features=256, bias=True)
|
442 |
-
(linear_v): Linear(in_features=256, out_features=256, bias=True)
|
443 |
-
(linear_out): Linear(in_features=256, out_features=256, bias=True)
|
444 |
-
(dropout): Dropout(p=0.0, inplace=False)
|
445 |
-
)
|
446 |
-
(feed_forward): PositionwiseFeedForward(
|
447 |
-
(w_1): Linear(in_features=256, out_features=2048, bias=True)
|
448 |
-
(w_2): Linear(in_features=2048, out_features=256, bias=True)
|
449 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
450 |
-
(activation): ReLU()
|
451 |
-
)
|
452 |
-
(norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
453 |
-
(norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
454 |
-
(norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
|
455 |
-
(dropout): Dropout(p=0.1, inplace=False)
|
456 |
-
)
|
457 |
-
)
|
458 |
-
)
|
459 |
-
(ctc): CTC(
|
460 |
-
(ctc_lo): Linear(in_features=256, out_features=613, bias=True)
|
461 |
-
(ctc_loss): CTCLoss()
|
462 |
-
)
|
463 |
-
(criterion_att): LabelSmoothingLoss(
|
464 |
-
(criterion): KLDivLoss()
|
465 |
-
)
|
466 |
-
)
|
467 |
-
|
468 |
-
Model summary:
|
469 |
-
Class Name: ESPnetASRModel
|
470 |
-
Total Number of model parameters: 27.56 M
|
471 |
-
Number of trainable parameters: 27.56 M (100.0%)
|
472 |
-
Size: 110.26 MB
|
473 |
-
Type: torch.float32
|
474 |
-
[v030] 2022-01-31 01:57:11,231 (abs_task:1136) INFO: Optimizer:
|
475 |
-
Adam (
|
476 |
-
Parameter Group 0
|
477 |
-
amsgrad: False
|
478 |
-
betas: (0.9, 0.999)
|
479 |
-
eps: 1e-08
|
480 |
-
initial_lr: 0.0002
|
481 |
-
lr: 8e-09
|
482 |
-
weight_decay: 0
|
483 |
-
)
|
484 |
-
[v030] 2022-01-31 01:57:11,232 (abs_task:1137) INFO: Scheduler: WarmupLR(warmup_steps=25000)
|
485 |
-
[v030] 2022-01-31 01:57:11,236 (abs_task:1146) INFO: Saving the configuration in exp/asr_train_asr_raw_en_word/config.yaml
|
486 |
-
[v030] 2022-01-31 01:57:11,750 (abs_task:1493) INFO: [train] dataset:
|
487 |
-
ESPnetDataset(
|
488 |
-
speech: {"path": "dump/raw/train/wav.scp", "type": "sound"}
|
489 |
-
text: {"path": "dump/raw/train/text", "type": "text"}
|
490 |
-
preprocess: <espnet2.train.preprocessor.CommonPreprocessor object at 0x1459a71f8310>)
|
491 |
-
[v030] 2022-01-31 01:57:11,751 (abs_task:1494) INFO: [train] Batch sampler: FoldedBatchSampler(N-batch=547, batch_size=20, shape_files=['exp/asr_stats_raw_en_word/train/speech_shape', 'exp/asr_stats_raw_en_word/train/text_shape.word'], sort_in_batch=descending, sort_batch=descending)
|
492 |
-
[v030] 2022-01-31 01:57:11,787 (abs_task:1495) INFO: [train] mini-batch sizes summary: N-batch=547, mean=19.9, min=6, max=20
|
493 |
-
[v030] 2022-01-31 01:57:12,032 (abs_task:1493) INFO: [valid] dataset:
|
494 |
-
ESPnetDataset(
|
495 |
-
speech: {"path": "dump/raw/valid/wav.scp", "type": "sound"}
|
496 |
-
text: {"path": "dump/raw/valid/text", "type": "text"}
|
497 |
-
preprocess: <espnet2.train.preprocessor.CommonPreprocessor object at 0x1459a3c2d520>)
|
498 |
-
[v030] 2022-01-31 01:57:12,033 (abs_task:1494) INFO: [valid] Batch sampler: FoldedBatchSampler(N-batch=178, batch_size=20, shape_files=['exp/asr_stats_raw_en_word/valid/speech_shape', 'exp/asr_stats_raw_en_word/valid/text_shape.word'], sort_in_batch=descending, sort_batch=descending)
|
499 |
-
[v030] 2022-01-31 01:57:12,033 (abs_task:1495) INFO: [valid] mini-batch sizes summary: N-batch=178, mean=20.0, min=19, max=20
|
500 |
-
[v030] 2022-01-31 01:57:12,059 (abs_task:1493) INFO: [plot_att] dataset:
|
501 |
-
ESPnetDataset(
|
502 |
-
speech: {"path": "dump/raw/valid/wav.scp", "type": "sound"}
|
503 |
-
text: {"path": "dump/raw/valid/text", "type": "text"}
|
504 |
-
preprocess: <espnet2.train.preprocessor.CommonPreprocessor object at 0x1459a3c2dfa0>)
|
505 |
-
[v030] 2022-01-31 01:57:12,059 (abs_task:1494) INFO: [plot_att] Batch sampler: UnsortedBatchSampler(N-batch=3559, batch_size=1, key_file=exp/asr_stats_raw_en_word/valid/speech_shape,
|
506 |
-
[v030] 2022-01-31 01:57:12,059 (abs_task:1495) INFO: [plot_att] mini-batch sizes summary: N-batch=3, mean=1.0, min=1, max=1
|
507 |
-
[v030] 2022-01-31 01:57:12,109 (trainer:273) INFO: 1/2epoch started
|
508 |
-
[v030] 2022-01-31 01:57:30,639 (trainer:653) INFO: 1epoch:train:1-27batch: iter_time=0.230, forward_time=0.249, loss=141.855, loss_att=47.067, loss_ctc=236.642, acc=4.953e-04, backward_time=0.036, optim_step_time=0.026, optim0_lr0=1.200e-07, train_time=0.632
|
509 |
-
[v030] 2022-01-31 01:57:39,688 (trainer:653) INFO: 1epoch:train:28-54batch: iter_time=0.188, forward_time=0.069, loss=142.019, loss_att=49.139, loss_ctc=234.898, acc=2.572e-04, backward_time=0.030, optim_step_time=0.024, optim0_lr0=3.360e-07, train_time=0.373
|
510 |
-
[v030] 2022-01-31 01:57:45,424 (trainer:653) INFO: 1epoch:train:55-81batch: iter_time=0.082, forward_time=0.058, loss=136.567, loss_att=49.707, loss_ctc=223.427, acc=4.240e-04, backward_time=0.029, optim_step_time=0.023, optim0_lr0=5.520e-07, train_time=0.213
|
511 |
-
[v030] 2022-01-31 01:57:51,546 (trainer:653) INFO: 1epoch:train:82-108batch: iter_time=0.103, forward_time=0.054, loss=140.269, loss_att=55.069, loss_ctc=225.468, acc=3.225e-04, backward_time=0.026, optim_step_time=0.022, optim0_lr0=7.680e-07, train_time=0.227
|
512 |
-
[v030] 2022-01-31 01:57:57,839 (trainer:653) INFO: 1epoch:train:109-135batch: iter_time=0.136, forward_time=0.034, loss=103.266, loss_att=50.612, loss_ctc=155.921, acc=2.684e-04, backward_time=0.023, optim_step_time=0.022, optim0_lr0=9.840e-07, train_time=0.233
|
513 |
-
[v030] 2022-01-31 01:58:07,335 (trainer:653) INFO: 1epoch:train:136-162batch: iter_time=0.133, forward_time=0.138, loss=86.321, loss_att=51.425, loss_ctc=121.217, acc=2.486e-04, backward_time=0.030, optim_step_time=0.022, optim0_lr0=1.200e-06, train_time=0.351
|
514 |
-
[v030] 2022-01-31 01:58:14,180 (trainer:653) INFO: 1epoch:train:163-189batch: iter_time=0.118, forward_time=0.068, loss=59.836, loss_att=46.904, loss_ctc=72.767, acc=2.503e-04, backward_time=0.026, optim_step_time=0.022, optim0_lr0=1.416e-06, train_time=0.253
|
515 |
-
[v030] 2022-01-31 01:58:19,882 (trainer:653) INFO: 1epoch:train:190-216batch: iter_time=0.110, forward_time=0.035, loss=62.973, loss_att=55.924, loss_ctc=70.023, acc=2.639e-04, backward_time=0.023, optim_step_time=0.022, optim0_lr0=1.632e-06, train_time=0.211
|
516 |
-
[v030] 2022-01-31 01:58:25,457 (trainer:653) INFO: 1epoch:train:217-243batch: iter_time=0.080, forward_time=0.058, loss=50.057, loss_att=50.607, loss_ctc=49.508, acc=0.000e+00, backward_time=0.026, optim_step_time=0.023, optim0_lr0=1.848e-06, train_time=0.206
|
517 |
-
[v030] 2022-01-31 01:58:40,212 (trainer:653) INFO: 1epoch:train:244-270batch: iter_time=0.090, forward_time=0.381, loss=42.156, loss_att=44.508, loss_ctc=39.803, acc=0.001, backward_time=0.027, optim_step_time=0.023, optim0_lr0=2.064e-06, train_time=0.546
|
518 |
-
[v030] 2022-01-31 01:58:49,595 (trainer:653) INFO: 1epoch:train:271-297batch: iter_time=0.165, forward_time=0.104, loss=35.478, loss_att=38.536, loss_ctc=32.420, acc=0.004, backward_time=0.031, optim_step_time=0.025, optim0_lr0=2.280e-06, train_time=0.347
|
519 |
-
[v030] 2022-01-31 01:58:55,079 (trainer:653) INFO: 1epoch:train:298-324batch: iter_time=0.107, forward_time=0.033, loss=40.375, loss_att=43.577, loss_ctc=37.174, acc=0.014, backward_time=0.023, optim_step_time=0.022, optim0_lr0=2.496e-06, train_time=0.203
|
520 |
-
[v030] 2022-01-31 01:59:05,215 (trainer:653) INFO: 1epoch:train:325-351batch: iter_time=0.110, forward_time=0.102, loss=51.612, loss_att=55.003, loss_ctc=48.221, acc=0.023, backward_time=0.095, optim_step_time=0.032, optim0_lr0=2.712e-06, train_time=0.375
|
521 |
-
[v030] 2022-01-31 01:59:16,439 (trainer:653) INFO: 1epoch:train:352-378batch: iter_time=0.198, forward_time=0.133, loss=43.941, loss_att=47.359, loss_ctc=40.524, acc=0.052, backward_time=0.035, optim_step_time=0.025, optim0_lr0=2.928e-06, train_time=0.415
|
522 |
-
[v030] 2022-01-31 01:59:22,380 (trainer:653) INFO: 1epoch:train:379-405batch: iter_time=0.055, forward_time=0.095, loss=40.773, loss_att=44.045, loss_ctc=37.502, acc=0.088, backward_time=0.027, optim_step_time=0.023, optim0_lr0=3.144e-06, train_time=0.220
|
523 |
-
[v030] 2022-01-31 01:59:28,375 (trainer:653) INFO: 1epoch:train:406-432batch: iter_time=0.121, forward_time=0.037, loss=41.249, loss_att=44.500, loss_ctc=37.998, acc=0.114, backward_time=0.023, optim_step_time=0.022, optim0_lr0=3.360e-06, train_time=0.222
|
524 |
-
[v030] 2022-01-31 01:59:40,639 (trainer:653) INFO: 1epoch:train:433-459batch: iter_time=0.094, forward_time=0.278, loss=47.537, loss_att=50.761, loss_ctc=44.314, acc=0.119, backward_time=0.035, optim_step_time=0.025, optim0_lr0=3.576e-06, train_time=0.454
|
525 |
-
[v030] 2022-01-31 02:00:15,522 (trainer:653) INFO: 1epoch:train:460-486batch: iter_time=0.126, forward_time=1.014, loss=37.846, loss_att=40.791, loss_ctc=34.901, acc=0.159, backward_time=0.079, optim_step_time=0.031, optim0_lr0=3.792e-06, train_time=1.292
|
526 |
-
[v030] 2022-01-31 02:00:29,360 (trainer:653) INFO: 1epoch:train:487-513batch: iter_time=0.110, forward_time=0.262, loss=48.077, loss_att=51.027, loss_ctc=45.127, acc=0.149, backward_time=0.061, optim_step_time=0.033, optim0_lr0=4.008e-06, train_time=0.512
|
527 |
-
[v030] 2022-01-31 02:00:38,109 (trainer:653) INFO: 1epoch:train:514-540batch: iter_time=0.097, forward_time=0.136, loss=35.719, loss_att=38.304, loss_ctc=33.134, acc=0.214, backward_time=0.036, optim_step_time=0.027, optim0_lr0=4.224e-06, train_time=0.324
|
528 |
-
[v030] 2022-01-31 02:02:20,408 (trainer:328) INFO: 1epoch results: [train] iter_time=0.124, forward_time=0.165, loss=68.972, loss_att=47.724, loss_ctc=90.219, acc=0.049, backward_time=0.036, optim_step_time=0.025, optim0_lr0=2.200e-06, train_time=0.379, time=3 minutes and 28.1 seconds, total_count=547, gpu_max_cached_mem_GB=4.266, [valid] loss=42.143, loss_att=42.969, loss_ctc=41.317, acc=0.226, cer=0.917, wer=1.000, cer_ctc=1.000, time=34.01 seconds, total_count=178, gpu_max_cached_mem_GB=4.891, [att_plot] time=1 minute and 6.1 seconds, total_count=0, gpu_max_cached_mem_GB=4.891
|
529 |
-
[v030] 2022-01-31 02:02:24,238 (trainer:375) INFO: The best model has been updated: train.loss, valid.loss, train.acc, valid.acc
|
530 |
-
[v030] 2022-01-31 02:02:24,290 (trainer:261) INFO: 2/2epoch started. Estimated time to finish: 5 minutes and 12.15 seconds
|
531 |
-
[v030] 2022-01-31 02:02:39,772 (trainer:653) INFO: 2epoch:train:1-27batch: iter_time=0.200, forward_time=0.227, loss=37.230, loss_att=39.722, loss_ctc=34.738, acc=0.219, backward_time=0.046, optim_step_time=0.026, optim0_lr0=4.496e-06, train_time=0.540
|
532 |
-
[v030] 2022-01-31 02:02:46,635 (trainer:653) INFO: 2epoch:train:28-54batch: iter_time=0.143, forward_time=0.046, loss=35.185, loss_att=37.229, loss_ctc=33.140, acc=0.229, backward_time=0.024, optim_step_time=0.022, optim0_lr0=4.712e-06, train_time=0.254
|
533 |
-
[v030] 2022-01-31 02:02:54,463 (trainer:653) INFO: 2epoch:train:55-81batch: iter_time=0.136, forward_time=0.073, loss=42.493, loss_att=44.580, loss_ctc=40.405, acc=0.204, backward_time=0.026, optim_step_time=0.032, optim0_lr0=4.928e-06, train_time=0.290
|
534 |
-
[v030] 2022-01-31 02:03:02,383 (trainer:653) INFO: 2epoch:train:82-108batch: iter_time=0.100, forward_time=0.122, loss=44.804, loss_att=46.801, loss_ctc=42.807, acc=0.193, backward_time=0.026, optim_step_time=0.023, optim0_lr0=5.144e-06, train_time=0.293
|
535 |
-
[v030] 2022-01-31 02:03:10,308 (trainer:653) INFO: 2epoch:train:109-135batch: iter_time=0.091, forward_time=0.111, loss=37.952, loss_att=39.864, loss_ctc=36.040, acc=0.221, backward_time=0.045, optim_step_time=0.023, optim0_lr0=5.360e-06, train_time=0.293
|
536 |
-
[v030] 2022-01-31 02:03:29,729 (trainer:653) INFO: 2epoch:train:136-162batch: iter_time=0.116, forward_time=0.490, loss=36.738, loss_att=38.327, loss_ctc=35.149, acc=0.230, backward_time=0.046, optim_step_time=0.026, optim0_lr0=5.576e-06, train_time=0.719
|
537 |
-
[v030] 2022-01-31 02:03:41,965 (trainer:653) INFO: 2epoch:train:163-189batch: iter_time=0.154, forward_time=0.226, loss=33.661, loss_att=34.931, loss_ctc=32.390, acc=0.232, backward_time=0.028, optim_step_time=0.023, optim0_lr0=5.792e-06, train_time=0.453
|
538 |
-
[v030] 2022-01-31 02:03:52,572 (trainer:653) INFO: 2epoch:train:190-216batch: iter_time=0.108, forward_time=0.164, loss=37.235, loss_att=38.509, loss_ctc=35.961, acc=0.236, backward_time=0.069, optim_step_time=0.023, optim0_lr0=6.008e-06, train_time=0.393
|
539 |
-
[v030] 2022-01-31 02:04:09,632 (trainer:653) INFO: 2epoch:train:217-243batch: iter_time=0.137, forward_time=0.341, loss=36.192, loss_att=37.130, loss_ctc=35.254, acc=0.226, backward_time=0.074, optim_step_time=0.031, optim0_lr0=6.224e-06, train_time=0.631
|
540 |
-
[v030] 2022-01-31 02:04:24,555 (trainer:653) INFO: 2epoch:train:244-270batch: iter_time=0.165, forward_time=0.303, loss=34.811, loss_att=35.738, loss_ctc=33.883, acc=0.241, backward_time=0.029, optim_step_time=0.029, optim0_lr0=6.440e-06, train_time=0.552
|
541 |
-
[v030] 2022-01-31 02:04:41,937 (trainer:653) INFO: 2epoch:train:271-297batch: iter_time=0.122, forward_time=0.326, loss=35.547, loss_att=36.079, loss_ctc=35.016, acc=0.238, backward_time=0.109, optim_step_time=0.036, optim0_lr0=6.656e-06, train_time=0.644
|
542 |
-
[v030] 2022-01-31 02:04:54,885 (trainer:653) INFO: 2epoch:train:298-324batch: iter_time=0.096, forward_time=0.271, loss=34.077, loss_att=34.718, loss_ctc=33.436, acc=0.243, backward_time=0.042, optim_step_time=0.026, optim0_lr0=6.872e-06, train_time=0.479
|
543 |
-
[v030] 2022-01-31 02:05:05,565 (trainer:653) INFO: 2epoch:train:325-351batch: iter_time=0.225, forward_time=0.076, loss=35.986, loss_att=36.355, loss_ctc=35.616, acc=0.242, backward_time=0.042, optim_step_time=0.024, optim0_lr0=7.088e-06, train_time=0.395
|
544 |
-
[v030] 2022-01-31 02:05:14,712 (trainer:653) INFO: 2epoch:train:352-378batch: iter_time=0.117, forward_time=0.152, loss=36.101, loss_att=36.323, loss_ctc=35.880, acc=0.243, backward_time=0.025, optim_step_time=0.023, optim0_lr0=7.304e-06, train_time=0.338
|
545 |
-
[v030] 2022-01-31 02:05:31,872 (trainer:653) INFO: 2epoch:train:379-405batch: iter_time=0.080, forward_time=0.434, loss=31.234, loss_att=31.494, loss_ctc=30.973, acc=0.279, backward_time=0.056, optim_step_time=0.027, optim0_lr0=7.520e-06, train_time=0.635
|
546 |
-
[v030] 2022-01-31 02:05:51,754 (trainer:653) INFO: 2epoch:train:406-432batch: iter_time=0.093, forward_time=0.395, loss=34.236, loss_att=34.248, loss_ctc=34.225, acc=0.278, backward_time=0.135, optim_step_time=0.037, optim0_lr0=7.736e-06, train_time=0.735
|
547 |
-
[v030] 2022-01-31 02:06:02,624 (trainer:653) INFO: 2epoch:train:433-459batch: iter_time=0.127, forward_time=0.178, loss=30.387, loss_att=30.207, loss_ctc=30.567, acc=0.282, backward_time=0.037, optim_step_time=0.027, optim0_lr0=7.952e-06, train_time=0.403
|
548 |
-
[v030] 2022-01-31 02:06:13,226 (trainer:653) INFO: 2epoch:train:460-486batch: iter_time=0.101, forward_time=0.167, loss=34.295, loss_att=34.131, loss_ctc=34.460, acc=0.281, backward_time=0.060, optim_step_time=0.030, optim0_lr0=8.168e-06, train_time=0.392
|
549 |
-
[v030] 2022-01-31 02:06:23,189 (trainer:653) INFO: 2epoch:train:487-513batch: iter_time=0.091, forward_time=0.160, loss=31.108, loss_att=30.714, loss_ctc=31.502, acc=0.286, backward_time=0.057, optim_step_time=0.032, optim0_lr0=8.384e-06, train_time=0.368
|
550 |
-
[v030] 2022-01-31 02:06:30,758 (trainer:653) INFO: 2epoch:train:514-540batch: iter_time=0.154, forward_time=0.051, loss=34.146, loss_att=33.576, loss_ctc=34.716, acc=0.262, backward_time=0.026, optim_step_time=0.028, optim0_lr0=8.600e-06, train_time=0.280
|
551 |
-
[v030] 2022-01-31 02:08:56,243 (trainer:328) INFO: 2epoch results: [train] iter_time=0.128, forward_time=0.235, loss=35.545, loss_att=36.390, loss_ctc=34.699, acc=0.245, backward_time=0.054, optim_step_time=0.028, optim0_lr0=6.576e-06, train_time=0.479, time=4 minutes and 23.23 seconds, total_count=1094, gpu_max_cached_mem_GB=4.891, [valid] loss=34.559, loss_att=31.596, loss_ctc=37.522, acc=0.298, cer=0.977, wer=1.000, cer_ctc=1.000, time=1 minute and 19.1 seconds, total_count=356, gpu_max_cached_mem_GB=4.891, [att_plot] time=49.47 seconds, total_count=0, gpu_max_cached_mem_GB=4.891
|
552 |
-
[v030] 2022-01-31 02:08:59,763 (trainer:375) INFO: The best model has been updated: train.loss, valid.loss, train.acc, valid.acc
|
553 |
-
[v030] 2022-01-31 02:08:59,768 (trainer:431) INFO: The training was finished at 2 epochs
|
554 |
-
# Accounting: begin_time=1643612152
|
555 |
-
# Accounting: end_time=1643612955
|
556 |
-
# Accounting: time=803 threads=1
|
557 |
-
# Finished at Mon Jan 31 02:09:15 EST 2022 with status 0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/train.acc.ave.pth
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
train.acc.ave_5best.pth
|
|
|
|
exp/asr_train_asr_raw_en_word/train.acc.ave_1best.pth
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
2epoch.pth
|
|
|
|
exp/asr_train_asr_raw_en_word/train.acc.ave_5best.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:b6856ac64a7105bd9c0809ace8b562e044d2375eab41818f8c205a5cafefab1c
|
3 |
-
size 110483405
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/train.acc.best.pth
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
195epoch.pth
|
|
|
|
exp/asr_train_asr_raw_en_word/train.log
DELETED
The diff for this file is too large to render.
See raw diff
|
|
exp/asr_train_asr_raw_en_word/train.loss.ave.pth
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
train.loss.ave_5best.pth
|
|
|
|
exp/asr_train_asr_raw_en_word/train.loss.ave_1best.pth
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
2epoch.pth
|
|
|
|
exp/asr_train_asr_raw_en_word/train.loss.ave_5best.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:b49bebd712995835c9df6a9b32244362261deef949cde007de65204d0ea1473b
|
3 |
-
size 110483405
|
|
|
|
|
|
|
|
exp/asr_train_asr_raw_en_word/train.loss.best.pth
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
200epoch.pth
|
|
|
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/RESULTS.md
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.100ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.101ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.102ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.103ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.104ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.105ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.106ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.107ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.108ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.109ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.10ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.110ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.111ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.112ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.113ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.114ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.115ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.116ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.117ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.118ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.119ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.11ep.png
RENAMED
File without changes
|
exp/{asr_train_asr_raw_en_word β asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.120ep.png
RENAMED
File without changes
|