shenyunhang commited on
Commit
9fd0f03
·
verified ·
1 Parent(s): e84baf6

Delete vita_tts_ckpt

Browse files
vita_tts_ckpt/codec/final.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c04d618827b0e2778280d7c4701d6c8450c4104de4c1a1dca2b5c3120017c7a
3
- size 253718273
 
 
 
 
vita_tts_ckpt/codec/model.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "resblock": "1",
3
- "num_gpus": 8,
4
- "batch_size": 160,
5
- "learning_rate": 0.0002,
6
- "adam_b1": 0.5,
7
- "adam_b2": 0.9,
8
- "lr_decay": 0.98,
9
- "seed": 1234,
10
-
11
- "upsample_rates": [8,5,5,3],
12
- "upsample_kernel_sizes": [16,11,11,5],
13
- "upsample_initial_channel": 512,
14
- "resblock_kernel_sizes": [3,7,11],
15
- "resblock_dilation_sizes": [[1,3,5], [1,3,5], [1,3,5]],
16
-
17
- "segment_size": 24000,
18
- "num_mels": 80,
19
- "num_freq": 1025,
20
- "n_fft": 1024,
21
- "hop_size": 240,
22
- "win_size": 1024,
23
-
24
- "sampling_rate": 24000,
25
-
26
- "n_code_groups": 1,
27
- "residul_layer": 1,
28
- "n_codes": 1024,
29
- "codebook_loss_lambda": 1.0,
30
- "commitment_loss_lambda": 0.25,
31
- "global_code_num": 8,
32
- "global_feature_conv":[128, 64, 128, 3, 1],
33
- "global_tokens": [473,975,419,219,565,121,550,616],
34
-
35
- "fmin": 0,
36
- "fmax": 8000,
37
- "fmax_for_loss": null,
38
-
39
- "num_workers": 12
40
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
vita_tts_ckpt/decoder/final.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d3274da94685758ce37209759ac7542ea8f2d6b47c00d1cc18c051d0d33cc3e
3
- size 1194900847
 
 
 
 
vita_tts_ckpt/decoder/model.json DELETED
@@ -1,72 +0,0 @@
1
- [
2
- 896,
3
- 1024,
4
- {
5
- "accum_grad": 3,
6
- "char_list": [],
7
- "debugmode": 0,
8
- "encoder_criterion": "ce",
9
- "encoder_drop_rate": 0.1,
10
- "encoder_input_dim": 896,
11
- "encoder_layer_config": "transformer",
12
- "encoder_output_dim": 896,
13
- "encoder_pre_norm_type": "ln",
14
- "encoder_upsample_rate": 9,
15
- "kv_cache_prefix_finetune": 0,
16
- "epochs": 100,
17
- "eps": 1e-08,
18
- "eps_decay": 0.8,
19
- "gpu_id": null,
20
- "gpu_num": 1,
21
- "grad_clip": 5,
22
- "grad_noise": false,
23
- "idim": 896,
24
- "init_lr": 0.0005,
25
- "lsm_weight": 0.0,
26
- "max_batch_size": 25,
27
- "max_duration": 256,
28
- "max_mem": 20000,
29
- "mtlalpha": 0.5,
30
- "n_iter_processes": 8,
31
- "noam_warmup_steps": 4000,
32
- "odim": 1024,
33
- "opt": "noamw",
34
- "rank": 0,
35
- "report_interval_iters": 100,
36
- "resume_trainer": false,
37
- "save_interval_iters": 2000,
38
- "seed": 19832,
39
- "sort_duration": true,
40
- "start_decay_epoch": 5,
41
- "stop_learning_rate": 1e-05,
42
- "sycn_batchnorm": false,
43
- "tensorboard_dir": null,
44
- "train_dtype": "bfloat16",
45
- "transformer_attention_dim": 896,
46
- "transformer_attention_dropout_rate": 0.1,
47
- "transformer_attention_heads": 14,
48
- "transformer_chunk_size": [
49
- 1
50
- ],
51
- "transformer_concat_after": false,
52
- "transformer_dropout_rate": 0.1,
53
- "transformer_dynamic_chunks": false,
54
- "transformer_input_dim": 896,
55
- "transformer_input_layer": "linear",
56
- "transformer_left_chunks": [
57
- -1
58
- ],
59
- "transformer_linear_units": 4864,
60
- "transformer_normalize_before": true,
61
- "transformer_num_blocks": 4,
62
- "transformer_output_dim": 896,
63
- "transformer_pos_enc_class": "rel-enc",
64
- "transformer_positional_dropout_rate": 0.1,
65
- "transformer_positionwise_conv_kernel_size": 1,
66
- "transformer_positionwise_layer_type": "linear",
67
- "use_zero_redun_opt": false,
68
- "verbose": 0,
69
- "weight_decay": 0.05,
70
- "world_size": 1
71
- }
72
- ]