Commit
•
75a9512
1
Parent(s):
d7a0e80
Saving train state of step 500
Browse files
checkpoint-500-epoch-6/optimizer.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3652763351
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4320d7a26b7d96a0c047dc0203176c69c66cc2ebf822d3da84e9f65cd71dda9
|
3 |
size 3652763351
|
checkpoint-500-epoch-6/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2588462170
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e0f6ba977980b21a89cb91f622f81621c81c1a61d8082dcda5b0e182daab1a4d
|
3 |
size 2588462170
|
checkpoint-500-epoch-6/random_states_0.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14344
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d5d5be8b16873cb20ec804deed210623d59e4380818972742d65841aa4d14ce
|
3 |
size 14344
|
checkpoint-500-epoch-6/scheduler.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1000
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a6db3feeafb620e767981163694df9114aa7589fa6f6c074e43f210e5e85019c
|
3 |
size 1000
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"ParlerTTSForConditionalGeneration"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/fsx/yoach/tmp/artefacts/training-400M-punctuated-v2/",
|
3 |
"architectures": [
|
4 |
"ParlerTTSForConditionalGeneration"
|
5 |
],
|
run.sh
CHANGED
@@ -29,11 +29,11 @@ python run_parler_tts_training.py \
|
|
29 |
--gradient_accumulation_steps 4 \
|
30 |
--gradient_checkpointing true \
|
31 |
--per_device_train_batch_size 32 \
|
32 |
-
--learning_rate
|
33 |
--adam_beta1 0.9 \
|
34 |
--adam_beta2 0.99 \
|
35 |
--weight_decay 0.01 \
|
36 |
-
--warmup_steps
|
37 |
--logging_steps 2 \
|
38 |
--freeze_text_encoder true \
|
39 |
--audio_encoder_per_device_batch_size 4 \
|
|
|
29 |
--gradient_accumulation_steps 4 \
|
30 |
--gradient_checkpointing true \
|
31 |
--per_device_train_batch_size 32 \
|
32 |
+
--learning_rate 8e-5 \
|
33 |
--adam_beta1 0.9 \
|
34 |
--adam_beta2 0.99 \
|
35 |
--weight_decay 0.01 \
|
36 |
+
--warmup_steps 250 \
|
37 |
--logging_steps 2 \
|
38 |
--freeze_text_encoder true \
|
39 |
--audio_encoder_per_device_batch_size 4 \
|