AntonioMS's picture
add preprocessing
9382c4c
|
raw
history blame
5.07 kB
metadata
language: gl
license: apache-2.0
datasets:
  - CRPIH_UVigo-GL-Voices/Sabela
tags:
  - TTS
  - speech-synthesis
  - Galician
  - female-speaker
  - VITS
  - coqui.ai

Celtia: Nos Project's Galician TTS Model

Model description

This model was trained from scratch using the Coqui TTS Python library on the corpus Nos_Celtia-GL.

A live inference demo can be found in our official page, here.

This model was trained using graphemes. A preprocessing with the Cotovía tool is needed for the input text.

Intended uses and limitations

You can use this model to generate synthetic speech in Galician.

How to use

Usage

Cotovía preprocessor

To generate fonectic transcriptions, the Cotovía tool is needed. The tool can be downloaded from the SourceForge website. The required debian packages are cotovia_0.5_amd64.deb and cotovia-lang-gl_0.5_all.deb, that can be installed with the following commands:

sudo dpkg -i cotovia_0.5_amd64.deb
sudo dpkg -i cotovia-lang-gl_0.5_all.deb

The tool can be used to generate the phonetic transcription of the text. The following command can be used to generate the phonetic transcription of a text string:

echo "Era unha avioneta... O piloto era pequeno, que se chega a ser dos grandes, tómbate!" | cotovia -p -n -S | iconv -f iso88591 -t utf8

The output of the command is the phonetic transcription of the input text. This string may be used in the inference part, as shown next.

Required libraries:

pip install TTS

Synthesize speech using python and the script preprocess.py, avaliable in this repository:

python preprocess.py text model_path config_path

This script takes a text input, preprocesses it with the cotovia tool, synthesizes speech from the preprocessed text, and saves the output as a .wav file.

Training

Hyperparameter

The model is based on VITS proposed by Kim et al. The following hyperparameters were set in the coqui framework.

Hyperparameter Value
Model vits
Batch Size 26
Eval Batch Size 16
Mixed Precision true
Window Length 1024
Hop Length 256
FTT size 1024
Num Mels 80
Phonemizer null
Phoneme Lenguage en-us
Text Cleaners multilingual_cleaners
Formatter nos_fonemas
Optimizer adam
Adam betas (0.8, 0.99)
Adam eps 1e-09
Adam weight decay 0.01
Learning Rate Gen 0.0002
Lr. schedurer Gen ExponentialLR
Lr. schedurer Gamma Gen 0.999875
Learning Rate Disc 0.0002
Lr. schedurer Disc ExponentialLR
Lr. schedurer Gamma Disc 0.999875

The model was trained for 457900 steps.

The nos_fonemas formatter is a modification of the LJSpeech formatter with one extra column for the normalized input (extended numbers and acronyms).

Additional information

Authors

Carmen Magariños

Contact information

For further information, send an email to proxecto.nos@usc.gal

Licensing Information

Apache License, Version 2.0

Funding

This research was funded by “The Nós project: Galician in the society and economy of Artificial Intelligence”, resulting from the agreement 2021-CP080 between the Xunta de Galicia and the University of Santiago de Compostela, and thanks to the Investigo program, within the National Recovery, Transformation and Resilience Plan, within the framework of the European Recovery Fund (NextGenerationEU).

Citation information

If you use this model, please cite as follows:

Magariños, Carmen. 2023. Nos_TTS-celtia-vits-graphemes. URL: https://huggingface.co/proxectonos/Nos_TTS-celtia-vits-graphemes