This repo was forked from NVIDIA/Tacotron2 for inference test only (not for training).
Because I didn't know flask well, I forked CodeDem/flask-musing-streaming.
If you want to test NVIDIA Tacotron2 models in jupyter notebook
, you better try inference model NVIDIA/Tacotron2 .
-
Install PyTorch 1.0 (You Need NVIDIA CUDA GPUs!)
-
pip install -r requirement.txt
-
clone this repo:
https://github.com/NVIDIA/waveglow.git
or
git submodule init; git submodule update
-
you may need models tacotron2, waveglow both :
-
NVIDIA/Tacotron2's model for inference demo: Tacotron 2 , WaveGlow
-
or My trained models:
Tacotron2: English_90k_steps(ljspeech dataset), Korean_162k_steps(kss dataset)
Waveglow: waveglow_152k_steps using Korean dataset
-
python app.py
or You can test tts on console: python console_test.py
in config.json
, you can change models' path.
You can see Warning! Decoder Max
on console.
In this case, your synthesized audio will have 11 seconds length and weired sounds.
This problems many happen in my korean trained model, but hardly happen in my english trained model.
I can't find any difference from synthesized audio between waveglow_256channels.pt
(waveglow demo) and my waveglow_152k
.