Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Help needed w.r.t. inference #90

Open
nirmeshshah opened this issue Dec 15, 2020 · 0 comments
Open

Help needed w.r.t. inference #90

nirmeshshah opened this issue Dec 15, 2020 · 0 comments

Comments

@nirmeshshah
Copy link

nirmeshshah commented Dec 15, 2020

Hi I have few doubts ,

  1. Is example1.wav the reference audio file whose style has to be captured while synthesizing samples in inference.ipynb? Do I need to have text and corresponding wav file for inference in mellotron well in advance ?. Usually, I have text which I want to synthesized and reference audio of completely different utterance whose style has to be captured. I am unable to map this with the existing inference.ipynb. Can anyone please give some more clarity on this ?

  2. How to run this model as a standalone TTS ?

  3. If I have trained my model on single speaker data how can I update Define Speaker Set section in inference.ipynb as it seems it is given for multispeaker only.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant