Replies: 2 comments 6 replies
-
Sure, this would be cool addition. It should be relatively easy to add. I was thinking in breaking the generated text into sentences and passing them to the TTS. |
Beta Was this translation helpful? Give feedback.
3 replies
-
I just stumbled upon this and am blown away by the quality: https://catid.io/posts/tts/ |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Currently, the
talk-llama
demo waits for the model's output to finish generating before displaying it and sending it to the TTS.Would it be useful to have the TTS run on a separate thread, and queue up partial responses as they're generated? E.g. for a response that goes:
We could have the first line, or even fragment "The meaning of life is subjective" being read aloud, and during this time push the rest of the response to the queue. At the same time, we display the response as a stream instead of wait for it to finish.
I've done a quick and loose implementation of this and voice chatting with the bot seems more prompt and fun.
Beta Was this translation helpful? Give feedback.
All reactions