Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] existing streaming latency is still takes time, #417

Closed
kunci115 opened this issue Jul 23, 2024 · 6 comments
Closed

[Feature] existing streaming latency is still takes time, #417

kunci115 opened this issue Jul 23, 2024 · 6 comments
Labels
enhancement New feature or request stale

Comments

@kunci115
Copy link

kunci115 commented Jul 23, 2024

streaming in 4090 tooks more than 2 second depend on length of token, is there a way to yield it/return while the engine still generating?

@kunci115 kunci115 added the enhancement New feature or request label Jul 23, 2024
@Stardust-minus
Copy link
Member

PR Welcome

@PoTaTo-Mika
Copy link
Collaborator

Please compile the model, or try the quantized version.

@kunci115
Copy link
Author

kunci115 commented Jul 24, 2024

@PoTaTo-Mika what do you mean by compile the model ? also how to do quantized version? since I only do steps for inference in english documentation https://speech.fish.audio/en/inference/#2-create-a-directory-structure-similar-to-the-following-within-the-ref_data-folder

@PoTaTo-Mika
Copy link
Collaborator

there's a python file called quantize.py, you can view the file and choose to quantize.
image

@kunci115
Copy link
Author

there's a python file called quantize.py, you can view the file and choose to quantize. image

its creating me a folder quantized version of the model now, just run it like previous run with that checkpoints model? still got the same latency

Copy link
Contributor

This issue is stale because it has been open for 30 days with no activity.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request stale
Projects
None yet
Development

No branches or pull requests

4 participants