-
Notifications
You must be signed in to change notification settings - Fork 90
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
latency is high #29
Comments
已收到。
|
使用fastapi的时候 deepspeed=True未生效。我改了一下代码强制 deepspeed=True时候报以下错误。
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
use the docker "ghcr.io/coqui-ai/xtts-streaming-server", when post request, I get the follow first chunk time:
when I use the local inference.
first chunk: 0.2440338134765625s
The text was updated successfully, but these errors were encountered: