-
-
Notifications
You must be signed in to change notification settings - Fork 662
Open
Description
Hello,
I would I have liked to make this an email or a Discussion tab, but neither options were available. Sorry for creating an issue instead.
As part of a recent project I have created, I have implemented streaming inference for Kokoro TTS. You can view the implementation in this web demo.
On CPU, 90% of the inference time is spent on streamable ops, so the Time To First Sound can be decreased heavily. This without any major refactor, retraining or by using sentence splitting.
I'd be happy to build a PR to incorporate the streaming into this repo, but this is quite a bit of additional work given it would be a new API. Therefore I would only do it if there is genuine interest in the PR being merged.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels