You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It's not currently supported on the server, but you may break up your input text into multiple smaller inputs and make an API call for each input separately.
I'm doing this for the voice chat feature. The speech is generated from an LLM reply in sentence chunks (rather than generating the speech from the full reply). Take a look at the SentenceChunker class, and it's usage within the code.
It is a bit slow on my machine to wait for the entire response to generate.
Is streaming technically possible?
Thanks
The text was updated successfully, but these errors were encountered: