Yep, we're getting true streaming responses - we're streaming an OpenAI LLM response and can see the results coming back a word or two at a time - make sure your client is handling the streaming response, not everything does! Still struggling to get this deployed to always-on premium though!
Published May 29, 2024
Version 1.0