Event banner
AMA: GPT-4o Audio model revolutionizes your Copilot and other AI applications
Event Ended
Wednesday, Oct 09, 2024, 09:00 AM PDTEvent details
Unlock the potential of your applications with the latest GPT-4o-realtime API with Audio, now available on Azure on October 1st, 2024. Join us to explore how this model, integrated as part of the new...
EricStarker
Updated Dec 27, 2024
Travis_Wilson_MSFT
Microsoft
We don't yet have higher-level abstractions for Teams specifically, but between OpenAI and Microsoft we've started some client library coverage to expose the new capabilities of the gpt-4o-realtime-preview model and the /realtime API:
- The OpenAI .NET SDK (https://github.com/openai/openai-dotnet) (as well as the AOAI companion library, Azure.AI.OpenAI) already has early support for a client integrated into 2.1.0-beta.1
- Python has an early standalone client we're iterating on at aoai-realtime-audio-sdk/python at main · Azure-Samples/aoai-realtime-audio-sdk (github.com)
- JavaScript has an early standalone library: openai/openai-realtime-api-beta: Node.js + JavaScript reference client for the Realtime API (beta) (github.com) and we also one at aoai-realtime-audio-sdk/javascript at main · Azure-Samples/aoai-realtime-audio-sdk (github.com)
We've already seen developers prototype applications with multiple agents talking to people (and each other!) using the /realtime capabilities and the results are very cool. Very possible with the tools we have today!
scastle15
Oct 09, 2024Copper Contributor
Excellent, thanks for the reply, I'll take a look.