We are excited to announce the public preview of Azure OpenAI Global Batch offering, designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, with target 24-hour turnaround time, at 50% less cost than global standard.
This is a quote from our customer:
"Ontada is at the unique position of serving providers, patients and life science partners with data-driven insights. We leverage the Azure OpenAI batch API to process tens of millions of unstructured documents efficiently, enhancing our ability to extract valuable clinical information. What would have taken months to process now takes just a week. This significantly improves evidence-based medicine practice and accelerates life science product R&D. Partnering with Microsoft, we are advancing AI-driven oncology research, aiming for breakthroughs in personalized cancer care and drug development."
Sagran Moodley, Chief Innovation and Technology Officer, Ontada
Why Azure OpenAI Global Batch?
Supported Models
The following models currently support the global batch:
Model |
Supported Versions |
gpt4-o |
2024-05-13 |
gpt-4o-mini |
2024-07-18 |
gpt-4 |
turbo-2024-04-09 |
gpt-4 |
0613 |
gpt-35-turbo |
0125 |
gpt-35-turbo |
1106 |
gpt-35-turbo |
0613 |
For the most up-to-date information on regions and models, please refer to our models page.
Key Use Cases
The Azure OpenAI Batch API opens up new possibilities across various industries and applications:
Getting Started
Ready to try Azure OpenAI Batch API? Take it for a spin here.
Learn more
Using images in your batch input
Default batch token quota allocation and requesting increase
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.