We are excited to announce the General Availability of Azure OpenAI Global Batch offering, designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, a 24-hour turnaround time, at 50% less cost than global standard.
This is a quote from one of our customers:
"Ontada is at the unique position of serving providers, patients and life science partners with data-driven insights. We leverage the Azure OpenAI batch API to process tens of millions of unstructured documents efficiently, enhancing our ability to extract valuable clinical information. What would have taken months to process now takes just a week. This significantly improves evidence-based medicine practice and accelerates life science product R&D. Partnering with Microsoft, we are advancing AI-driven oncology research, aiming for breakthroughs in personalized cancer care and drug development."
Sagran Moodley, Chief Innovation and Technology Officer, Ontada
Why Azure OpenAI Global Batch?
- Benefit 50% lower costs, enabling you to either introduce new workloads or run existing workloads more frequently, thereby increasing overall business value.
- Efficiently handle large-scale workloads that would be impractical to process in real-time, significantly reducing processing times.
- Minimize engineering overhead for job management with a high resource quota, allowing you to queue and process gigabytes of data with ease. Substantially high quotas for batch.
New feature: Dynamic Quota - no more quota exceed errors!
When you enable dynamic quota for your deployments, you can opportunistically take advantage of more quota when extra capacity is available.!
Supported Models
The following models currently support the global batch:
Model |
Supported Versions |
gpt4o |
2024-08-06 |
gpt4o |
2024-05-13 |
gpt-4o-mini |
2024-07-18 |
gpt-4 |
turbo-2024-04-09 |
gpt-4 |
0613 |
gpt-35-turbo |
0125 |
gpt-35-turbo |
1106 |
gpt-35-turbo |
0613 |
For the most up-to-date information on regions and models, please refer to our models page.
Key Use Cases
The Azure OpenAI Batch API opens up new possibilities across various industries and applications:
- Large-Scale Data Processing: Quickly analyze extensive datasets in parallel, enabling faster decision-making and insights.
- Content Generation: Automate the creation of large volumes of text, such as product descriptions, articles, and more.
- Document Review and Summarization: Streamline the review and summarization of lengthy documents, saving valuable time and resources.
- Customer Support Automation: Enhance customer support by handling numerous queries simultaneously, ensuring faster and more efficient responses.
- Data Extraction and Analysis: Extract and analyze information from vast amounts of unstructured data, unlocking valuable insights.
- Natural Language Processing (NLP) Tasks: Perform sentiment analysis, translation, and other NLP tasks on large datasets effortlessly.
- Marketing and Personalization: Generate personalized content and recommendations at-scale, improving engagement and customer satisfaction.
Getting Started
Ready to try Azure OpenAI Batch API? Take it for a spin here.
Learn more
Using images in your batch input
Default batch token quota allocation and requesting increase