Blog Post

AI - Azure AI services Blog
2 MIN READ

Announcing Azure OpenAI Global Batch General availability: At scale processing with 50% less cost!

Sethu_Raman's avatar
Sethu_Raman
Microsoft
Oct 23, 2024

We are excited to announce the General Availability of Azure OpenAI Global Batch offering, designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, a 24-hour turnaround time, at 50% less cost than global standard.

This is a quote from one of our customers:

 

 "Ontada is at the unique position of serving providers, patients and life science partners with data-driven insights. We leverage the Azure OpenAI batch API to process tens of millions of unstructured documents efficiently, enhancing our ability to extract valuable clinical information. What would have taken months to process now takes just a week. This significantly improves evidence-based medicine practice and accelerates life science product R&D. Partnering with Microsoft, we are advancing AI-driven oncology research, aiming for breakthroughs in personalized cancer care and drug development."

 Sagran Moodley, Chief Innovation and Technology Officer, Ontada

 

 Why Azure OpenAI Global Batch?

  • Benefit 50% lower costs, enabling you to either introduce new workloads or run existing workloads more frequently, thereby increasing overall business value.
  • Efficiently handle large-scale workloads that would be impractical to process in real-time, significantly reducing processing times.
  • Minimize engineering overhead for job management with a high resource quota, allowing you to queue and process gigabytes of data with ease. Substantially high quotas for batch.

 

New feature: Dynamic Quota - no more quota exceed errors!

When you enable dynamic quota for your deployments, you can opportunistically take advantage of more quota when extra capacity is available.!

 

 

Supported Models

The following models currently support the global batch:

Model

Supported Versions

gpt4o

2024-08-06

gpt4o

2024-05-13

gpt-4o-mini

2024-07-18

gpt-4

turbo-2024-04-09

gpt-4

0613

gpt-35-turbo

0125

gpt-35-turbo

1106

gpt-35-turbo

0613

 

For the most up-to-date information on regions and models, please refer to our models page.

 

Key Use Cases

The Azure OpenAI Batch API opens up new possibilities across various industries and applications:

  1. Large-Scale Data Processing: Quickly analyze extensive datasets in parallel, enabling faster decision-making and insights.
  2. Content Generation: Automate the creation of large volumes of text, such as product descriptions, articles, and more.
  3. Document Review and Summarization: Streamline the review and summarization of lengthy documents, saving valuable time and resources.
  4. Customer Support Automation: Enhance customer support by handling numerous queries simultaneously, ensuring faster and more efficient responses.
  5. Data Extraction and Analysis: Extract and analyze information from vast amounts of unstructured data, unlocking valuable insights.
  6. Natural Language Processing (NLP) Tasks: Perform sentiment analysis, translation, and other NLP tasks on large datasets effortlessly.
  7. Marketing and Personalization: Generate personalized content and recommendations at-scale, improving engagement and customer satisfaction.

Getting Started

Ready to try Azure OpenAI Batch API? Take it for a spin here.

Learn more

Using images in your batch input

Default batch token quota allocation and requesting increase

Supported regions

Published Oct 23, 2024
Version 1.0