Blog Post

Azure AI Foundry Blog
2 MIN READ

Fine-tuning gpt-oss-20b Now Available on Managed Compute

NandiniMuralidharan's avatar
Aug 28, 2025

Start fine-tuning your own multilingual gpt-oss-20B model today.

Earlier this month, we made available OpenAI’s open‑source model gpt‑oss on Azure AI Foundry and Windows AI Foundry. Today, you can fine-tune gpt‑oss‑20b using Managed Compute on Azure — available in preview and accessible via notebook.  

As open-weight models gain traction, developers are increasingly looking for ways to customize and deploy them efficiently. Large language models like gpt-oss-20B are revolutionizing AI applications. But what if you could fine-tune them for your domain or language with just a single H100? 

Why Fine-tune gpt-oss-20B? 

gpt-oss-20B is a powerful, open-source model. Fine-tuning enables: 

  • Domain-specific adaptation (e.g., customer support, education) 
  • Multilingual reasoning (English, Spanish, French, Italian, German) 
  • Efficient training with LoRA adapters—only 1% of parameters are updated 

With just a few clicks, you can fine-tune gpt-oss-20B for multi-lingual chain-of-thought reasoning. The platform handles distributed training, resource scaling, and output packaging. 

 

Image – Select gpt-oss-20b

Fine-tuning Use Cases & Applications 

From multilingual chatbots and assistants that deliver seamless global interactions, to domain-specific reasoning tools tailored for legal, financial, and technical decision-making, organizations can unlock new efficiencies and insights. 

Category 

Use Case

Multilingual Chatbots & Assistants 

Real-time support, translation, and conversational agents across languages 

Domain-Specific Reasoning Tools 

Legal, medical, financial, and technical copilots for reasoning and decision support 

Global Customer Support 

Automated ticket triage, sentiment analysis, and escalation routing 

Developer & Data Tools 

Code generation, tool-use agents, and data science copilots 

Why use Managed Compute? 

Managed Compute is a deployment option within Azure AI Foundry Models that lets you run large language models (LLMs), SLMs, HuggingFace models and custom models fully hosted on Azure infrastructure. Azure Managed Compute is a powerful deployment option for models not available via standard (pay-go) endpoints. It gives you:  

  • Custom model support: Deploy open-source or third-party models  
  • GPU flexibility: tested on Standard_NC96ads_A100_v4 and Standard_ND96isr_H100_v5 
  • Detailed control: Configure inference servers, protocols, and advanced settings  
  • Full integration: Works with Azure ML SDK, CLI, Prompt Flow, and REST APIs  
  • Enterprise-ready: Supports VNet, private endpoints, quotas, and scaling policies 

This setup is ideal for developers who want to fine-tune and deploy models without managing infrastructure.  

Getting Started

Currently, fine-tuning gpt-oss-20B is available only via notebook with Azure ML. There’s no UI support in the Azure yet (UI enablement coming soon), so developers work directly with code in this scenario. This notebook walks through: 

  • Setting up Azure ML compute 
  • Configuring LoRA adapters 
  • Training with multilingual datasets 
  • Deploying for chat completion tasks 
➡️ Get started with this GitHub Notebook 

Learn More with these Resources  

🧠 Get started with Azure AI Fine-tuning on Microsoft Learn 

👩‍💻 Learn more about gpt-oss on Azure AI Foundry and Windows AI Foundry 

👋 Continue the conversation on Discord   

 

Updated Sep 01, 2025
Version 10.0
No CommentsBe the first to comment