We’re thrilled to announce that MiniMax-M2, the latest breakthrough from MiniMax, is now available in Azure AI Foundry through Hugging Face. Built for developers, this model advances capabilities for what’s possible in coding, multi-turn reasoning, and agentic workflows—while delivering enhanced efficiency and scalability.
What makes MiniMax-M2 different?
MiniMax-M2 isn’t just another large language model—it’s a 230B-parameter Mixture of Experts (MoE) architecture that activates 10B parameters per task, ensuring better performance at a lower cost. This design enables:
- Enhanced efficiency: Achieve top-tier results up to 8% of the cost of comparable models.
- Increased context handling: With an industry-leading 204K token context window and 131K output capacity, MiniMax-M2 can process entire codebases, multi-file projects, and long-form documentation without losing coherence.
- Commercial ready: Released under Apache 2.0, MiniMax-M2 is open-source and ready to deploy into your workflow.
The model was ranked #5 overall on the Artificial Analysis Intelligence Index, making MiniMax-M2 one of the highest-ranked open-source model globally, outperforming many proprietary systems in reasoning, coding, and language understanding. For organizations seeking high-throughput, low-latency deployments, MiniMax-M2 runs seamlessly on an 8xH100 setup using vLLM, making it both powerful and practical.
Figure 1. The MiniMax-M2's abilities are very close to similar, top performing models like DeepSeek-V3.3, GLM-4.6, and Gemini 2.5 Pro.The graphic above compares MiniMax-M2’s performance across multiple industry-standard benchmarks against leading models like DeepSeek-V3.2, GLM-4.6, and Gemini 2.5 Pro. While proprietary models such as GPT-5 (thinking) and Claude Sonnet 4.5 remain strong in certain areas, MiniMax-M2 delivers competitive results as an open-source solution, offering enterprise-grade performance for organizations seeking high-quality AI without compromising scalability or flexibility.
Why it matters for developers
MiniMax-M2 is built for modern development workflows. Whether you’re generating production-ready code, automating agentic tasks, or managing large-scale projects, this model delivers accuracy, speed, and flexibility while keeping infrastructure costs in check.
- Mixture of Experts Architecture: 230B total parameters, 10B active per task for cost-effective scalability.
- Ultra-Large Context Window: 204K tokens for comprehensive project understanding.
- Advanced Coding Intelligence: Optimized for code generation, debugging, multi-file editing, and test-driven development.
- Agentic Workflow Support: Handles complex tool integrations and multi-step problem-solving with ease.
- Open Source Freedom: Apache 2.0 license for commercial use.
MiniMax-M2 can support finance and legal workflows by automating document-heavy tasks. In finance, it could help generate audit reports, investment summaries, and portfolio analyses by processing large datasets and regulatory guidelines in a single pass, which can improve accuracy and reduce manual effort. In legal, it could assist with case law research by summarizing extensive statutes and precedents, extracting relevant insights, and providing context-specific recommendations. With its large context window and reasoning capabilities, MiniMax-M2 can enable faster, more efficient handling of complex information, allowing professionals to focus on higher-value activities.
Get started today
MiniMax-M2 is now live in Azure AI Foundry, explore its capabilities and try it today.