Available Models
Model specifications, capabilities, and selection guidance for the MiniMax-M2 family
Available Models
MiniMax-M2 offers two production-ready models optimized for different development workflows. Both models share the same sparse MoE architecture with 230B total parameters and 10B active per token.
Quick Comparison
| Feature | MiniMax-M2 | MiniMax-M2.1 |
|---|---|---|
| Release | Oct 2025 | Dec 2025 |
| Core Focus | Agent workflows & general coding | Multi-language development & native apps |
| Primary Languages | Python, JavaScript (strong) | Rust, Java, Golang, C++, Kotlin, Obj-C, TS/JS (enhanced) |
| Native Mobile | Good | Excellent (Android/iOS) |
| Response Style | Fast, ~100 tokens/s | More concise, lower token consumption |
| Design Understanding | Standard | Enhanced aesthetic comprehension |
| VIBE Score | Not benchmarked | 88.6 (Web: 91.5, Android: 89.7) |
| Context Window | 128K tokens | 128K tokens |
| Streaming | ✅ Yes | ✅ Yes |
| Thinking Traces | ✅ Yes | ✅ Yes (more efficient) |
| Pricing | $0.50 / $1.50 per 1M tokens | $0.50 / $1.50 per 1M tokens |
MiniMax-M2
"A model born for Agents and code" — Stable, cost-optimized, battle-tested.
Model ID: MiniMax-M2
Released: October 27, 2025
Strengths
- ✅ Agent-First Design: Reliable coordination of Shell, Browser, Python interpreter, and MCP tools
- ✅ Tool Integration: Industry-leading tool-use performance, matches top overseas models
- ✅ Deep Search: Near best-in-class research and information retrieval capabilities
- ✅ Production Stability: Internally validated by MiniMax engineering teams before release
- ✅ Fast Inference: ~100 tokens/second throughput for responsive interactions
- ✅ Cost Efficiency: Balanced intelligence, speed, and affordability
Ideal Use Cases
- Python and JavaScript-heavy agent workflows
- Cost-sensitive production deployments
- Shell automation and CLI tool orchestration
- Browser automation and web scraping tasks
- General-purpose coding assistance
- Teams prioritizing inference speed over specialized language support
Supported Platforms
- OpenAI-compatible Chat Completions API
- Claude-compatible Messages API
- Claude Code, Cline, Kilo Code, RooCode, BlackBox
- Factory AI's Droid
MiniMax-M2.1
"Real-world complex tasks, multilingual excellence" — Enhanced quality, broader language coverage.
Model ID: MiniMax-M2.1
Released: December 23, 2025
Major Improvements Over M2
🌍 Systematically Enhanced Multi-Language Programming
Strong performance across Rust, Java, Golang, C++, Kotlin, Objective-C, TypeScript, JavaScript, and more. Achieves industry-leading multilingual capabilities, surpassing Claude Sonnet 4.5 and approaching Claude Opus 4.5 in multi-language scenarios.
⚡ Efficiency Revolution
Generates more concise responses and thought chains, significantly reducing token consumption while improving response speed. Delivers higher quality with lower operational costs.
📱 Native Mobile Development
Substantially strengthened native Android and iOS development capabilities. Enhanced understanding of platform-specific patterns, frameworks, and best practices.
🎨 Design & Aesthetic Comprehension
Improved aesthetic judgment for web applications, better alignment with modern design principles and user experience standards.
🏢 Office & Composite Workflows
Enhanced handling of composite instruction constraints for office scenarios, supporting complex multi-step business processes.
Benchmark Highlights
- VIBE Aggregate: 88.6 (Visual & Interactive Benchmark for Evaluation)
- VIBE-Web: 91.5
- VIBE-Android: 89.7
- SWE-bench Verified: Strong framework generalization across diverse agent tooling
- Multilingual Scenarios: Outperforms Claude Sonnet 4.5, approaches Claude Opus 4.5
Ideal Use Cases
- Multi-language codebases (Rust, Java, Golang, C++, etc.)
- Native Android and iOS application development
- Web applications requiring strong design sensibility
- Complex office automation and business workflows
- Teams prioritizing output quality and token efficiency
- Projects requiring broad framework and language support
Supported Platforms
- OpenAI-compatible Chat Completions API
- Claude-compatible Messages API
- Claude Code, Cline, Kilo Code, RooCode, BlackBox
- Factory AI's Droid
- GUI automation via text commands
How to Choose
Choose MiniMax-M2 if you:
- Primarily work with Python and JavaScript
- Need fast, stable agent workflows with proven reliability
- Prioritize inference speed for interactive applications
- Run cost-sensitive production workloads
Choose MiniMax-M2.1 if you:
- Develop in Rust, Java, Golang, C++, Kotlin, or Objective-C
- Build native Android or iOS applications
- Require higher output quality with lower token costs
- Work on web applications where design matters
- Handle complex multi-step business processes
How to Specify Models
Pass the model field in your API request:
Chat Completions (OpenAI format)
Messages API (Claude format)
Default Behavior
If the model field is omitted, requests default to MiniMax-M2.1 for enhanced multi-language capabilities and improved efficiency.
Pricing
Both models use identical pricing:
| Token Type | Rate |
|---|---|
| Input tokens | $0.50 per million |
| Output tokens | $1.50 per million |
Usage charges appear in your billing dashboard with per-model breakdowns, allowing you to track consumption across M2 and M2.1 independently.
Recommended Parameters
For optimal performance with both models:
Adjust temperature downward (0.3–0.7) for deterministic outputs in compliance-sensitive or production-critical scenarios.
Access Methods
API Integration:
- OpenAI SDK: Set
base_urltohttps://minimax-m2.com/api/v1/ - Anthropic SDK: Set
base_urltohttps://minimax-m2.com/api/
Direct API:
Developer Tools:
Migration Notes
Switching from M2 to M2.1
- Update the
modelfield: Change"MiniMax-M2"to"MiniMax-M2.1"in your request body - No API changes required: Both models support identical endpoint contracts
- Monitor token usage: M2.1's concise responses may reduce your token consumption
- Test prompts: M2.1's enhanced capabilities may produce different (often better) outputs for the same inputs
Backward Compatibility
All M2 prompts, system messages, and tool configurations work unchanged with M2.1. You can switch models without modifying application logic.
Official Resources
- M2 Announcement: https://www.minimax.io/news/minimax-m2
- M2.1 Announcement: https://www.minimax.io/news/minimax-m21
- API Documentation: https://platform.minimax.io/docs/guides/text-generation
- Agent Product: https://agent.minimax.io/
- Open-source Weights: HuggingFace (MiniMaxAI organization)