MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning, tool use, and multi-step task execution while maintaining low latency and deployment efficiency. The model excels in code generation, multi-file editing, compile-run-fix loops, and test-validated repair, showing strong results on SWE-Bench Verified, Multi-SWE-Bench, and Terminal-Bench. It also performs competitively in agentic evaluations such as BrowseComp and GAIA, effectively handling long-horizon planning, retrieval, and recovery from execution errors. Benchmarked by Artificial Analysis, MiniMax-M2 ranks among the top open-source models for composite intelligence, spanning mathematics, science, and instruction-following. Its small activation footprint enables fast inference, high concurrency, and improved unit economics, making it well-suited for large-scale agents, developer assistants, and reasoning-driven applications that require responsiveness and cost efficiency. Key specifications include a 262K token context window, 4K token max output, and PRO access tier. Pricing is competitive at $0.20/$1.00 per 1M input/output tokens. Capabilities include functions, code, and streaming. MiniMax highly recommends preserving reasoning between turns to avoid performance degradation.
✅ Best For
🚀 Capabilities
❌ Limitations
Specifications
| Provider | minimax |
| Context Window | 262,144 tokens |
| Max Output | 4,096 tokens |
| Minimum Plan | Premium |
Pricing
| Input Price | $0.2000 / 1M tokens |
| Output Price | $1.0000 / 1M tokens |
💡 With PRO subscription, cost is reduced by 20%