MiniMax M2.1, now open source and available on Hugging Face, is setting new standards in real-world development and agent applications by achieving state-of-the-art (SOTA) performance on coding benchmarks such as SWE, VIBE, and Multi-SWE. Demonstrating superior capabilities, it surpasses notable models like Gemini 3 Pro and Claude Sonnet 4.5. With a configuration of 10 billion active parameters and a total of 230 billion parameters in a Mixture of Experts (MoE) architecture, MiniMax M2.1 offers significant advancements in computational efficiency and effectiveness for developers and AI agents. This matters because it provides the AI community with a powerful, open-source tool that enhances coding efficiency and innovation in AI applications.
Read Full Article: MiniMax M2.1: Open Source SOTA for Dev & Agents