123B Devstral 2 and 24B Small 2 (2 variants; dense)
Modified MIT (Devstral 2) / Apache 2.0 (Small 2)
Next-generation agentic coding model family; 256K context; SOTA open-weight on SWE-bench Verified (72.2%, huge jump from original Devstral's 46.8%); 7x more cost-efficient than Claude Sonnet for real-world coding tasks; business context awareness similar to Le Chat's conversational memory; ships with Mistral Vibe CLI for natural language code automation and vibe coding workflows; Devstral 2 requires 4+ H100 GPUs; Small 2 runs on consumer hardware (single GPU) and supports image inputs for multimodal agents; integrated with Kilo Code, Cline, and Zed IDE; currently free API access.