671B / 37B (2 variants: standard
MIT
Speciale; MoE),First DeepSeek to integrate thinking into tool-use; hybrid thinking/non-thinking modes; standard version reaches GPT-5 level (93.1% AIME, 92.5% HMMT); Speciale variant for extreme reasoning with gold medals in IMO/CMO/ICPC/IOI 2025 (99.2% HMMT, 35/42 IMO); combines theorem-proving from Math-V2; massive agent training (1,800+ environments).