106B / 12B (MoE)
MIT
Post-trained on GLM-4.5-Air-Base using SFT and RL; trained on 512 H200 GPUs with prime-rl framework; SOTA performance for size on math (90.8% AIME 2024), code, and reasoning; fully open-sourced with complete RL stack and environments.