8B Nano, 49B Super, 253B Ultra (3 sizes; Llama 4-based)
Apache 2.0
Open reasoning family on Llama 4; Nano (8B) for PC/edge, Super (49B) for single GPU with best throughput, Ultra (253B) for maximum agentic accuracy; 20% improved accuracy vs base models, 5x faster inference; excels in multi-agent collaboration, workflow automation, and domain-specific fine-tuning; compute-efficient for enterprise agents.