3B / 8B / 14B (3 sizes × 3 variants: Base
Apache 2.0
Instruct, Reasoning; dense),Multimodal edge-optimized family (text + vision); 128K-256K context; single GPU deployment; Base for foundation tasks, Instruct for chat/assistants, Reasoning for complex logic; 14B Reasoning achieves 85% on AIME 2025; can run on laptops/phones/drones; efficient token generation.