← Back to models

Llama Nemotron

NVIDIA (USA) March 18 2025

Parameters

8B Nano, 49B Super, 253B Ultra (3 sizes; Llama 4-based)

License

Apache 2.0

Key Features

Open reasoning family on Llama 4; Nano (8B) for PC/edge, Super (49B) for single GPU with best throughput, Ultra (253B) for maximum agentic accuracy; 20% improved accuracy vs base models, 5x faster inference; excels in multi-agent collaboration, workflow automation, and domain-specific fine-tuning; compute-efficient for enterprise agents.

Paper / Source

https://arxiv.org/abs/2505.00949