Hugging Face BlogAI & LLMs

NVIDIA's New 4-Billion-Parameter Model Brings High-Performance AI to the Edge

NVIDIA has released the Nemotron 3 Nano 4B, a small language model engineered for deployment directly on devices, from workstations to embedded systems. The model represents a strategic move to put capable AI into environments where data privacy, low latency, and cost are primary concerns.

Built with a hybrid Mamba-Transformer architecture, the 4-billion-parameter model is designed to operate efficiently on NVIDIA's edge platforms like Jetson Orin Nano and DGX Spark, as well as on consumer RTX GPUs. Benchmarks on an RTX 4070 show it leads its class in instruction following, gaming agency, and memory efficiency. It also exhibits strong tool-use performance and reduced hallucination.

The model was created using NVIDIA's Nemotron Elastic framework, which pruned and distilled it from the larger Nemotron Nano 9B v2. This process involved a trained router that intelligently reduced the model's depth, width, and components to hit the 4B parameter target, followed by retraining. The result is a model that maintains much of its predecessor's reasoning at a fraction of the size.

For engineers, the practical release formats are key. The model is offered in FP8 and 4-bit GGUF quantized versions, which recover full benchmark accuracy while significantly boosting speed. On a Jetson Orin Nano, the quantized model generates text roughly twice as fast as the 9B version it was derived from.

Available now on Hugging Face, the Nemotron 3 Nano 4B supports major inference engines like Transformers, vLLM, and Llama.cpp, providing flexibility for integration into local applications, from conversational agents to robotic systems.

Source: Hugging Face Blog

← Back to News