
Nemotron 3 Nano 4B: NVIDIA Edge Model Runs on 8GB
NVIDIA's Nemotron 3 Nano 4B packs a Mamba-dominant hybrid architecture, 262K token context, and 95.4% on MATH500 into a model that fits an 8GB Jetson Orin Nano.

NVIDIA's Nemotron 3 Nano 4B packs a Mamba-dominant hybrid architecture, 262K token context, and 95.4% on MATH500 into a model that fits an 8GB Jetson Orin Nano.

NVIDIA Nemotron 3 Super is a 120B-parameter open model with 12B active at inference, combining Mamba-2, LatentMoE, and Multi-Token Prediction for agentic workloads with a 1M token context window.

NVIDIA releases Nemotron 3 Super, a 120B-parameter open model with only 12B active at inference, combining Mamba-2 and Transformer layers for agentic AI workloads with a 1M token context window.

Comparing Moonshot AI's trillion-parameter Kimi K2.5 with NVIDIA's Mamba2-MoE hybrid Nemotron 3 Nano 30B-A3B - frontier intelligence versus a model engineered for maximum throughput, 1M context, and 10x lower cost.