NVIDIA Launches Nemotron 3 Super: Open 120B MoE Model Optimized for Agentic AI Systems
NVIDIA released Nemotron 3 Super, a fully open 120B parameter (12B active) hybrid Mamba-Transformer MoE model with 1M token context, designed for efficient multi-agent applications and high-throughput reasoning. It features latent MoE, multi-token prediction, and NVFP4 training, outperforming larger models in speed and efficiency while releasing weights, data, and recipes. Available now on Hugging Face and inference platforms.
Delivers state-of-the-art open agentic AI efficiency, accelerating multi-agent development on NVIDIA hardware and democratizing advanced capabilities.