NVIDIA Nemotron 3 Super: Open Model for Agentic AI


The race for agentic AI infrastructure just shifted dramatically. NVIDIA released Nemotron 3 Super this week, a 120 billion parameter open model specifically designed for the throughput demands of autonomous agents. While everyone debates whether Claude or GPT handles coding tasks better, NVIDIA quietly solved the real bottleneck: agents that can actually hold a million tokens in context without melting your budget.

Through implementing multi-agent systems at scale, I have seen how context explosion kills agentic workflows. Agents constantly exchange full histories, reasoning chains, and tool outputs. A software development agent that needs to reason over an entire codebase cannot do so when the model forgets the first half of the context. Nemotron 3 Super addresses this with a native 1M token context window and a hybrid architecture that makes long context practical rather than theoretical.

AspectSpecification
Total Parameters120B with 12B active per token
Context Window1M tokens native
ArchitectureHybrid Mamba-Transformer MoE
Throughput5x previous Nemotron, 2.2x vs GPT-OSS-120B
LicenseNVIDIA Nemotron Open Model License
AvailabilityHugging Face, NVIDIA NIM, major cloud providers

Why a Hybrid Architecture Matters for Agents

Traditional transformers have a fundamental problem with long contexts: the key-value cache grows linearly with sequence length. When agents pass around conversation histories that span thousands of turns, memory consumption spirals out of control. This is why most production agent systems implement aggressive context pruning, often losing critical information in the process.

Nemotron 3 Super takes a different approach by combining Mamba state space layers with transformer attention. The Mamba layers compress context into a rolling, fixed-size state rather than maintaining the full key-value cache. Meanwhile, transformer layers preserve the precise associative recall capabilities needed when agents must locate specific facts in lengthy documents.

The practical result is that you can actually load an entire codebase into context at once. No document segmentation. No retrieval pipeline to retrieve retrieved chunks. The agent sees everything and can reason across the full span of information. For engineers building agentic AI systems, this eliminates one of the most frustrating architectural constraints.

Benchmark Performance That Matters

The headline numbers look impressive, but the benchmarks that matter for agent developers tell a more nuanced story:

PinchBench (85.6%): This benchmark specifically measures how well models perform as the reasoning core of autonomous agents. Nemotron 3 Super achieves the best score among open models in its class, indicating strong real-world agent capabilities.

DeepResearch Bench (#1): NVIDIA’s AI-Q research agent powered by Nemotron 3 Super currently holds the top position on both DeepResearch Bench leaderboards. These benchmarks measure multistep research across large document sets while maintaining reasoning coherence. For agents that need to synthesize information from multiple sources, this matters.

Throughput (478 tokens/second): Raw intelligence is worthless if your agent spends minutes waiting for each response. Nemotron 3 Super achieves up to 2.2x higher throughput than comparable open models and 7.5x higher than Qwen 3.5-122B in high-volume settings.

The Artificial Analysis Intelligence Index ranks Nemotron 3 Super at 36 points, placing it above average among open models but below frontier closed models like GPT-5.4 (57) and Claude Opus 4.6 (53). The tradeoff is clear: you sacrifice some raw capability for open weights, throughput, and a context window five times larger than Claude’s 200K limit.

The Technical Innovations Under the Hood

Three architectural decisions make Nemotron 3 Super particularly suited for AI agent development:

Latent MoE: Traditional mixture-of-experts models route tokens to different expert networks. Nemotron 3 Super compresses tokens before routing, enabling four times as many expert specialists for the same inference cost. More experts means more specialized reasoning capabilities without proportional cost increases.

Multi-token Prediction (MTP): Rather than predicting one token at a time, the model predicts multiple future tokens in a single forward pass. For structured generation tasks common in agent systems, this delivers up to 3x wall-clock speedups. When your agent generates JSON tool calls or code blocks, it completes faster.

Native NVFP4 Training: Instead of training at full precision and quantizing afterward, NVIDIA trained Nemotron 3 Super natively in 4-bit precision from the first gradient update. This enables 4x improved memory and compute efficiency on NVIDIA B200 GPUs compared to FP8 on H100, while maintaining accuracy. The efficiency gains compound when running multiple agent instances.

When to Choose Nemotron 3 Super

The decision framework for selecting the right LLM depends heavily on your use case:

Choose Nemotron 3 Super if:

  • Your agents need to reason over entire codebases or document sets
  • You require high throughput for production-scale agent deployments
  • You want open weights for enterprise data control and customization
  • You are building multi-agent systems where context explosion is a real problem
  • You can deploy on NVIDIA infrastructure

Consider alternatives if:

  • Raw benchmark accuracy is the primary concern (GPT-5.4, Claude Opus remain ahead)
  • You need maximum coding performance specifically (Claude Opus 4.6 edges out in agentic coding tasks)
  • Your context requirements fit within 200K tokens

The comparison with Claude 4.1 Opus is instructive: Claude averages 58.7 on agentic benchmarks versus 56.6 for Nemotron 3 Super. But Nemotron offers 5x the context window. For many real-world agent architectures, that context advantage outweighs a few percentage points on benchmarks.

Practical Deployment Options

NVIDIA has made Nemotron 3 Super accessible across multiple platforms:

  • Hugging Face: Full weights available for self-hosting
  • NVIDIA NIM: Optimized inference containers for enterprise deployment
  • Cloud Providers: Available on Baseten, Cloudflare, Coreweave, DeepInfra, Fireworks AI, FriendliAI, and Google Cloud
  • OpenRouter: API access alongside other models
  • Perplexity Pro: Integrated into Perplexity’s infrastructure

For engineers who have been running models locally, the open weights mean you can host Nemotron 3 Super on your own infrastructure if you have access to NVIDIA GPUs. The 12B active parameters make inference more tractable than the full 120B would suggest.

The Bigger Picture for Open Models

Nemotron 3 Super represents a broader trend in open model development. Companies like NVIDIA, Meta, and AI2 are proving that open weights models can compete in specialized domains even if they lag frontier closed models on general benchmarks.

The training recipe matters here. NVIDIA trained on 25 trillion tokens using their native 4-bit format, with additional focus on 10 billion reasoning-specific tokens and 15 million coding problems. The reinforcement learning phase used over 1.2 million environment rollouts across 21 configurations. This level of investment in agentic-specific training data is what separates models that work well for chat from models that work well for autonomous tasks.

Warning: The NVIDIA Nemotron Open Model License is permissive but not completely open. Review the license terms carefully if you plan commercial deployment or redistribution. The license supports enterprise data control, which is a key differentiator for many production use cases.

What This Means for AI Engineers

If you are building agentic systems today, Nemotron 3 Super deserves serious evaluation. The combination of open weights, massive context window, and optimized throughput addresses real constraints that production agents face.

The architecture innovations, particularly the hybrid Mamba-Transformer approach and native low-precision training, point toward where agent-specific models are heading. We are moving beyond general-purpose LLMs toward models explicitly designed for the token economics of multi-agent workflows.

For most engineers, the practical next step is testing Nemotron 3 Super against your specific agent workloads via one of the cloud providers. The throughput improvements and context capacity may justify the switch from closed models, particularly for workloads where you are currently losing information due to context limitations.

Frequently Asked Questions

How does Nemotron 3 Super compare to Claude for coding agents?

Claude Opus 4.6 edges out slightly on agentic coding benchmarks (58.7 vs 56.6 average). However, Nemotron 3 Super offers a 1M token context window versus Claude’s 200K, which matters significantly when agents need to reason over entire codebases.

Can I run Nemotron 3 Super locally?

Yes, the open weights are available on Hugging Face. However, the 120B total parameters require substantial GPU memory even with only 12B active parameters. Most practical deployments use cloud providers or NVIDIA NIM containers.

What is the Mamba architecture and why does it matter?

Mamba is a state space model architecture that provides linear-time complexity with respect to sequence length, unlike transformer attention which scales quadratically. The hybrid approach preserves the precision of attention while gaining the efficiency benefits of state space models for long contexts.

Sources

If you are building production AI systems and want to understand the fundamentals that power real agent architectures, join the AI Engineering community where we share implementation experience across different model choices and deployment strategies.

Inside the community, you will find practical guidance on model selection, agent architecture patterns, and the infrastructure decisions that separate demo projects from production systems.

Zen van Riel

Zen van Riel

Senior AI Engineer at GitHub | Ex-Microsoft

I went from a $500/month internship to Senior Engineer at GitHub. Now I teach 30,000+ engineers on YouTube and coach engineers toward $200K+ AI careers in the AI Engineering community.

Blog last updated