LLM AtlasLLM AtlasSearch models

NVIDIA

Nemotron 3 Super 120B

Open weightopen-weight

NVIDIA's flagship 120B/12B-active LatentMoE model with 1M context, trained on 25T tokens. Strong on agentic workflows, reasoning, and long-context tasks. Requires 8x H100-80GB.

Last verified: 2026-03-29Confidence: HighSources: 5
textreasoningcodeopen-source
Input price
Unpublished
Output price
Unpublished
Context window
1,048,576
Max output
32,768
Release date
2026-03-11
Access
open-weight, self-hosted, hosted, api
License
NVIDIA Nemotron Open Model License
Last verified
2026-03-29

Capability profile

Radar view of the model's practical strengths. This chart is backed by textual summaries below for crawlability.

Benchmark summary

Scores 90.21 on AIME25, 60.47 on SWE-Bench, and maintains 91.75 accuracy at 1M context on RULER.

No benchmark series is attached to this model yet. Source links and product metadata are available below.

Strengths

  • 1M token context
  • LatentMoE efficiency
  • Strong agentic capability
  • Multilingual (7 languages)
  • Configurable reasoning mode

Trade-offs

  • Requires 8x H100+
  • Large model size
  • Recent release

Crawlable benchmark analysis

Nemotron 3 Super 120B is positioned as an enterprise llm model with published scores that emphasize its practical fit for buyers evaluating the entry.

Published scores highlight reasoning 88/100, coding 82/100, enterprise readiness 88/100, vision 40/100, speed 72/100, and safety 80/100.

Pricing is not fully published for this entry. With a context window of 1,048,576 tokens, it supports large-document analysis and retrieval workflows.

Benchmark coverage is still limited for this entry, so this section focuses on published metadata and deployment fit.

Sources

Provider and distribution links used to verify this model record.

Last verified: 2026-03-29

Related models

OpenAI

GPT-5.4

OpenAI

OpenAI's GPT-5.4, the most capable and efficient frontier model for professional work. First general-purpose model with native computer-use capabilities. Combines industry-leading coding from GPT-5.3-Codex with improved agentic workflows.

Score 933 sources
textreasoningtool-usevisionapihosted
Context
1,000,000
Input
$0.005/1K tok
Output
$0.02/1K tok
Coverage
Full profile

Anthropic

Claude Sonnet 4.6

Claude 4.6

Anthropic's current Sonnet tier for fast frontier reasoning, coding, and long-context agent work.

Score 923 sources
textvisionreasoningcodetool-useapihosted
Context
1,000,000
Input
$0.003/1K tok
Output
$0.02/1K tok
Coverage
Full profile

Anthropic

Claude Opus 4.6

Claude 1M

Anthropic's most intelligent Claude model for complex agents, coding, and deep reasoning, with 1M token context and 128K output.

Score 913 sources
textvisionreasoningapihosted
Context
1,000,000
Input
$0.005/1K tok
Output
$0.03/1K tok
Coverage
Full profile