Close Menu
Altcoin ObserverAltcoin Observer
  • Regulation
  • Bitcoin
  • Altcoins
  • Market
  • Analysis
  • DeFi
  • Security
  • Ethereum
Categories
  • Altcoins (3,169)
  • Analysis (3,293)
  • Bitcoin (3,908)
  • Blockchain (2,157)
  • DeFi (2,623)
  • Ethereum (2,615)
  • Event (119)
  • Exclusive Deep Dive (1)
  • Landscape Ads (2)
  • Market (2,714)
  • Press Releases (12)
  • Reddit (2,600)
  • Regulation (2,474)
  • Security (3,678)
  • Thought Leadership (3)
  • Videos (44)
Hand picked
  • BTC Rush Telegram game (featured even on CMC) has scammed it’s users
  • Nvidia Launches Nemotron 3 Super, a 120GB Open AI Model Built for Agentic Workloads – Bitcoin News
  • Ethereum Staking Surpasses 32% – Yet ETH Still Lacks ONE KEY Driver
  • eToro buys standalone wallet company Zengo for $70 million
  • Crypto for a home? Coinbase brings token-backed down payments to housing market
We are social
  • Facebook
  • Twitter
  • Instagram
  • YouTube
Facebook X (Twitter) Instagram
  • About us
  • Disclaimer
  • Terms of service
  • Privacy policy
  • Contact us
Facebook X (Twitter) Instagram YouTube LinkedIn
Altcoin ObserverAltcoin Observer
  • Regulation
  • Bitcoin
  • Altcoins
  • Market
  • Analysis
  • DeFi
  • Security
  • Ethereum
Events
Altcoin ObserverAltcoin Observer
Home»Bitcoin»Nvidia Launches Nemotron 3 Super, a 120GB Open AI Model Built for Agentic Workloads – Bitcoin News
Bitcoin

Nvidia Launches Nemotron 3 Super, a 120GB Open AI Model Built for Agentic Workloads – Bitcoin News

April 20, 2026No Comments
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
Share
Facebook Twitter LinkedIn Pinterest Email


Key points to remember:

  • Nvidia has released Nemotron 3 Super, a 120 B parameter open MoE model enabling only 12.7 B parameters per forward pass.
  • Nemotron 3 Super delivers up to 7.5x higher throughput than Qwen3.5-122B-A10B in agent workloads on settings of 8,000 inputs/64,000 outputs.
  • The model is fully open sourced under the Nvidia Nemotron Open Model license, with checkpoints and training data on Hugging Face.

Nvidia Launches Nemotron 3 Super With 7.5x Throughput Gains Over Qwen3.5-122B

The latest Nvidia model only activates 12.7 billion parameters per pass through using a Mixture-of-Experts (MoE) architecture, meaning most of its weight remains unused during inference. This design choice directly targets two issues developers face when deploying multi-stage AI agents: the added cost of extended reasoning chains and the increasing token usage that can multiply up to 15 times in multi-agent pipelines.

Nemotron 3 Super is the second model in Nvidia’s Nemotron 3 family, after the Nemotron 3 Nano from December 2025. Nvidia has announced the release around March 10, 2026.

The model uses a hybrid Mamba-Transformer skeleton on 88 layers. Mamba-2 blocks handle long sequences with linear timing efficiency, while Transformer attention layers preserve precise recall. This combination gives the model native support for pop-ups up to a million tokens without the memory penalties typical of purely attentional designs.

Nvidia has also integrated a LatentMoE routing system that compresses token embeddings into low-rank space before sending them to 512 experts per layer, activating 22 at a time. The company says this allows for approximately four times more experts for the same inference cost compared to standard MoE approaches, and allows for finer task specialization, such as separating Python logic from SQL management at the expert level.

Nvidia launches Nemotron 3 Super, a 120GB open AI model designed for agentic workloads
Image source: NVIDIA blog.

Multi-token prediction layers, using two heads of shared weight, accelerate thought chain generation and enable native speculative decoding. On structured tasks, Nvidia reports up to three times faster generation.

The model was pre-trained on 25 trillion tokens in two phases. The first phase used 20 trillion large data tokens. The second used five trillion high-quality tokens optimized for benchmark performance. A final expansion phase on 51 billion tokens extended the native context to one million tokens. Post-training included supervised fine-tuning on approximately seven million samples and reinforcement learning in 21 environments with over 1.2 million deployments.

In benchmarks, Nemotron 3 Super scored 83.73 on MMLU-Pro, 90.21 on AIME25 and 60.47 on SWE-Bench using OpenHands. On PinchBench, it achieved 85.6%, the highest score among open models in its category. In a long context assessment, he scored 91.64 on RULER 1M.

Compared to GPT-OSS-120B, Nemotron 3 Super delivers 2.2x higher throughput at 8k input and 64k output. Against Qwen3.5-122B-A10B, this figure reaches 7.5 times. Nvidia also reports more than five times higher throughput and up to two times higher accuracy than the previous generation Nemotron Super.

Nvidia trained the model end-to-end in its NVFP4 four-bit floating point format, optimized for Blackwell GPUs. On the B200 hardware, Nvidia claims that inference runs up to four times faster than FP8 on the H100, with no reported loss of accuracy. FP8 and NVFP4 quantized checkpoints maintain 99.8% or greater total accuracy.

The model also powers the Nvidia AI-Q search agent, which reached the top spot in the Deepresearch Bench rankings.

Nemotron 3 Super is fully open sourced under the Nvidia Nemotron Open Model license. Checkpoints in BF16, FP8, and NVFP4 formats, as well as pre-training data, post-training samples, and reinforcement learning environments, are available on Hugging Face. Inference is supported through Nvidia NIM, build.nvidia.com, Perplexity, Openrouter, Together AI, Google Cloud, AWS, Azure, and Coreweave, with on-premises options through Dell Enterprise Hub and HPE.

Developers can access training recipes, fine-tuning guides, and inference cookbooks through the NeMo platform using vLLM, SGLang, and TensorRT-LLM.



Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleEthereum Staking Surpasses 32% – Yet ETH Still Lacks ONE KEY Driver
Next Article BTC Rush Telegram game (featured even on CMC) has scammed it’s users

Related Posts

Bitcoin

Earnings season over – traders now watching the Fed

April 20, 2026
Bitcoin

US-Based Bitcoin ETFs See Inflows of Around $1 Billion Last Week: Report

April 19, 2026
Bitcoin

Bitcoin Network Improves as Difficulty Decreases by 2.43% and Hash Price Increases by 13.65%

April 19, 2026
Add A Comment
Leave A Reply Cancel Reply

Single Page Post
Share
  • Facebook
  • Twitter
  • Instagram
  • YouTube
Featured Content
Event

Dutch Blockchain Week 2026 strengthens position as Europe’s leading B2B blockchain event week

April 14, 2026

Amsterdam, April 2026 – Dutch Blockchain Week 2026 is rapidly evolving into one of Europe’s…

Event

Global Games Show Riyadh: The Ultimate Creator & Influencer Hub

March 31, 2026

The fast-evolving gaming ecosystem of Riyadh is powered by solid national investment, a flourishing esports…

1 2 3 … 82 Next
  • Facebook
  • Twitter
  • Instagram
  • YouTube

Ethereum Staking Surpasses 32% – Yet ETH Still Lacks ONE KEY Driver

April 20, 2026

Ethereum Sees Demand Rise in the US – But Why Is ETH Stuck Near the Breakout?

April 20, 2026

AAVE Price Falls 18% as KelpDAO Exploit Spreads – What Happened?

April 19, 2026
Facebook X (Twitter) Instagram LinkedIn
  • About us
  • Disclaimer
  • Terms of service
  • Privacy policy
  • Contact us
© 2026 Altcoin Observer. all rights reserved by Tech Team.

Type above and press Enter to search. Press Esc to cancel.

bitcoin
Bitcoin (BTC) $ 74,214.00
ethereum
Ethereum (ETH) $ 2,267.47
tether
Tether (USDT) $ 1.00
xrp
XRP (XRP) $ 1.40
bnb
BNB (BNB) $ 617.86
usd-coin
USDC (USDC) $ 0.999782
solana
Solana (SOL) $ 83.85
tron
TRON (TRX) $ 0.332029
figure-heloc
Figure Heloc (FIGR_HELOC) $ 1.04
staked-ether
Lido Staked Ether (STETH) $ 2,265.05