The post Nvidia Drops Nemotron 3 Super Amid $26 Billion Open-Model AI Bet—America’s Answer to Qwen? appeared on BitcoinEthereumNews.com. In brief Nvidia launchedThe post Nvidia Drops Nemotron 3 Super Amid $26 Billion Open-Model AI Bet—America’s Answer to Qwen? appeared on BitcoinEthereumNews.com. In brief Nvidia launched

Nvidia Drops Nemotron 3 Super Amid $26 Billion Open-Model AI Bet—America’s Answer to Qwen?

2026/03/13 03:44
Okuma süresi: 5 dk
Bu içerikle ilgili geri bildirim veya endişeleriniz için lütfen [email protected] üzerinden bizimle iletişime geçin.

In brief

  • Nvidia launched Nemotron 3 Super, a 120B open-weight AI model optimized for autonomous agents and ultra-long context tasks.
  • The hybrid Mamba-Transformer MoE architecture delivers faster reasoning and over 5× throughput while running at 4-bit precision.
  • Nvidia’s $26 billion investment into open-source AI wants to counter China’s rise in the field.

Nvidia just shipped Nemotron 3 Super, a 120-billion-parameter open-weight model built to do one thing well: run autonomous AI agents without bleeding your compute budget dry.

That’s not a small problem. Multi-agent systems generate a lot more tokens than a normal chat—every tool call, reasoning step, and slice of context gets re-sent from scratch. As a result, costs explode, models tend to drift, and the agents slowly forget what they were supposed to be doing in the first place… or at least decrease in accuracy.

Nemotron 3 Super is Nvidia’s answer to all of that. The model runs 12 billion active parameters out of 120 billion total, using a mixture-of-experts (MoE) design that keeps inference cheap while retaining the reasoning depth complex workflows need. It packs a 1-million-token context window, so agents can hold an entire codebase, or nearly 750,000 words in memory before collapsing.

To build its model, Nvidia combined three components that rarely appear together in the same architecture: Mamba-2 state-space layers—a faster, memory-efficient alternative to attention for handling long token streams—along with Transformer attention layers for precise recall, and a new “Latent MoE” design that compresses token embeddings before routing them to experts. That allows the model to activate four times as many specialists at the same compute cost.

The model was also pretrained natively in NVFP4, Nvidia’s 4-bit floating-point format. In practice, that means the system learned to operate accurately within 4-bit arithmetic from the very first gradient update, rather than being trained at high precision and compressed afterward, which often causes models to lose accuracy.

For context, a model’s precision is measured in bits. Full precision, known as FP32, is the gold standard—but it is also extremely expensive to run at scale. Developers often reduce precision to save compute while trying to preserve useful performance.

Think of it like shrinking a 4K image down to 1080p: The picture still looks the same at a glance, just with less detail. Normally, dropping from 32-bit precision all the way to 4-bit would cripple a model’s reasoning ability. Nemotron avoids that problem by learning to operate at low precision from the start, instead of being squeezed into it later.

Compared to its own predecessor, Nemotron 3 Super delivers more than five times the throughput. Against external rivals, it’s 2.2x faster than OpenAI’s GPT-OSS 120B on inference throughput, and 7.5x faster than Alibaba’s Qwen3.5-122B.

We ran our own quick test. The reasoning held up well, including on prompts that were deliberately vague, badly worded, or based on wrong information. The model caught small errors in context without being asked to, handled math and logic problems cleanly, and didn’t fall apart when the question itself was slightly off.

The full training pipeline is public: weights on Hugging Face, 10 trillion curated pretraining tokens seen over 25 trillion total during training, 40 million post-training samples, and reinforcement learning recipes across 21 environment configurations. Perplexity, Palantir, Cadence, and Siemens are already integrating the model in their workflows.

The $26 billion bet

The model may be one piece of a larger strategy. A 2025 financial filing shows Nvidia plans to spend $26 billion over the next five years building open-weight AI models. Executives confirmed it, too.

Bryan Catanzaro, VP of applied deep learning research, told Wired the company recently finished pretraining a 550-billion-parameter model. Nvidia released its first Nemotron model back in November 2023, but that filing makes clear this is no longer a side project.

The investment is strategic considering Nvidia’s chips are still the default infrastructure for training and running frontier models. Models tuned to its hardware give customers a built-in reason to stay on Nvidia despite efforts from competitors to use other hardware. But there’s a more urgent pressure behind the move: America is losing the open-source AI race, and losing it fast.

Chinese open models went from barely 1.2% of global open-model usage in late 2024 to roughly 30% by the end of 2025, according to research by OpenRouter and Andreessen Horowitz. Alibaba’s Qwen overtook Meta’s Llama as the most-used self-hosted open-source model, according to Runpod. American companies including Airbnb adopted it for customer service. Startups worldwide are building on top of it. Beyond market share, that kind of adoption creates infrastructure dependencies that are hard to reverse.

While U.S. giants like OpenAI, Anthropic, and Google keep their best models locked behind APIs, Chinese labs from DeepSeek to Alibaba have been flooding the open ecosystem. Meta was the one major American player competing in open source with Llama, but Zuckerberg recently signaled the company might not make future models fully open.

The gap between “best proprietary model” and “best open model” used to be massive—and in America’s favor. That gap is now very small, and the open side of the ledger is increasingly Chinese.

There’s also a hardware threat underneath all of this. A new DeepSeek model is widely expected to drop soon, and it’s rumored to have been trained entirely on chips made by Huawei—a sanctioned Chinese company. If that’s confirmed, then it would give developers around the world, particularly in China, a concrete reason to start testing Huawei’s hardware. China’s Ziphu AI is already doing that.

That’s the scenario Nvidia most needs to prevent: Chinese open models and Chinese chips building an ecosystem that doesn’t need Nvidia at all.

Daily Debrief Newsletter

Start every day with the top news stories right now, plus original features, a podcast, videos and more.

Source: https://decrypt.co/360929/nvidia-drops-nemotron-3-super-26-billion-open-model-ai-bet

Piyasa Fırsatı
Belong Logosu
Belong Fiyatı(LONG)
$0.002097
$0.002097$0.002097
+5.48%
USD
Belong (LONG) Canlı Fiyat Grafiği
Sorumluluk Reddi: Bu sitede yeniden yayınlanan makaleler, halka açık platformlardan alınmıştır ve yalnızca bilgilendirme amaçlıdır. MEXC'nin görüşlerini yansıtmayabilir. Tüm hakları telif sahiplerine aittir. Herhangi bir içeriğin üçüncü taraf haklarını ihlal ettiğini düşünüyorsanız, kaldırılması için lütfen [email protected] ile iletişime geçin. MEXC, içeriğin doğruluğu, eksiksizliği veya güncelliği konusunda hiçbir garanti vermez ve sağlanan bilgilere dayalı olarak alınan herhangi bir eylemden sorumlu değildir. İçerik, finansal, yasal veya diğer profesyonel tavsiye niteliğinde değildir ve MEXC tarafından bir tavsiye veya onay olarak değerlendirilmemelidir.

Ayrıca Şunları da Beğenebilirsiniz

Disney Pockets $2.2 Billion For Filming Outside America

Disney Pockets $2.2 Billion For Filming Outside America

The post Disney Pockets $2.2 Billion For Filming Outside America appeared on BitcoinEthereumNews.com. Disney has made $2.2 billion from filming productions like ‘Avengers: Endgame’ in the U.K. ©Marvel Studios 2018 Disney has been handed $2.2 billion by the government of the United Kingdom over the past 15 years in return for filming movies and streaming shows in the country according to analysis of more than 400 company filings Disney is believed to be the biggest single beneficiary of the Audio-Visual Expenditure Credit (AVEC) in the U.K. which gives studios a cash reimbursement of up to 25.5% of the money they spend there. The generous fiscal incentives have attracted all of the major Hollywood studios to the U.K. and the country has reeled in the returns from it. Data from the British Film Institute (BFI) shows that foreign studios contributed around 87% of the $2.2 billion (£1.6 billion) spent on making films in the U.K. last year. It is a 7.6% increase on the sum spent in 2019 and is in stark contrast to the picture in the United States. According to permit issuing office FilmLA, the number of on-location shooting days in Los Angeles fell 35.7% from 2019 to 2024 making it the second-least productive year since 1995 aside from 2020 when it was the height of the pandemic. The outlook hasn’t improved since then with FilmLA’s latest data showing that between April and June this year there was a 6.2% drop in shooting days on the same period a year ago. It followed a 22.4% decline in the first quarter with FilmLA noting that “each drop reflected the impact of global production cutbacks and California’s ongoing loss of work to rival territories.” The one-two punch of the pandemic followed by the 2023 SAG-AFTRA strikes put Hollywood on the ropes just as the U.K. began drafting a plan to improve its fiscal incentives…
Paylaş
BitcoinEthereumNews2025/09/18 07:20
DEXTools raises $3 million to launch its perpetual DEX, "PerpTools".

DEXTools raises $3 million to launch its perpetual DEX, "PerpTools".

PANews reported on March 13 that, according to Cryptopolitan, DeFi data analytics platform DEXTools announced the completion of a $3 million funding round to launch
Paylaş
PANews2026/03/13 09:28
Exclusive interview with Smokey The Bera, co-founder of Berachain: How the innovative PoL public chain solves the liquidity problem and may be launched in a few months

Exclusive interview with Smokey The Bera, co-founder of Berachain: How the innovative PoL public chain solves the liquidity problem and may be launched in a few months

Recently, PANews interviewed Smokey The Bera, co-founder of Berachain, to unravel the background of the establishment of this anonymous project, Berachain's PoL mechanism, the latest developments, and answered widely concerned topics such as airdrop expectations and new opportunities in the DeFi field.
Paylaş
PANews2024/07/03 13:00