NVIDIA announces Vera Rubin POD featuring 1,152 GPUs across 40 racks, delivering 60 exaflops and 10x better inference performance per watt than Blackwell. (ReadNVIDIA announces Vera Rubin POD featuring 1,152 GPUs across 40 racks, delivering 60 exaflops and 10x better inference performance per watt than Blackwell. (Read

NVIDIA Unveils Vera Rubin POD 40-Rack AI Supercomputer for Agentic Workloads

2026/03/17 03:48
3 min read
For feedback or concerns regarding this content, please contact us at [email protected]

NVIDIA Unveils Vera Rubin POD 40-Rack AI Supercomputer for Agentic Workloads

Iris Coleman Mar 16, 2026 19:48

NVIDIA announces Vera Rubin POD featuring 1,152 GPUs across 40 racks, delivering 60 exaflops and 10x better inference performance per watt than Blackwell.

NVIDIA Unveils Vera Rubin POD 40-Rack AI Supercomputer for Agentic Workloads

NVIDIA just dropped the specs on its most ambitious AI infrastructure play yet. The Vera Rubin POD packs 1,152 Rubin GPUs across 40 racks, delivering 60 exaflops of compute power and 10 petabytes per second of total scale-up bandwidth. Production units ship in the second half of 2026.

The numbers here are staggering: 1.2 quadrillion transistors, nearly 20,000 NVIDIA dies, all engineered to function as a single coherent supercomputer. NVIDIA claims 4x better training performance and 10x better inference performance per watt compared to its current Blackwell architecture—with token costs dropping to one-tenth of current levels.

Five Purpose-Built Rack Systems

The POD combines five distinct rack-scale systems, each targeting specific bottlenecks in modern AI workloads:

Vera Rubin NVL72 serves as the core compute engine. Each rack integrates 72 Rubin GPUs and 36 Vera CPUs connected through NVLink 6, which pushes 3.6 TB/s bandwidth per GPU—more total bandwidth than the entire global internet, according to NVIDIA. The system targets all four AI scaling laws: pretraining, post-training, test-time scaling, and agentic scaling.

Groq 3 LPX racks tackle the latency problem. With 256 language processing units per rack using SRAM-only architecture, these pair with NVL72 to deliver what NVIDIA claims is 35x more tokens and 10x more revenue opportunity for trillion-parameter models versus Blackwell.

Vera CPU racks provide sandbox environments for agent testing. A single rack sustains over 22,500 concurrent reinforcement learning environments—critical for validating agentic AI outputs before deployment.

BlueField-4 STX racks introduce what NVIDIA calls "AI-native storage" through the CMX context memory platform. By offloading KV cache to dedicated high-bandwidth storage, the system claims 5x higher tokens-per-second and 5x better power efficiency than traditional approaches.

Spectrum-6 SPX networking racks tie everything together with 102.4 Tb/s switches featuring co-packaged optics.

The Token Economics Argument

NVIDIA frames this around a specific market reality: token consumption now exceeds 10 quadrillion annually, and the shift from human-AI to AI-AI interactions will accelerate that growth dramatically. Modern agentic systems generate massive reasoning token volumes while expanding KV cache requirements—exactly the bottleneck this architecture targets.

Third-party SemiAnalysis InferenceMax benchmarks cited by NVIDIA show current Blackwell systems already deliver 50x better performance per watt and 35x lower cost per token compared to H200. Vera Rubin aims to extend that lead.

Thermal and Power Engineering

The third-generation MGX rack architecture introduces Intelligent Power Smoothing with 6x more rack-level energy storage (400 joules per GPU) than previous generations. This reduces peak current demands by up to 25% and eliminates the need for massive battery packs.

All racks operate at 45°C warm-water inlet temperatures, enabling data centers in many climates to use ambient air cooling. NVIDIA claims this frees enough power to add 10% more racks in the same facility power budget.

Looking Ahead

Beyond the initial POD configuration, NVIDIA previewed Vera Rubin Ultra NVL576 scaling to 576 GPUs across eight racks, and the next-generation Kyber architecture targeting NVL1152 with 144 GPUs per rack. The roadmap suggests NVIDIA sees multi-rack NVLink domains as the future of AI infrastructure—not just bigger GPUs, but fundamentally different system architectures.

For enterprises planning AI infrastructure investments, the message is clear: the economics of AI compute are shifting from chip-level to facility-level optimization. Those building out data centers now face a choice between current-generation systems and waiting for Vera Rubin availability in late 2026.

Image source: Shutterstock
  • nvidia
  • ai infrastructure
  • vera rubin
  • data centers
  • enterprise ai
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Renewal Fuels Expands Patent Portfolio and Leadership Team for Fusion Energy Commercialization

Renewal Fuels Expands Patent Portfolio and Leadership Team for Fusion Energy Commercialization

Renewal Fuels files 8 new patents for Texatron™ fusion tech and appoints key leaders to drive commercialization strategy for clean energy generation. The post Renewal
Share
Citybuzz2026/03/16 23:20
BetFury is at SBC Summit Lisbon 2025: Affiliate Growth in Focus

BetFury is at SBC Summit Lisbon 2025: Affiliate Growth in Focus

The post BetFury is at SBC Summit Lisbon 2025: Affiliate Growth in Focus appeared on BitcoinEthereumNews.com. Press Releases are sponsored content and not a part of Finbold’s editorial content. For a full disclaimer, please . Crypto assets/products can be highly risky. Never invest unless you’re prepared to lose all the money you invest. Curacao, Curacao, September 17th, 2025, Chainwire BetFury steps onto the stage of SBC Summit Lisbon 2025 — one of the key gatherings in the iGaming calendar. From 16 to 18 September, the platform showcases its brand strength, deepens affiliate connections, and outlines its plans for global expansion. BetFury continues to play a role in the evolving crypto and iGaming partnership landscape. BetFury’s Participation at SBC Summit The SBC Summit gathers over 25,000 delegates, including 6,000+ affiliates — the largest concentration of affiliate professionals in iGaming. For BetFury, this isn’t just visibility, it’s a strategic chance to present its Affiliate Program to the right audience. Face-to-face meetings, dedicated networking zones, and affiliate-focused sessions make Lisbon the ideal ground to build new partnerships and strengthen existing ones. BetFury Meets Affiliate Leaders at its Massive Stand BetFury arrives at the summit with a massive stand placed right in the center of the Affiliate zone. Designed as a true meeting hub, the stand combines large LED screens, a sleek interior, and the best coffee at the event — but its core mission goes far beyond style. Here, BetFury’s team welcomes partners and affiliates to discuss tailored collaborations, explore growth opportunities across multiple GEOs, and expand its global Affiliate Program. To make the experience even more engaging, the stand also hosts: Affiliate Lottery — a branded drum filled with exclusive offers and personalized deals for affiliates. Merch Kits — premium giveaways to boost brand recognition and leave visitors with a lasting conference memory. Besides, at SBC Summit Lisbon, attendees have a chance to meet the BetFury team along…
Share
BitcoinEthereumNews2025/09/18 01:20
Polymarket’s Iran Betting Surge Helped Spark Washington’s Prediction Market Crackdown Bill

Polymarket’s Iran Betting Surge Helped Spark Washington’s Prediction Market Crackdown Bill

A spike in Iran-related betting volume on Polymarket drew Congressional scrutiny and helped accelerate a new bill targeting crypto prediction markets in the US.
Share
coinlineup2026/03/17 06:06