Nvidia Vera Rubin: 1.3M components, 10x efficiency, 72 GPUs ship H2 2026 to Meta, OpenAI

GigaNectar Team

Close-up view of Nvidia Vera Rubin superchip showing dual Rubin GPU dies and Vera CPU with green circuit board and metallic cooling components

Nvidia announced its next-generation AI system called Vera Rubin at the Consumer Electronics Show in January 2026. The system consists of 1.3 million components and delivers 10 times more performance per watt compared to its predecessor, Grace Blackwell. This efficiency gain addresses energy consumption challenges in AI infrastructure deployment.

The chips are primarily manufactured by Taiwan Semiconductor Manufacturing Company, with more than 80 suppliers across at least 20 countries contributing components. Major customers including Meta, OpenAI, and Anthropic plan to deploy Vera Rubin systems starting in the second half of 2026. Nvidia’s AI infrastructure head Dion Harris stated the company provides suppliers with detailed forecasts to meet production demands.

Nvidia Vera Rubin AI System

Next-generation rack-scale computing with 1.3 million components

Technical Specifications

Vera Rubin represents Nvidia’s most advanced rack-scale AI system. Each rack contains 72 Rubin graphics processing units and 36 Vera central processing units, totaling about 1,300 microchips compared to Grace Blackwell’s 864 chips.

1.3M
Total Components
10x
Performance Per Watt vs Grace Blackwell
72
Rubin GPUs Per Rack
36
Vera CPUs Per Rack
~2 Tons
Rack Weight
1,300
Total Microchips

The Vera Rubin superchip contains two Rubin GPUs and one Vera CPU, with 17,000 components in total. The system consumes about twice as much power as Grace Blackwell, but the 10 times performance per watt improvement makes it significantly more efficient for AI workloads. According to Mizuho Securities analyst Jordan Klein, the key metric is tokens per power consumed, which directly impacts return on investment.

System Architecture

Vera Rubin integrates components from a global supply chain spanning more than 80 suppliers in at least 20 countries including China, Vietnam, Thailand, Mexico, Israel, and the United States.

🔲
Processing Units
72 Rubin GPUs + 36 Vera CPUs
Core chips primarily manufactured by TSMC. Nvidia plans to manufacture Blackwell GPUs at TSMC’s Arizona fabrication facilities as part of its $500 billion US infrastructure investment through 2029.
💧
Liquid Cooling System
100% Liquid Cooled
First Nvidia system with complete liquid cooling. Harris stated this design helps data centers consume much less water than traditional evaporative cooling methods.
🔧
Modular Design
18 Compute Trays
Each superchip slides out of compute trays in seconds for maintenance and repairs. Grace Blackwell components are soldered to boards, making Vera Rubin’s modular approach a significant improvement.
🌍
Global Supply Chain
80+ Suppliers, 20+ Countries
Components sourced internationally including liquid cooling elements, power systems, and compute trays from suppliers in China, Vietnam, Thailand, Mexico, Israel, and the United States.
Power Efficiency
2x Power Draw, 10x Performance Gain
The system consumes about twice the power of Grace Blackwell but delivers 10 times more performance per watt, resulting in significantly better efficiency for AI workloads.
🏭
Manufacturing Locations
US, Taiwan, Mexico
Racks manufactured in the United States, Taiwan, and at a new Foxconn plant in Mexico to meet global demand.

Performance Comparison

Vera Rubin contains about 1,300 total microchips compared to Grace Blackwell’s 864. The performance per watt improvement represents a 10-fold increase, addressing energy efficiency concerns in AI infrastructure.

Development Timeline

Vera Rubin development follows Grace Blackwell’s 2024 production launch. CEO Jensen Huang announced the system entered full production in January 2026 at the Consumer Electronics Show in Las Vegas.

2024
Grace Blackwell entered production with 864 microchips per rack, establishing a new standard for rack-scale AI computing density and performance.
January 6, 2026
CEO Jensen Huang announced Vera Rubin in full production at the Consumer Electronics Show in Las Vegas, detailing the system’s specifications and capabilities.
February 13, 2026
Nvidia showed the Vera Rubin superchip to reporters at headquarters in Santa Clara, California, featuring two Rubin GPUs and one Vera CPU with 17,000 components.
Late February 2026
Meta announced plans to deploy Vera Rubin in data centers by 2027, becoming one of the first confirmed customers.
Second Half 2026
Vera Rubin scheduled to ship to customers including OpenAI, Anthropic, Amazon, Google, and Microsoft.

Expected Customer Deployments

Nvidia’s list of expected Vera Rubin customers includes major technology companies and AI research organizations planning deployments in the second half of 2026 and into 2027.

Meta
Deploying by 2027
OpenAI
Expected Customer
Anthropic
Expected Customer
Amazon
Expected Customer
Google
Expected Customer
Microsoft
Expected Customer

Key Technical Improvements

Vera Rubin introduces several technical improvements over Grace Blackwell, including enhanced energy efficiency, simplified maintenance procedures, and water conservation through liquid cooling.

Energy Efficiency
Delivers 10 times more performance per watt compared to Grace Blackwell. While consuming about twice the power, the performance improvement makes it significantly more efficient for AI workloads.
🔧
Simplified Maintenance
Modular design allows superchips to slide out from 18 compute trays in seconds. Grace Blackwell components are soldered to boards, making Vera Rubin’s serviceability a major improvement.
💧
Water Conservation
First Nvidia system with 100% liquid cooling. Harris stated this design helps data centers consume much less water than traditional evaporative cooling methods.
📊
Token Processing Efficiency
Optimized for tokens per power consumed. Mizuho Securities analyst Jordan Klein stated this metric determines return on investment for AI infrastructure spending.
🏗️
US Manufacturing Investment
Part of Nvidia’s plan to manufacture up to $500 billion of AI infrastructure in the United States through 2029, including Blackwell GPUs at TSMC’s Arizona fabrication facilities.
🔲
Increased Chip Density
Contains about 1,300 total microchips compared to Grace Blackwell’s 864, packed into rack-scale systems designed for maximum computational density and efficiency.

The system is scheduled to ship in the second half of 2026. Futurum Group estimates pricing at approximately 25% higher than Grace Blackwell, placing Vera Rubin systems around $3.5 million to $4 million per rack. Nvidia faces competition from AMD, which plans to ship its first rack-scale system called Helios later in 2026 with a major commitment from Meta for up to 6 gigawatts of capacity.

Daniel Newman from Futurum Group stated these systems integrate compute, networking, cabling, and cooling into single racks built for efficiency and performance, differing from traditional server construction. Harris acknowledged competition, stating the development represents a complex endeavor. The system addresses memory costs through detailed supplier forecasts as global AI demand creates component shortages. Related developments include tariff impacts on chip manufacturing, AI model developments, and processor ecosystem initiatives.

Leave a comment