Catenaa, Tuesday, January 06, 2025- AI powerhouse Nvidia announced the launch of its next-generation Vera Rubin superchip at CES 2026 on Monday in Las Vegas.
One of six chips that make up what Nvidia is now calling its Rubin platform, Vera Rubin combines one Vera CPU and two Rubin GPUs in a single processor.
Nvidia is framing the Rubin platform as ideal for agentic AI, advanced reasoning models, and mixture-of-experts (MoE) models, which combine a series of “expert” AIs and route queries to the appropriate one depending on the question a user asks.
“Rubin arrives at exactly the right moment, as AI computing demand for both training and inference is going through the roof,” Nvidia CEO Jensen Huang said in a statement.
“With our annual cadence of delivering a new generation of AI supercomputers, and extreme codesign across six new chips, Rubin takes a giant leap toward the next frontier of AI.”
Shares of Nvidia rose by over 2% on Tuesday after the launch.
In addition to the Vera CPU and Rubin GPUs, the Rubin platform includes four other networking and storage chips: the Nvidia NVLink 6 Switch, Nvidia ConnectX-9 SuperNIC, Nvidia BlueField-4 DPU, and Nvidia Spectrum-6 Ethernet Switch.
All of that can then be packaged together into Nvidia’s Vera Rubin NVL72 server, which combines 72 GPUs into a single system. Join several NVL72s together, and you get Nvidia’s DGX SuperPOD, a kind of massive AI supercomputer.
These huge systems are what hyperscalers like Microsoft, Google, Amazon, and social media giant Meta are spending billions of dollars to get their hands on.
Nvidia is also touting its AI storage, called Nvidia Inference Context Memory Storage, which the company says is necessary to store and share data generated by trillion-parameter and multi-step reasoning AI models.
All of this is meant to make the Rubin platform more efficient than Nvidia’s previous-generation Grace Blackwell offering.
