In the present day on the Client Electronics present, Nvidia CEO Jensen Huang formally launched the corporate’s new Rubin computing structure, which he described because the cutting-edge in AI {hardware}. The brand new structure is presently in manufacturing and is anticipated to ramp up additional within the second half of the 12 months.
“Vera Rubin is designed to deal with this elementary problem that now we have: The quantity of computation needed for AI is skyrocketing.” Huang instructed the viewers. “In the present day, I can inform you that Vera Rubin is in full manufacturing.”
The Rubin structure, which was first introduced in 2024, is the most recent results of Nvidia’s relentless {hardware} growth cycle, which has reworked Nvidia into probably the most useful company on the earth. The Rubin structure will substitute the Blackwell structure, which in flip, changed the Hopper and Lovelace architectures.
Rubin chips are already slated to be used by almost each main cloud supplier, together with high-profile Nvidia partnerships with Anthropic, OpenAI, and Amazon Web Services. Rubin methods will even be utilized in HPE’s Blue Lion supercomputer and the upcoming Doudna supercomputer at Lawrence Berkeley Nationwide Lab.
Named for the astronomer Vera Florence Cooper Rubin, the Rubin structure consists of six separate chips designed for use in live performance. The Rubin GPU stands on the heart, however the structure additionally addresses rising bottlenecks in storage and interconnection with new enhancements within the Bluefield and NVLink methods respectively. The structure additionally features a new Vera CPU, designed for agentic reasoning.
Explaining the advantages of the brand new storage, Nvidia’s senior director of AI infrastructure options Dion Harris pointed to the rising cache-related reminiscence calls for of contemporary AI methods.
“As you begin to allow new varieties of workflows, like agentic AI or long-term duties, that places numerous stress and necessities in your KV cache,” Harris instructed reporters on a name, referring to a memory system used by AI models to condense inputs. “So we’ve launched a brand new tier of storage that connects externally to the compute system, which lets you scale your storage pool way more effectively.”
Techcrunch occasion
San Francisco
|
October 13-15, 2026
As anticipated, the brand new structure additionally represents a big advance in pace and energy effectivity. In accordance with Nvidia’s assessments, the Rubin structure will function three and a half occasions quicker than the earlier Blackwell structure on model-training duties and 5 occasions quicker on inference duties, reaching as excessive as 50 petaflops. The brand new platform will even help eight occasions extra inference compute per watt.
Rubin’s new capabilities come amid intense competitors to construct AI infrastructure, which has seen each AI labs and cloud suppliers scramble for Nvidia chips in addition to the services essential to energy them. On an earnings name in October 2025, Huang estimated that between $3 trillion and $4 trillion will likely be spent on AI infrastructure over the following 5 years.

