Nvidia Unveils ‘Vera Rubin’ Platform as Next Generation of AI Computing at CES 2026

Compared to the current Blackwell architecture, Rubin offers 3.5 times faster training speed and reduces inference costs by a factor of 10.

Image source: NVIDIA livestream screenshot

Image source: NVIDIA livestream screenshot

TMTPOST -- Nvidia on Monday introduced its next-generation artificial intelligence computing platform, called Vera Rubin, at the Consumer Electronics Show (CES) 2026, marking what the company describes as a major step forward in performance, efficiency and scalability for training and running large AI models.

The new platform is designed to support emerging AI workloads such as autonomous “agentic” systems, advanced reasoning models and mixture-of-experts (MoE) architectures, which rely on routing user queries dynamically across multiple specialized models.

At the heart of the platform is the Vera Rubin superchip, which integrates one Vera central processing unit (CPU) with two Rubin graphics processing units (GPUs) into a single package. It is one of six interconnected chips that together form Nvidia’s broader Rubin architecture.

“Rubin arrives at a moment when the demand for AI computing for both training and inference is accelerating at an unprecedented pace,” Nvidia chief executive Jensen Huang said during the company’s keynote presentation. “With deep co-design across compute, networking and storage, this platform is built for the next frontier of AI.”

A Platform Built for Large-Scale AI

Beyond the main processor, the Rubin platform includes a suite of new networking and infrastructure components: the NVLink 6 Switch for high-speed GPU interconnects, the ConnectX-9 SuperNIC for network acceleration, the BlueField-4 data processing unit (DPU) for offloading infrastructure workloads, and the Spectrum-6 Ethernet Switch for large-scale data center networking.

These components can be assembled into Nvidia’s new NVL72 server system, which integrates 72 GPUs into a single rack-scale unit. Multiple NVL72 systems can then be combined into larger clusters known as DGX SuperPODs, which are used by hyperscale cloud providers and AI developers to train frontier models.

Customers for these systems include major cloud and technology firms such as Microsoft, Google, Amazon and Meta, all of which are investing heavily in AI infrastructure.

Nvidia also introduced a new storage architecture called Inference Context Memory Storage, designed to manage the massive volumes of data generated by trillion-parameter and multi-step reasoning models and to allow that data to be shared efficiently across large AI systems.

Efficiency Gains Over Previous Systems

Nvidia said the Rubin platform delivers significant efficiency improvements over its previous Grace Blackwell generation.

According to the company, Rubin can reduce the number of GPUs required to train certain mixture-of-experts models by up to four times, allowing companies either to cut costs or redeploy hardware to other workloads. Nvidia also claims that the platform can reduce the cost of AI inference — the process of generating outputs from trained models — by up to ten times per token.

Inference costs have become a growing concern for AI developers, as large language and multimodal models consume vast amounts of computing power and electricity when processing text, images and video. Lower token costs could significantly improve the total cost of ownership for enterprises deploying AI at scale.

Nvidia said the platform is already being sampled by partners and is now in full production.

Market Position and Competition

Nvidia’s dominance in AI chips has propelled it to the top ranks of global technology companies by market capitalization, although its valuation has fluctuated in recent months amid investor concerns about the pace and sustainability of AI spending.

The company also faces rising competition. Advanced Micro Devices (AMD) is developing its own rack-scale AI systems, while major cloud providers such as Google and Amazon are expanding the use of in-house chips for some workloads, including those supporting AI start-up Anthropic.

Google is also in talks with other technology firms about broader adoption of its custom processors in third-party data centers, according to people familiar with the matter.

Even so, analysts say Nvidia retains a substantial lead in AI hardware, software integration and developer ecosystem. Its strategy of delivering a new generation of AI platforms on an annual cadence could make it difficult for rivals to close the gap in the near term.

“With Rubin, Nvidia is not just selling faster chips — it is selling a tightly integrated AI computing stack,” said one industry analyst. “That makes it much harder for competitors to match the full system performance and ecosystem support that Nvidia now offers.”

As AI applications move beyond experimentation into large-scale deployment across industries, Nvidia is betting that demand for powerful, efficient and flexible AI infrastructure will continue to rise — and that Rubin will become the backbone of that next phase.

本文系作者 zhangxinyue 授权钛媒体发表,并经钛媒体编辑,转载请注明出处、作者和本文链接
本内容来源于钛媒体钛度号,文章内容仅供参考、交流、学习,不构成投资建议。
想和千万钛媒体用户分享你的新奇观点和发现,点击这里投稿 。创业或融资寻求报道,点击这里

敬原创,有钛度,得赞赏

赞赏支持
发表评论
0 / 300

根据《网络安全法》实名制要求,请绑定手机号后发表评论

登录后输入评论内容

快报

更多

15:54

大族激光:第一季度净利润同比增长116.59%

15:53

韩国4月24日起将电子烟与传统卷烟同等监管

15:46

外交部:中国核政策为世界注入稳定性和确定性

15:44

华为余承东:手机定价压力较大,后期可能涨价

15:43

巴基斯坦称已为所有“特殊访客”做好特别安保安排

15:42

外交部:中方支持相关方继续保持停火和谈判势头

15:38

伊朗外交部:目前暂无与美国进行第二轮谈判的计划

15:37

外交部:望有关方以负责任态度恪守停火协定,为海峡通行恢复正常提供必要条件

15:36

因AI电力需求及电网老化,美公用事业公司未来五年资本支出计划猛增至1.4万亿美元

15:34

“AI药物递送第一股”剂泰科技正式通过港交所聆讯

15:33

永辉超市:大规模闭店已结束,目前调改店已实现收入增长

15:31

最高法:正在抓紧起草关于依法妥善审理涉人工智能纠纷案件的意见

15:30

英特尔14A工艺整装待发,瑞银猜测其客户或包括英伟达及苹果等

15:28

伊朗议会要员:伊朗已决定继续与美国进行谈判

15:27

国债期货收盘表现分化,30年期主力合约涨0.01%

15:27

华友钴业陈红良:企业已拿到精矿出口相关批文,在途周期约3个月

15:25

港股群核科技拉升超100%

15:23

严惩“网络黑嘴”造谣诋毁,最高法例举涉胖某公司案件

15:22

一季度北京GDP同比增长5.9%

15:19

力源信息:在存储产业链紧张背景下,公司存储相关业务销售额大幅提升

扫描下载App