At the 2026 CES International Consumer Electronics Show in Las Vegas, AI leader NVIDIA officially launched its highly anticipated next-generation AI platform—Rubin. At the core of this platform, the new super chip Vera Rubin made its debut, marking another significant leap for NVIDIA in the field of high-performance computing.
According to AIbase, the Rubin platform is designed to provide more efficient computing power for current agent AI and large-scale reasoning models. The Vera Rubin chip features an innovative integrated design, packaging a Vera CPU and two Rubin GPUs in a single processor. Compared to the previous Blackwell architecture, the new platform has achieved a qualitative leap in efficiency. Official data from NVIDIA shows that when training the same mixture-of-experts model (MoE), the Rubin platform requires only a quarter of the number of GPUs needed previously.
In addition to the core chip, NVIDIA also showcased a range of network and storage components, including the NVLink6 switch and the ConnectX-9 super NIC. These technologies together form the Vera Rubin NVL72 server, capable of building large-scale supercomputing clusters.
For enterprises, the direct benefit brought by the new architecture is a significant reduction in costs. AIbase reported that the Rubin platform can reduce the cost of processing "tokens" during the AI inference stage by 90%. This means that when processing massive amounts of language, image, and video data, companies will see significant optimization in energy consumption and overall operational costs. Despite competition from Advanced Micro Devices (AMD) and tech giants like Google and Amazon developing their own chips, NVIDIA aims to continue solidifying its dominant position in the global AI computing market by maintaining a fast-paced product iteration cycle each year.
Key Points:
🚀 New Architecture Launch: NVIDIA officially launched the Rubin platform, which includes the Vera Rubin super chip with new CPU and GPU cores integrated.
📉 Significant Efficiency Improvement: Compared to previous products, the number of GPUs required to train similar AI models has been reduced by 75%, and the cost of inference has dropped by 90%.
🌐 Comprehensive Computing Solution: The new platform covers a complete ecosystem from switches to storage solutions, aiming to support complex AI model reasoning at the scale of trillions of parameters.
