NVIDIA Launches Rubin Platform with Six New Chips for Advanced AI Supercomputing
NVIDIA has launched the Rubin platform, featuring six new chips aimed at advancing AI supercomputing, with a notable 10x reduction in inference token costs and a requirement of 4x fewer GPUs for training mixture-of-experts models compared to its predecessor. Key components include the NVIDIA Vera CPU and Rubin GPU, and the platform is set to support large-scale AI factories in next-generation data centers, with partnerships including Microsoft, AWS, and Google Cloud. Collaborations with Red Hat and CoreWeave will further optimize the AI stack for enhanced performance and efficiency.

NVIDIA has unveiled the Rubin platform, featuring six new chips designed to enhance AI supercomputing capabilities. The platform offers up to 10x reduction in inference token costs and requires 4x fewer GPUs for training mixture-of-experts (MoE) models compared to the Blackwell platform.
Key components include the NVIDIA Vera CPU, Rubin GPU, NVLink 6 Switch, ConnectX-9 SuperNIC, BlueField-4 DPU, and Spectrum-6 Ethernet Switch. The platform supports massive-scale AI factories and will be integrated into next-generation data centers by partners like Microsoft, AWS, and Google Cloud, with availability expected in the second half of 2026. Significant collaborations with companies such as Red Hat and CoreWeave aim to optimize the AI stack for the Rubin platform, enhancing both performance and efficiency for AI workloads.




Comments