NVIDIA has unveiled its latest AI compute platform, Vera Rubin, set for release in H2 2026. The new platform builds upon the previous generation's advancements with significant improvements across multiple dimensions.
Key features of Vera Rubin include:
- Increased Compute Power: Each GPU offers a 2.8x boost in FP4 compute performance compared to B200 GPUs.
- Enhanced Memory Bandwidth: With HBM4, each GPU achieves nearly three times the memory bandwidth (22 TB/s) relative to Blackwell GPUs, addressing the primary bottleneck in large language model inference.
- Doubled Interconnect Speed: NVLink bandwidth has doubled, enhancing data transfer efficiency between components.
At rack scale, these improvements translate into a 5–6x throughput increase over GB200 NVL72. The Vera Rubin platform introduces new chips such as the Rubin CPX for long-context workloads and BlueField-4 DPU for networking and storage offload, indicating NVIDIA's growing role in providing comprehensive data center solutions.
The platform consists of six key components:
- Rubin GPU: Core training and dense inference compute.
- Vera CPU: 88-core Arm v9
Read the full article at Towards AI - Medium
Want to create content about this topic? Use Nemati AI tools to generate articles, social posts, and more.

![[AINews] The Unreasonable Effectiveness of Closing the Loop](/_next/image?url=https%3A%2F%2Fmedia.nemati.ai%2Fmedia%2Fblog%2Fimages%2Farticles%2F600e22851bc7453b.webp&w=3840&q=75)



