CES 2026: Jensen Huang Lays Out NVIDIA’s Plan for the Physical AI Era
At CES 2026, NVIDIA focused its keynote on how AI infrastructure is scaling across the global economy. CEO Jensen Huang spoke about data centers, robots, autonomous vehicles, and healthcare, with a clear emphasis on systems already moving into production.
There were no announcements about new GeForce GPUs during this keynote. Gaming updates were placed outside the main event. The stage was used to explain how AI is being deployed into real industries and how NVIDIA is building the systems that support it.
Vera Rubin Enters Production
Vera Rubin is now in full production.
Vera Rubin is a rack-scale AI computing system designed by NVIDIA. It combines the Vera CPU, Rubin GPU, NVLink 6, ConnectX-9, BlueField-4, and Spectrum-X Ethernet into a single platform for large-scale training, inference, and reasoning.
This system is designed for companies operating large AI clusters, sometimes called AI factories. These systems handle model training, real-time inference, and reasoning workloads that require high bandwidth, large memory pools, and efficient power use.
System-Level Design for AI Workloads
AI workloads are changing. Models are larger, and reasoning models generate more tokens per task. This places pressure on compute, networking, and memory.
NVIDIA is addressing this by designing chips, networking, security, and cooling as a single system. This improves throughput and efficiency across an entire rack instead of optimizing one component at a time.
For customers, this approach reduces training time, improves inference efficiency, and lowers the cost per token at the data center level.
Large Open Model Release
Alongside the keynote, NVIDIA released new open models, datasets, and tools covering multiple AI categories.
The released model families include:
