The AI chip wars just took an unexpected turn. Nvidia CEO Jensen Huang is preparing to showcase a new generation of CPU processors designed specifically for agentic AI at the company's GTC conference this week, marking a strategic pivot from the GPU-centric approach that propelled the chipmaker to a $2 trillion valuation. Both Nvidia and rival AMD are experiencing surging demand for traditional CPUs as enterprises discover that autonomous AI agents require fundamentally different computing architectures than the large language models that dominated 2024 and 2025.
Nvidia built a chipmaking empire on graphics processors, but the next phase of AI is forcing even Jensen Huang to rethink the playbook. At this week's GTC conference in San Jose, industry insiders expect the leather-jacket-wearing CEO to detail a new family of CPUs purpose-built for agentic AI - the autonomous software agents that reason, plan, and act independently rather than simply generating text responses.
The timing couldn't be more telling. While Nvidia has dominated AI infrastructure with its H100 and H200 GPUs that power OpenAI, Anthropic, and nearly every other major LLM, the emergence of agentic systems is exposing gaps in pure GPU architectures. These AI agents don't just process massive matrix calculations - they orchestrate workflows, manage memory, make sequential decisions, and interact with external tools. That requires the kind of general-purpose computing power that CPUs have delivered for decades.
AMD saw the shift coming first. The company's EPYC server processors have posted double-digit growth for six consecutive quarters, with CEO Lisa Su attributing much of the momentum to AI inference workloads during the company's recent earnings call. But Nvidia isn't ceding territory. The company acquired server CPU designer Arm-based startup Mellanox in 2020 for $7 billion and has been quietly building out its Grace CPU architecture alongside its GPU roadmap.
What makes agentic AI different comes down to workload characteristics. Training GPT-4 or Claude required massive parallel processing - exactly what GPUs excel at. But an AI agent booking your travel, managing your calendar, and coordinating with other agents spends more time on sequential logic, API calls, and state management. According to enterprise infrastructure data, these workloads show 60-70% CPU utilization versus 30-40% GPU, inverting the ratio seen in LLM training.












