A controversial research paper is throwing cold water on the AI industry's agent dreams. Published mid-2025, "Hallucination Stations: On Some Basic Limitations of Transformer-Based Language Models" claims to mathematically prove that large language models can't reliably handle complex computational and agentic tasks. But as Google, OpenAI, and dozens of startups pour billions into agent AI, they're betting the math is wrong - or at least incomplete.
The big AI companies promised 2025 would be "the year of the AI agents." It turned out to be the year of talking about AI agents. Now a research paper is suggesting the wait might be permanent.
Published without fanfare during the height of agent hype, "Hallucination Stations: On Some Basic Limitations of Transformer-Based Language Models" delivers a mathematical gut punch to the agentic AI vision. The paper, authored by former SAP CTO Vishal Sikka and his teenage prodigy son, claims to prove that LLMs are fundamentally incapable of carrying out computational and agentic tasks beyond a certain complexity. Even reasoning models that go beyond pure word prediction won't fix the problem, according to their analysis.
"There is no way they can be reliable," Sikka told Wired in a recent interview. The researcher, who studied under AI pioneer John McCarthy before his career at SAP, Infosys, and Oracle, now runs AI services startup Vianai. His verdict on agents running critical systems like nuclear power plants? Forget it. You might get one to file some papers and save time, but mistakes are inevitable.
The timing couldn't be more awkward for an industry that's bet its future on autonomous AI systems. Google's Demis Hassabis just reported breakthroughs in minimizing hallucinations at Davos this week, while hyperscalers and startups race to ship agent products. But the mathematical critique has support from an unlikely source - itself.









