Google is building multiple security layers into Chrome's upcoming AI agents before their public rollout, addressing concerns that autonomous browsing features could expose users to data theft and financial fraud. The company revealed a sophisticated defense system using observer models and strict user consent protocols for sensitive actions like banking and purchases.
Google just lifted the curtain on how it plans to keep Chrome's AI agents from going rogue when they launch in the coming months. The tech giant detailed a comprehensive security framework that puts multiple AI models to work monitoring each other - a necessary precaution as browsers race to deploy autonomous agents that can shop, book tickets, and navigate websites on your behalf.
The stakes couldn't be higher. These agentic browsing features promise to revolutionize how we interact with the web, but they also open new attack vectors that could drain bank accounts or steal sensitive data with a single malicious prompt. Google first previewed these capabilities in September, but the security details remained under wraps until now.
At the heart of Google's defense strategy sits what the company calls a User Alignment Critic - essentially an AI referee built using Gemini that scrutinizes every action another AI model plans to take. When the main planner model decides to click a button or fill out a form, the critic reviews the metadata and asks a crucial question: does this actually serve what the user wanted?
"If the critic model thinks that the planned tasks don't serve the user's goal, it asks the planner model to rethink the strategy," Google explained in its technical disclosure. Crucially, this oversight model never sees the actual web content - only the proposed actions - creating an additional privacy buffer.
But Google's security net extends far beyond AI-to-AI oversight. The company has developed something called Agent Origin Sets, which create rigid boundaries around what parts of websites the AI can access and interact with. Think of it as digital cordoning - the AI might be allowed to read product listings on a shopping site but blocked from touching banner ads or tracking pixels.
"This bounds the threat vector of cross-origin data leaks," Google noted, addressing one of the key vulnerabilities security researchers have identified with browser agents. "This also gives the browser the ability to enforce some of that separation, such as by not even sending to the model data that is outside the readable set."












