{"id":92537,"date":"2025-12-23T12:34:43","date_gmt":"2025-12-23T09:34:43","guid":{"rendered":"https:\/\/forklog.com\/en\/?p=92537"},"modified":"2025-12-23T12:35:21","modified_gmt":"2025-12-23T09:35:21","slug":"prompt-injection-emerges-as-a-major-threat-to-ai-browsers","status":"publish","type":"post","link":"https:\/\/forklog.com\/en\/prompt-injection-emerges-as-a-major-threat-to-ai-browsers\/","title":{"rendered":"Prompt Injection Emerges as a Major Threat to AI Browsers"},"content":{"rendered":"<p>OpenAI <a href=\"https:\/\/openai.com\/index\/hardening-atlas-against-prompt-injection\/\">disclosed<\/a> vulnerabilities in AI browsers and measures to bolster the security of its own solution \u2014 <a href=\"https:\/\/forklog.com\/en\/news\/openai-launches-ai-powered-browser-atlas\">Atlas<\/a>.<\/p>\n<p>The company acknowledged that &#8220;prompt injection&#8221; attacks, which manipulate agents into executing malicious instructions, pose a risk that is unlikely to disappear soon.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;Such vulnerabilities, like fraud and social engineering on the internet, are unlikely to be completely eradicated,&#8221; wrote OpenAI.<\/p>\n<\/blockquote>\n<p>It noted that the &#8220;agent mode&#8221; in Atlas &#8220;increases the threat surface.&#8221;<\/p>\n<p>Besides Sam Altman&#8217;s startup, other experts have also taken note of the issue. In early December, the UK&#8217;s National Cyber Security Centre <a href=\"https:\/\/www.ncsc.gov.uk\/news\/mistaking-ai-vulnerability-could-lead-to-large-scale-breaches\">warned<\/a> that attacks involving malicious prompt integration &#8220;will never disappear.&#8221; The government advised cybersecurity specialists to focus on mitigating risks and consequences rather than attempting to eliminate the problem.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;We view this as a long-term AI security issue and will continuously strengthen our defenses,&#8221; OpenAI noted.<\/p>\n<\/blockquote>\n<h2 class=\"wp-block-heading\">Countermeasures<\/h2>\n<p>Prompt injection is a method of manipulating AI by deliberately adding text to its input, causing it to ignore original instructions.<\/p>\n<p>OpenAI reported the implementation of a proactive rapid response cycle, which shows promising results in identifying new attack strategies before they appear &#8220;in real-world conditions.&#8221;<\/p>\n<p>Anthropic and Google <a href=\"https:\/\/security.googleblog.com\/2025\/12\/architecting-security-for-agentic.html\">express<\/a> similar views. Competitors suggest employing multi-layered defenses and conducting constant stress tests.<\/p>\n<p>OpenAI uses an &#8220;automated attacker based on <span data-descr=\"large language model\" class=\"old_tooltip\">LLM<\/span>&#8221; \u2014 an AI bot trained to act as a hacker seeking ways to infiltrate the agent with malicious prompts.<\/p>\n<p>The artificial fraudster can test the exploitation of vulnerabilities in a simulator, which demonstrates the actions of the compromised neural network. The bot then studies the response, adjusts its actions, and makes a second attempt, then a third, and so on.<\/p>\n<p>Outsiders do not have access to information about the internal reasoning of the target AI. In theory, the &#8220;virtual hacker&#8221; should find vulnerabilities faster than a real attacker.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;Our AI assistant can prompt the agent to execute complex, long-term malicious processes that unfold over dozens or even hundreds of steps. We have observed new attack strategies that did not manifest in our red team campaign or external reports,&#8221; states the OpenAI blog.<\/p>\n<\/blockquote>\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"578\" src=\"https:\/\/forklog.com\/wp-content\/uploads\/img-9f9b5c344fc0a040-1698611497246417-1024x578.png\" alt=\"image\" class=\"wp-image-272319\" srcset=\"https:\/\/forklog.com\/wp-content\/uploads\/img-9f9b5c344fc0a040-1698611497246417-1024x578.png 1024w, https:\/\/forklog.com\/wp-content\/uploads\/img-9f9b5c344fc0a040-1698611497246417-300x169.png 300w, https:\/\/forklog.com\/wp-content\/uploads\/img-9f9b5c344fc0a040-1698611497246417-768x434.png 768w, https:\/\/forklog.com\/wp-content\/uploads\/img-9f9b5c344fc0a040-1698611497246417.png 1194w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">Test demonstration. Source: OpenAI blog.<\/figcaption><\/figure>\n<p>In the example provided, the automated attacker sent an email to the user. The AI agent then scanned the email service and executed hidden instructions, sending a resignation message instead of composing an out-of-office reply.<\/p>\n<p>Following a security update, the &#8220;agent mode&#8221; was able to detect the sudden prompt injection attempt and flag it for the user.<\/p>\n<p>OpenAI emphasized that while it is challenging to reliably defend against such attacks, it relies on large-scale testing and rapid fix cycles.<\/p>\n<h2 class=\"wp-block-heading\">User Recommendations<\/h2>\n<p>Rami McCarthy, Chief Security Researcher at Wiz, <a href=\"https:\/\/techcrunch.com\/2025\/12\/22\/openai-says-ai-browsers-may-always-be-vulnerable-to-prompt-injection-attacks\/\">highlighted<\/a> that reinforcement learning is one of the main ways to continuously adapt to malicious behavior, but it is only part of the picture.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;A useful way to think about risks in AI systems is autonomy multiplied by access. Agent browsers are in a challenging part of this space: moderate autonomy combined with very high access. Many current recommendations reflect this trade-off. Limiting access after logging in primarily reduces vulnerability, while requiring confirmation request checks limits autonomy,&#8221; said the expert.<\/p>\n<\/blockquote>\n<p>These two recommendations were provided by OpenAI to users to reduce risk. The startup also suggested giving agents specific instructions rather than granting access to email and asking them to &#8220;take any necessary actions.&#8221;<\/p>\n<p>McCarthy noted that, as of today, browsers with built-in AI agents do not offer enough benefits to justify the risk profile.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;This balance will evolve, but today the trade-offs are still very real,&#8221; he concluded.<\/p>\n<\/blockquote>\n<p>Back in November, Microsoft experts <a href=\"https:\/\/forklog.com\/en\/news\/microsoft-identifies-ai-agent-vulnerabilities-following-extensive-testing\">introduced<\/a> a testing environment for AI agents and identified vulnerabilities inherent in modern digital assistants.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>OpenAI disclosed vulnerabilities in AI browsers and measures to bolster the security of its solution \u2014 Atlas.<\/p>\n","protected":false},"author":1,"featured_media":92538,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"select":"1","news_style_id":"1","cryptorium_level":"","_short_excerpt_text":"OpenAI disclosed vulnerabilities in AI browsers and security measures for Atlas.","creation_source":"","_metatest_mainpost_news_update":false,"footnotes":""},"categories":[3],"tags":[438,1111,1190],"class_list":["post-92537","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news-and-analysis","tag-artificial-intelligence","tag-cybersecurity","tag-openai"],"aioseo_notices":[],"amp_enabled":true,"views":"152","promo_type":"1","layout_type":"1","short_excerpt":"OpenAI disclosed vulnerabilities in AI browsers and security measures for Atlas.","is_update":"","_links":{"self":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/92537","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/comments?post=92537"}],"version-history":[{"count":1,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/92537\/revisions"}],"predecessor-version":[{"id":92539,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/92537\/revisions\/92539"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media\/92538"}],"wp:attachment":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media?parent=92537"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/categories?post=92537"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/tags?post=92537"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}