Google Stops First AI-Generated Zero-Day Exploit Attack
Key Takeaways
- Google intercepted what may be the first AI-generated zero-day exploit before a planned mass attack
- The vulnerability bypassed two-factor authentication in an open-source web admin tool
- Chinese and North Korean threat actors are increasingly using AI for vulnerability discovery
What Google Found
Google's Threat Intelligence Group (GTIG) says it disrupted what appears to be the first documented case of hackers using AI to create a working zero-day exploit. The attackers planned to use the vulnerability in a mass exploitation campaign before Google stepped in.
The flaw targeted two-factor authentication in what Google describes as a "popular open-source, web-based system administration tool." Google did not name the specific software. The company worked with the affected vendor to patch the vulnerability before attackers could deploy it at scale.
Google's team stopped short of confirming AI was definitely used. But they expressed high confidence based on evidence inside the exploit code.
“Although we do not believe Gemini was used, based on the structure and content of these exploits, we have high confidence that the actor likely leveraged an AI model to support the discovery and weaponization of this vulnerability.”
— Google Threat Intelligence Group
How Google Identified AI Involvement
The exploit code contained several telltale signs of AI generation. Google's analysts found unusually detailed educational-style comments throughout the code. These explanatory notes are typical of how AI models structure output when asked to write functional code.
The Python script also contained a hallucinated CVSS security score. CVSS scores are standardized vulnerability ratings. An AI model apparently invented one that didn't exist, a common error when language models fill in details they weren't given.
Google noted the vulnerability itself was a "high-level semantic logic flaw." This type of bug is harder to find through traditional automated scanning. It requires understanding what a developer intended the code to do, not just finding crashes or malformed inputs. AI models are increasingly capable of this contextual reasoning.
State-Backed Hackers Are Using AI for Exploit Development
The GTIG report documents a broader pattern. Chinese and North Korean threat actors have been experimenting with AI to speed up vulnerability discovery, exploit development, and automated testing.
Google observed attackers using carefully crafted prompts to make AI models act as security auditors. One example prompt instructed the AI to analyze router firmware for remote code execution vulnerabilities.
“You are currently a network security expert specializing in embedded devices, specifically routers. I am currently researching a certain embedded device, and I have extracted its file system. I am auditing it for pre-authentication remote code execution (RCE) vulnerabilities.”
— Example attacker prompt, per Google
This prompt engineering technique turns general-purpose AI models into specialized security tools. The attackers don't need to build custom AI systems. They manipulate existing models into performing expert-level analysis.
Another recent zero-day affecting open-source systems
The Wooyun-Legacy Problem
Attackers have also started using a specialized vulnerability database to train their AI-assisted attacks. A GitHub project called "wooyun-legacy" operates as a plugin for Claude Code, Anthropic's AI coding assistant.
The repository contains more than 85,000 real-world vulnerability cases collected from a Chinese bug bounty platform. By feeding this data to AI models, attackers can prime them to recognize similar flaws in new codebases.
Google explained the technique works through in-context learning. The vulnerability examples teach the model to approach code analysis like an experienced security researcher. This helps the AI identify logic flaws that a base model without this context might miss.
How AI models approach code generation differently by language
Why This Attack Matters
Zero-day exploits are valuable because they target unknown vulnerabilities. Defenders have no patch available. Victims have no warning. Finding these bugs traditionally requires significant expertise and time.
AI changes that calculus. A model that can analyze code semantically and understand developer intent can find logic flaws faster than manual review. The attackers in this case appear to have used AI not just to find the bug, but to write working exploit code.
The target, a 2FA bypass in a web admin tool, suggests the attackers wanted widespread access to systems protected by standard security measures. Mass exploitation of such a flaw could compromise thousands of servers running the affected software.
Related security development in authentication and encryption
What Happens Next
Google's intervention stopped this specific campaign. But the techniques are now documented and will likely spread. Other threat actors will study the approach and build on it.
Defensive teams should expect AI-assisted exploit development to become standard practice for sophisticated attackers. The barrier to finding complex logic flaws has dropped. Traditional scanning tools that look for known patterns will miss these semantic vulnerabilities.
Google did not specify which AI model the attackers used. The company noted it was not Gemini, Google's own AI. The wooyun-legacy project's integration with Claude Code suggests Anthropic's model may have been involved, though Google did not confirm this.
Logicity's Take
Frequently Asked Questions
What is a zero-day exploit?
A zero-day exploit targets a software vulnerability unknown to the vendor. There is no patch available when attackers use it. The name refers to defenders having zero days to prepare.
How did Google know the exploit was AI-generated?
The code contained educational-style comments, structured formatting, and a hallucinated CVSS security score. These patterns are common in AI-generated code but rare in human-written exploits.
Which software was targeted?
Google described it as a popular open-source, web-based system administration tool. The company did not name the specific software, likely to protect systems that haven't yet been patched.
Was the attack successful?
No. Google says it worked with the affected vendor to disclose the flaw before attackers could launch their planned mass exploitation campaign.
What AI model did the attackers use?
Google confirmed the attackers did not use Gemini. The company did not identify which AI model was used. The wooyun-legacy project's Claude Code integration suggests Anthropic's model may have been involved.
Need Help Implementing This?
Source: mint / Aman Gupta
Manaal Khan
Tech & Innovation Writer
Related Articles
Browse all
Robotaxi Companies Are Hiding How Often Humans Take the Wheel
Autonomous vehicle firms like Waymo and Tesla are under scrutiny for refusing to disclose how often remote operators step in to control their self-driving cars. A Senate investigation reveals major gaps in transparency, raising safety and accountability concerns.

Wisconsin Governor Throws a Wrench in Age Verification Plans
Wisconsin Governor Tony Evers has vetoed a bill that would have required residents to verify their age before accessing adult content online, citing concerns over privacy and data security. This move comes as several other states have already implemented similar age check requirements. The veto has significant implications for the future of online age verification.

Apple's App Store Empire Under Siege: The Battle for the Future of Tech
The long-running feud between Apple and Epic Games has reached a boiling point, with Apple preparing to take its case to the Supreme Court. The tech giant is fighting to maintain control over its App Store, while Epic Games is pushing for more freedom for developers. The outcome could have far-reaching implications for the entire tech industry.

Tesla's Remote Parking Feature: The Investigation That Didn't Quite Park Itself
The US auto safety regulators have closed their investigation into Tesla's remote parking feature, but what does this mean for the future of autonomous driving? We dive into the details of the investigation and what it reveals about the technology. The National Highway Traffic Safety Administration found that crashes were rare and minor, but the investigation's closure doesn't necessarily mean the feature is completely safe.
Also Read

Why AI Writes Better Rust Than Python
For a decade, Python won because it was fast to ship. But AI coding assistants have flipped the script. Strong type systems and compiler feedback loops make Rust, Go, and Swift easier for AI agents than dynamic languages. The evidence is piling up in production codebases.

Apple Brings End-to-End Encryption to iPhone-Android RCS in iOS 26.5
Apple and Google have jointly rolled out end-to-end encryption for RCS messages between iPhones and Android devices. The feature, arriving in iOS 26.5, means carriers and tech companies can no longer read cross-platform texts. It's the first time a major messaging protocol has offered interoperable E2EE across different providers.

Samsung May Use BOE Displays for Galaxy S27
Samsung is reportedly considering Chinese display maker BOE as a secondary supplier for the Galaxy S27. The move could help Samsung Electronics cut costs on panels and offset rising memory prices, though internal politics between Samsung divisions may complicate the decision.