All posts
Trending Tech

Google Stops First AI-Generated Zero-Day Exploit Attack

Manaal Khan12 May 2026 at 10:08 am5 min read

Key Takeaways

  • Google intercepted what may be the first AI-generated zero-day exploit before a planned mass attack
  • The vulnerability bypassed two-factor authentication in an open-source web admin tool
  • Chinese and North Korean threat actors are increasingly using AI for vulnerability discovery

What Google Found

Google's Threat Intelligence Group (GTIG) says it disrupted what appears to be the first documented case of hackers using AI to create a working zero-day exploit. The attackers planned to use the vulnerability in a mass exploitation campaign before Google stepped in.

The flaw targeted two-factor authentication in what Google describes as a "popular open-source, web-based system administration tool." Google did not name the specific software. The company worked with the affected vendor to patch the vulnerability before attackers could deploy it at scale.

Google's team stopped short of confirming AI was definitely used. But they expressed high confidence based on evidence inside the exploit code.

Although we do not believe Gemini was used, based on the structure and content of these exploits, we have high confidence that the actor likely leveraged an AI model to support the discovery and weaponization of this vulnerability.

— Google Threat Intelligence Group

How Google Identified AI Involvement

The exploit code contained several telltale signs of AI generation. Google's analysts found unusually detailed educational-style comments throughout the code. These explanatory notes are typical of how AI models structure output when asked to write functional code.

The Python script also contained a hallucinated CVSS security score. CVSS scores are standardized vulnerability ratings. An AI model apparently invented one that didn't exist, a common error when language models fill in details they weren't given.

Google noted the vulnerability itself was a "high-level semantic logic flaw." This type of bug is harder to find through traditional automated scanning. It requires understanding what a developer intended the code to do, not just finding crashes or malformed inputs. AI models are increasingly capable of this contextual reasoning.

State-Backed Hackers Are Using AI for Exploit Development

The GTIG report documents a broader pattern. Chinese and North Korean threat actors have been experimenting with AI to speed up vulnerability discovery, exploit development, and automated testing.

Google observed attackers using carefully crafted prompts to make AI models act as security auditors. One example prompt instructed the AI to analyze router firmware for remote code execution vulnerabilities.

You are currently a network security expert specializing in embedded devices, specifically routers. I am currently researching a certain embedded device, and I have extracted its file system. I am auditing it for pre-authentication remote code execution (RCE) vulnerabilities.

— Example attacker prompt, per Google

This prompt engineering technique turns general-purpose AI models into specialized security tools. The attackers don't need to build custom AI systems. They manipulate existing models into performing expert-level analysis.

Also Read
Dirty Frag Exploit Gives Root Access on Nearly All Linux Distros

Another recent zero-day affecting open-source systems

The Wooyun-Legacy Problem

Attackers have also started using a specialized vulnerability database to train their AI-assisted attacks. A GitHub project called "wooyun-legacy" operates as a plugin for Claude Code, Anthropic's AI coding assistant.

The repository contains more than 85,000 real-world vulnerability cases collected from a Chinese bug bounty platform. By feeding this data to AI models, attackers can prime them to recognize similar flaws in new codebases.

Google explained the technique works through in-context learning. The vulnerability examples teach the model to approach code analysis like an experienced security researcher. This helps the AI identify logic flaws that a base model without this context might miss.

Also Read
Why AI Writes Better Rust Than Python

How AI models approach code generation differently by language

Why This Attack Matters

Zero-day exploits are valuable because they target unknown vulnerabilities. Defenders have no patch available. Victims have no warning. Finding these bugs traditionally requires significant expertise and time.

AI changes that calculus. A model that can analyze code semantically and understand developer intent can find logic flaws faster than manual review. The attackers in this case appear to have used AI not just to find the bug, but to write working exploit code.

The target, a 2FA bypass in a web admin tool, suggests the attackers wanted widespread access to systems protected by standard security measures. Mass exploitation of such a flaw could compromise thousands of servers running the affected software.

Also Read
Apple Brings End-to-End Encryption to iPhone-Android RCS in iOS 26.5

Related security development in authentication and encryption

What Happens Next

Google's intervention stopped this specific campaign. But the techniques are now documented and will likely spread. Other threat actors will study the approach and build on it.

Defensive teams should expect AI-assisted exploit development to become standard practice for sophisticated attackers. The barrier to finding complex logic flaws has dropped. Traditional scanning tools that look for known patterns will miss these semantic vulnerabilities.

Google did not specify which AI model the attackers used. The company noted it was not Gemini, Google's own AI. The wooyun-legacy project's integration with Claude Code suggests Anthropic's model may have been involved, though Google did not confirm this.

ℹ️

Logicity's Take

Frequently Asked Questions

What is a zero-day exploit?

A zero-day exploit targets a software vulnerability unknown to the vendor. There is no patch available when attackers use it. The name refers to defenders having zero days to prepare.

How did Google know the exploit was AI-generated?

The code contained educational-style comments, structured formatting, and a hallucinated CVSS security score. These patterns are common in AI-generated code but rare in human-written exploits.

Which software was targeted?

Google described it as a popular open-source, web-based system administration tool. The company did not name the specific software, likely to protect systems that haven't yet been patched.

Was the attack successful?

No. Google says it worked with the affected vendor to disclose the flaw before attackers could launch their planned mass exploitation campaign.

What AI model did the attackers use?

Google confirmed the attackers did not use Gemini. The company did not identify which AI model was used. The wooyun-legacy project's Claude Code integration suggests Anthropic's model may have been involved.

ℹ️

Need Help Implementing This?

Source: mint / Aman Gupta

M

Manaal Khan

Tech & Innovation Writer

Related Articles

Tesla's Remote Parking Feature: The Investigation That Didn't Quite Park Itself
Trending Tech·8 min

Tesla's Remote Parking Feature: The Investigation That Didn't Quite Park Itself

The US auto safety regulators have closed their investigation into Tesla's remote parking feature, but what does this mean for the future of autonomous driving? We dive into the details of the investigation and what it reveals about the technology. The National Highway Traffic Safety Administration found that crashes were rare and minor, but the investigation's closure doesn't necessarily mean the feature is completely safe.