Google Identifies First AI-Generated Zero-Day Weaponized in the Wild
Google confirms the first documented case of an AI-developed zero-day exploit used in the wild, targeting a 2FA vulnerability in an open-source admin tool. The…

On May 11, the Google Threat Intelligence Group (GTIG) disclosed the first documented instance of a threat actor using an AI model to discover and weaponize a zero-day vulnerability in a real-world environment. The exploit, delivered via a Python script, bypasses two-factor authentication (2FA) on a widely used open-source web administration tool by exploiting a high-level semantic logic flaw. This official confirmation with "high confidence" from a primary vendor like Google shifts the conversation from theoretical alarmism to a measurable, active threat—one that was disrupted just prior to mass exploitation.
- The 2FA bypass does not rely on traditional memory corruption or input sanitization errors; instead, it targets a hard-coded trust assumption described by GTIG researchers as a semantic logic flaw. The attack requires valid user credentials to execute.
- Forensic analysis of the Python-based exploit revealed telltale LLM fingerprints, including extensive "educational" docstrings, a hallucinated CVSS score, and a "textbook Pythonic" structure featuring clean ANSI color classes.
- Google has categorically ruled out the use of its own Gemini model but has not identified the specific LLM used. The "high confidence" assessment is based on code artifacts and the sophisticated reasoning required to identify the specific vulnerability.
- The operation, categorized as a mass exploitation campaign, was dismantled before widespread impact thanks to a responsible disclosure process coordinated with the software vendor.
Exploiting Hard-Coded Trust Assumptions
According to GTIG, the exploit bypasses security not through memory corruption, but through a semantic logic failure. By leveraging existing valid credentials, the attacker manipulated the administrative tool's authorization logic, specifically targeting a hard-coded trust assumption within the code.
Researchers noted that while frontier LLMs often struggle with complex enterprise logic, they are demonstrating an increasing capacity for contextual reasoning. GTIG’s analysis suggests these models can interpret a developer's original intent and identify contradictions between 2FA enforcement logic and its hard-coded exceptions. Effectively, the AI identified an internal inconsistency in the code that human reviewers had long taken for granted.
Python Script Analysis Reveals Clear LLM Fingerprints
The primary artifact analyzed was the Python script used to execute the zero-day. GTIG identified stylistic characteristics rarely seen in manual code written by traditional threat actors. The file contained an "abundance of educational docstrings," a hallucinated CVSS score, and a highly structured formatting style described as "textbook Pythonic."
Notable details included comprehensive help menus and a clean, organized _C class for ANSI colors. While these elements do not contribute to the exploit's functionality, they serve as a signature of text generated from typical LLM training datasets. Furthermore, the inclusion of an invented CVSS score confirms that the model produced content that was statistically plausible but technically inaccurate.
Attribution: High Confidence Despite Unnamed Model
The Google Threat Intelligence Group issued a "high confidence" assessment regarding the use of an AI model in the discovery and weaponization phases. While the analysis explicitly excludes Gemini, the specific LLM remains unidentified. It is currently impossible to determine the exact engine used or the degree of human oversight involved in the process.
The attribution rests on the convergence of two factors: the nature of the flaw discovered—which requires contextual reasoning regarding authorization logic—and the specific stylistic artifacts found within the code. This evidence allowed GTIG to draw a direct line between LLM output and a malicious artifact found in the wild, moving beyond mere hypothesis.
Campaign Disrupted Before Mass Exploitation
The operation was identified and neutralized before it could scale into a full-blown mass exploitation event. Google collaborated with the vendor of the open-source tool (who remains unnamed) to patch the vulnerability and dismantle the campaign. It is not currently known if the exploit was successfully used against specific targets prior to intervention, nor how many actors were involved in the criminal network.
While responsible disclosure successfully mitigated the immediate risk, this case proves that the window between the discovery of a logical flaw and its weaponization has been drastically compressed by artificial intelligence.
"There’s a misconception that the AI vulnerability race is imminent. The reality is that it’s already begun. For every zero-day we can trace back to AI, there are probably many more out there" — John Hultquist, chief analyst at GTIG
Defense and Mitigation Strategies
Audit Hard-Coded Exceptions: Organizations should audit authorization logic and static trust assumptions in web-based administration tools, particularly where 2FA bypasses or conditional exceptions exist.
Re-evaluate 2FA Assumptions: While 2FA remains essential, it should be reinforced with behavioral analytics and post-authentication monitoring, as this exploit targets the session after initial access with valid credentials.
Strengthen Disclosure Channels: Maintain direct lines with security vendors and participate in responsible disclosure programs to reduce exposure windows, given that LLMs accelerate the timeline from discovery to weaponization.
Deploy AI Defensively: Use similar contextual reasoning tools to identify logical contradictions in internal software before threat actors can exploit them.
Google’s forensic documentation effectively closes the gap between academic proof-of-concept and criminal weaponization. Frontier LLMs are no longer just helping write payloads; they are reasoning through complex authorization logic to find exploitable contradictions. For enterprises, this means the next critical vulnerability may not come from a fuzzer or a human reverse engineer, but from a model that has analyzed code with a level of patience and consistency that is difficult to replicate manually.
Frequently Asked Questions
How does this zero-day differ from standard AI-generated scripts?
Google’s attribution focuses not just on code generation, but on the discovery and weaponization of a semantic logic flaw. An LLM successfully identified a hard-coded trust assumption and built a functional exploit around it.
How dangerous is this exploit if it requires valid credentials?
It poses a significant threat to system administration tools. Initial access is often gained through phishing or password reuse; bypassing 2FA allows an attacker to escalate control without facing the second layer of security.
What was the level of human involvement in creating the exploit?
The exact degree of human supervision remains unquantified. GTIG confirmed the use of AI but has not established whether a human operator guided the model, corrected its output, or if the model acted largely autonomously.
Information has been verified against cited sources and is current as of the time of publication.
Sources
- https://thehackernews.com/2026/05/hackers-used-ai-to-develop-first-known.html
- https://www.tradingview.com/news/cointelegraph:b7efb500b094b:0-hackers-used-ai-to-craft-zero-day-attack-to-bypass-2fa-google/
- https://www.bleepingcomputer.com/news/security/google-hackers-used-ai-to-develop-zero-day-exploit-for-web-admin-tool/
- https://www.infosecurity-magazine.com/news/hackers-using-ai-zero-day-first/
- https://www.csoonline.com/article/4169046/google-discovers-weaponized-zero-day-exploits-created-with-ai.html