How Google Caught Cybercriminals Using AI to Build a Zero-Day
The Dawn of AI Exploits: Google’s First Confirmed Detection
On May 11, 2026, the Google Threat Intelligence Group (GTIG) released a landmark report confirming the first known instance of a working zero-day exploit developed with the assistance of artificial intelligence. The discovery marks a critical turning point in cybersecurity, where large language models (LLMs) are being used not just for phishing, but for high-level vulnerability research and weaponization.
The Anatomy of the “AI Signature”
The exploit, a Python-based script, targeted a semantic logic flaw in a popular open-source system administration tool. While the hackers attempted to keep it quiet, the code contained “telltale signs” that it was generated by an AI:
-
Hallucinated Metrics: The script included a hallucinated CVSS (severity) score, a common trait where LLMs “make up” plausible-sounding data points.
-
Textbook Formatting: The code used a highly structured, “textbook Pythonic” format with clean
_C ANSIcolor classes and detailed help menus—elements often found in LLM training data but rarely in handwritten criminal exploits. -
Educational Docstrings: Unlike typical hacker scripts, which are often messy or uncommented, this one featured an abundance of detailed, educational docstrings.
The Targets and the “Agentic” Threat
The report highlighted that while this specific criminal group failed due to implementation errors, state-sponsored actors are becoming far more sophisticated:
-
2FA Bypass: The AI-generated zero-day was designed to bypass Two-Factor Authentication (2FA), highlighting a shift toward attacking the very security layers we rely on most.
-
North Korean Automation (APT45): This group was observed sending thousands of repetitive, automated prompts to recursively analyze vulnerabilities, building an arsenal that would be impossible for humans to manage alone.
-
Chinese “Jailbreaking” (UNC2814): These actors used “persona-driven” jailbreaks, forcing AI to act as a “senior C++ security expert” to find flaws in TP-Link firmware and other infrastructure.
The Counter-Offensive: AI vs. AI
At zyproo.online, we focus on the “architectural” defense. Google is already deploying its own AI-driven counter-measures:
-
“Big Sleep”: An autonomous vulnerability discovery agent that hunts for bugs in open-source software before hackers can find them.
-
“CodeMender”: An experimental tool designed to automatically generate and apply patches to the vulnerabilities that “Big Sleep” finds.
-
“PROMPTSPY” Protection: Google Play Protect has been updated to defend Android devices against a new AI-driven backdoor that uses Gemini APIs to interpret on-screen UI elements.











