The gap between human ingenuity and artificial intelligence in cyberwarfare has officially closed. The Google Threat Intelligence Group (GTIG) recently published a landmark Q2 2026 report detailing the first confirmed case of a working zero-day exploit developed entirely through AI assistance.
A cybercriminal syndicate successfully utilized a Large Language Model (LLM) to craft a Python-based exploit that bypasses Two-Factor Authentication (2FA) in a popular open-source web administration tool. While Google disrupted the operation before it could be launched at scale, the incident proves that AI can now identify and exploit high-level semantic logic flaws that traditional security scanners miss.
Inside the AI Exploit: “Textbook Pythonic” Structure
How did researchers know the exploit was AI-generated? The code left behind a “digital fingerprint” unique to LLM training outputs:
- Educational Docstrings: The script contained an unusual abundance of helpful, lecture-style comments.
- Logic Over Memory: The flaw wasn’t a common “memory leak” but a complex logic vulnerability—a hardcoded trust assumption in the 2FA enforcement code.
- Hallucinated Data: The code included a completely hallucinated CVSS (severity) score, a classic sign of AI “imagination.”
State-Sponsored AI: From China to North Korea
Beyond common criminals, state-sponsored APT (Advanced Persistent Threat) groups are now using AI as a force multiplier:
- UNC2814 (PRC-linked): Used “persona-driven jailbreaking” to force Gemini to act as a senior C/C++ security expert to probe TP-Link firmware for flaws.
- APT45 (DPRK-linked): Sent thousands of automated prompts to recursively analyze known CVEs, creating an “AI-augmented arsenal” of exploits that would be impossible for humans to build manually.
- LONGSTREAM (Russia-linked): Deployed malware that uses AI to generate “decoy logic.” The code includes dozens of fake daylight-saving time queries to appear benign to security software.
PROMPTSPY: The Malware That Uses Gemini as a “Brain”
Perhaps the most alarming discovery is PROMPTSPY, an Android backdoor that integrates Google’s Gemini API directly into its core execution.
How it works:
- UI Serialization: The malware “sees” your phone screen and turns it into XML data.
- AI Command: It sends that data to Gemini-2.5-flash-lite, asking the AI what to do next.
- Autonomous Navigation: Gemini sends back structured JSON commands like CLICK or SWIPE, allowing the malware to navigate banking apps or delete security alerts without any human hacker involvement.
The AI Supply Chain Attack: Targeting the Gateway
In late March 2026, the group TeamPCP executed a coordinated supply chain attack targeting AI development tools like LiteLLM and BerriAI. By embedding a credential stealer called SANDCLOCK into these repositories, they harvested AWS keys and GitHub tokens directly from developer environments.
By compromising LiteLLM—a tool used to manage multiple AI providers—the hackers gained a master key to enterprise AI infrastructure, allowing them to pivot into corporate networks at will.
Google’s Defensive Response
Google isn’t just watching; it’s using AI to fight back. The company has deployed the “Big Sleep” agent to find vulnerabilities and “CodeMender” to automatically patch them before hackers can strike. Additionally, Google Play Protect has been updated to proactively block all known variants of PROMPTSPY.
Top Recommendations for 2026:
- Audit AI Dependencies: Treat tools like LiteLLM as “Tier 0” critical infrastructure.
- Rotate API Keys: Dyamically rotate Gemini, Claude, and OpenAI keys to prevent them from being pooled by hackers.
- Verify CI/CD Tokens: Ensure your GitHub tokens have the absolute minimum permissions needed to function.