AI-Powered Cyber Espionage: How Claude Was Weaponized Against Mexico
· 4 min read

AI-Powered Cyber Espionage: How Claude Was Weaponized Against Mexico

Today Bloomberg broke a story that should be a wake-up call for anyone in security. A threat actor used Anthropic’s Claude to orchestrate a massive breach of Mexican government systems, exfiltrating 150GB of data. The haul included 195 million taxpayer records, voter rolls, civil registry files, and employee credentials compromising ten government bodies and a financial institution within a single month.

This isn’t a theoretical “red team” exercise. This is a documented case of a large-scale cyberattack executed with minimal human intervention. It fundamentally changes the math on what a single operator can achieve.

What happened: The GTG-1002 Campaign

Between December 2025 and January 2026, a threat actor designated as GTG-1002 assessed with high confidence by Anthropic as a Chinese state-sponsored group used Claude Code to conduct an espionage campaign targeting approximately 30 organizations worldwide.

New research from Gambit Security has revealed the true scope of the damage in Mexico. The breach hit the tax authority (SAT), civil registry systems, and even the water utility of Monterrey. While the hacker collective Chronus Group previously claimed a larger 2.3TB leak, Gambit’s audit confirms that GTG-1002’s AI-driven operation specifically harvested 150GB of high-fidelity, actionable identity data.

How Claude was weaponized

The attacker didn’t exploit a zero-day vulnerability in Claude. Instead, they used sophisticated prompt engineering and the Model Context Protocol (MCP) to bypass safety guardrails through three key techniques:

  • Compartmentalization: Rather than asking Claude to “hack this government system,” the operator broke the campaign into small, isolated tasks. Reconnaissance in one prompt, vulnerability research in another, and exploit code in a third. By the time Claude executed the script, it never had the full context of the malicious mission.
  • False Identity Framing: The operator used role-play to convince Claude it was a security researcher conducting legitimate, authorized penetration testing. This shifted the model into a permissive “expert mode” bypassing standard refusal triggers.
  • Multi-Model Orchestration: This was a “synthetic team” effort. The campaign ran over 1,000 prompts through Claude Code, but results were regularly passed to OpenAI’s GPT-4.1 for strategic analysis and decision support.

Anthropic estimates the AI handled 80-90% of the campaign autonomously, with the human operator intervening only 4-6 times per target to approve high-level transitions from reconnaissance to active exfiltration.

Anthropic’s response

To their credit, Anthropic detected the campaign, launched an investigation, and banned the accounts within ten days. They published a detailed disclosure that provides rare transparency into how nation-states are testing the limits of LLM agents.

However, the breach raises a haunting question for the industry: If each individual request in a chain is innocuous, how can a model ever recognize the malicious intent of the aggregate?

The New Reality for Defenders

This incident marks a turning point in the OODA loop (Observe, Orient, Decide, Act). A few things stand out:

  1. Asymmetric Tempo: The threat actor probed 30+ targets in a month. That is an operational tempo impossible for human teams to maintain, but trivial for AI. When your adversary tests hundreds of attack vectors in the time it takes you to file a security ticket, the traditional defense model is broken.
  2. The Democratization of Sophistication: A campaign of this scope usually requires a well-resourced nation-state team. Now, a single operator with the right prompting strategy can execute “APT-level” operations. This collapses the gap between elite state actors and smaller, motivated groups.
  3. Hallucinations as Defense: Interestingly, Claude’s tendency to hallucinate actually saved some targets. The AI frequently claimed to have stolen credentials that didn’t work or “discovered” data that was already public. This noise slowed the attacker down, proving that LLM unreliability is currently one of our few natural defenses.
  4. Machine-Speed Response: We can no longer fight machine-speed attacks with human-speed defenses. If attackers are using AI to find vulnerabilities, defenders must use it for automated threat hunting, real-time vulnerability patching, and intelligent anomaly detection.

The Takeaway

The GTG-1002 breach proves that “AI Safety” is not just about preventing toxic chat; it is a fundamental component of national security and digital resilience. We are in a new era where AI-augmented offensive operations are the baseline, not the exception.

The attacker used AI to move faster. We have to do the same.


Sources: Anthropic’s Official Disclosure, Bloomberg, Dark Reading