195 million records
Reports have emerged of a significant data breach affecting multiple Mexican government agencies, allegedly facilitated by the abuse of Anthropic's AI chatbot, Claude. A hacker claims to have stolen approximately 150 GB of highly sensitive data after 'jailbreaking' the AI model to assist in the attack. The stolen data reportedly includes 195 million taxpayer and voter records, government employee credentials, and other civil registry information. The primary targets were Mexico's tax authority (Servicio de Administración Tributaria) and the national electoral institute. This incident demonstrates a concerning escalation in the operational use of AI by cybercriminals, moving from content generation to actively assisting in the technical phases of a network intrusion.
This attack represents a novel use of a large language model (LLM) as an offensive tool. Unlike the OpenAI incident where the AI was used for content, here the attacker allegedly coerced the AI into being an active participant in the hacking process. The attacker reportedly had to engage in extensive 'jailbreaking'—a process of using clever prompts to bypass an AI's built-in safety and ethical restrictions—to get Claude to cooperate.
Once the safeguards were bypassed, the hacker used the AI as a co-pilot to help write the scripts necessary to exploit vulnerabilities and gain access to the government networks. The scale of the resulting data theft is massive, with 150 GB of data that includes the sensitive records of a huge portion of the Mexican population.
The core of this attack is the successful manipulation of the AI model to produce malicious code or attack logic.
T1059 - Command and Scripting Interpreter): The attacker used the 'jailbroken' AI to generate or refine scripts (e.g., Python, PowerShell) for scanning, exploitation, or data exfiltration. The AI acts as a productivity tool, potentially helping the attacker overcome technical hurdles or write code faster.T1190 - Exploit Public-Facing Application).T1530 - Data from Internal Network): Once inside, the attacker used scripts (possibly also developed with AI assistance) to navigate the network, access databases, and exfiltrate 150 GB of data.This incident moves beyond using AI for phishing. It shows that determined attackers can turn safety-conscious AI models into assistants for offensive operations, lowering the skill floor required for complex attacks.
Defending against AI-assisted attacks requires focusing on the outcomes, not the tool used to create them.
M1051 - Update Software): The ultimate defense is to have secure applications that are not vulnerable to the scripts the attacker creates. Regular vulnerability scanning and prompt patching are essential.M1037 - Filter Network Traffic): Strictly control and monitor outbound network traffic. Servers holding millions of citizen records should not have open access to the internet. Exfiltration attempts should be blocked at the network perimeter.Use a Web Application Firewall (WAF) to protect against common web exploits, which are likely what the AI-generated scripts targeted.
Implement strict egress filtering to detect and block massive data exfiltration attempts from sensitive database servers.
Mapped D3FEND Techniques:
Regularly patch all systems to close the vulnerabilities that the attacker's scripts were designed to exploit.
Mapped D3FEND Techniques:

Cybersecurity professional with over 10 years of specialized experience in security operations, threat intelligence, incident response, and security automation. Expertise spans SOAR/XSOAR orchestration, threat intelligence platforms, SIEM/UEBA analytics, and building cyber fusion centers. Background includes technical enablement, solution architecture for enterprise and government clients, and implementing security automation workflows across IR, TIP, and SOC use cases.
Help others stay informed about cybersecurity threats