On March 12, 2026, China's CERT (Computer Emergency Response Team) issued a public warning about severe security risks associated with the OpenClaw AI model. The national cybersecurity body cautioned that the model poses a significant threat, as it can be manipulated into performing harmful actions. These actions reportedly include the deletion of data, the exposure of sensitive information such as secret API keys, and the ability to load malicious content onto a user's system. The gravity of the warning is underscored by reports that the city of Beijing has moved to ban the use of the OpenClaw model. This event marks a critical moment in the governance of AI, where a national agency has formally identified a specific AI model as a direct security risk, moving beyond theoretical concerns to actionable warnings.
The warning from China's CERT did not specify CVEs but described functional vulnerabilities within the OpenClaw AI model that could be exploited by an attacker. This suggests issues with the model's safety alignment and its ability to interpret and refuse harmful instructions, a problem often referred to as 'prompt injection' or 'jailbreaking.'
These are not traditional software vulnerabilities but rather inherent risks in the way large language models process and act upon natural language inputs.
The warning implies that these manipulations are practical and repeatable. Furthermore, the report notes a related trend of malvertising campaigns impersonating popular AI agents, including OpenClaw and Claude Code, to distribute infostealing malware. In these campaigns, attackers use fake documentation pages to trick users into running malicious commands they believe are for installing or using the AI tool. This demonstrates that AI models are already being used as a powerful lure for social engineering.
The potential impact of these vulnerabilities is substantial. If an AI model can be reliably weaponized to delete data or execute code, it transforms from a productivity tool into a potential attack vector. An attacker could:
The ban in Beijing suggests that the Chinese government views this as a serious threat to national and corporate security.
Detecting misuse of an AI model is a novel challenge.
| Type | Value | Description | Context | Confidence |
|---|---|---|---|---|
| command_line_pattern | pip install [malicious_package] |
Malvertising campaigns trick users into installing malware via package managers, disguised as the AI tool. | Command-line logs, EDR logs | high |
| log_source | AI model audit logs / API logs | Monitor logs for unusual or suspicious prompts containing system commands, file paths, or attempts to break context. | Application-level logging | medium |
| file_name | Unexpected file deletion or creation | Monitor for file system changes that occur immediately after interaction with the AI model. | File integrity monitoring | medium |
C:\Windows, it should be blocked and flagged.D3-DA - Dynamic Analysis.Run AI models in isolated, containerized environments with strict limitations on file system and network access.
Mapped D3FEND Techniques:
Harden the configuration of applications that use AI models, ensuring they operate with the principle of least privilege.
Mapped D3FEND Techniques:
Educate users and developers about the risks of prompt injection and the importance of never trusting code or commands generated by an AI without thorough review.

Cybersecurity professional with over 10 years of specialized experience in security operations, threat intelligence, incident response, and security automation. Expertise spans SOAR/XSOAR orchestration, threat intelligence platforms, SIEM/UEBA analytics, and building cyber fusion centers. Background includes technical enablement, solution architecture for enterprise and government clients, and implementing security automation workflows across IR, TIP, and SOC use cases.
Help others stay informed about cybersecurity threats