On March 12, 2026, China's CERT (Computer Emergency Response Team) issued a public warning about severe security risks associated with the OpenClaw AI model. The national cybersecurity body cautioned that the model poses a significant threat, as it can be manipulated into performing harmful actions. These actions reportedly include the deletion of data, the exposure of sensitive information such as secret API keys, and the ability to load malicious content onto a user's system. The gravity of the warning is underscored by reports that the city of Beijing has moved to ban the use of the OpenClaw model. This event marks a critical moment in the governance of AI, where a national agency has formally identified a specific AI model as a direct security risk, moving beyond theoretical concerns to actionable warnings.
The warning from China's CERT did not specify CVEs but described functional vulnerabilities within the OpenClaw AI model that could be exploited by an attacker. This suggests issues with the model's safety alignment and its ability to interpret and refuse harmful instructions, a problem often referred to as 'prompt injection' or 'jailbreaking.'
These are not traditional software vulnerabilities but rather inherent risks in the way large language models process and act upon natural language inputs.
The warning implies that these manipulations are practical and repeatable. Furthermore, the report notes a related trend of malvertising campaigns impersonating popular AI agents, including OpenClaw and Claude Code, to distribute infostealing malware. In these campaigns, attackers use fake documentation pages to trick users into running malicious commands they believe are for installing or using the AI tool. This demonstrates that AI models are already being used as a powerful lure for social engineering.
The potential impact of these vulnerabilities is substantial. If an AI model can be reliably weaponized to delete data or execute code, it transforms from a productivity tool into a potential attack vector. An attacker could:
The ban in Beijing suggests that the Chinese government views this as a serious threat to national and corporate security.
Detecting misuse of an AI model is a novel challenge.
pip install [malicious_package]C:\Windows, it should be blocked and flagged.D3-DA - Dynamic Analysis.Run AI models in isolated, containerized environments with strict limitations on file system and network access.
Mapped D3FEND Techniques:
Harden the configuration of applications that use AI models, ensuring they operate with the principle of least privilege.
Mapped D3FEND Techniques:
Educate users and developers about the risks of prompt injection and the importance of never trusting code or commands generated by an AI without thorough review.
China's national CERT is reported to have issued a security warning about the OpenClaw AI model.

Cybersecurity professional with over 10 years of specialized experience in security operations, threat intelligence, incident response, and security automation. Expertise spans SOAR/XSOAR orchestration, threat intelligence platforms, SIEM/UEBA analytics, and building cyber fusion centers. Background includes technical enablement, solution architecture for enterprise and government clients, and implementing security automation workflows across IR, TIP, and SOC use cases.
Help others stay informed about cybersecurity threats
Every tactic, technique, and sub-technique used in this threat has been identified and mapped to the MITRE ATT&CK framework for consistent, actionable threat language.
Observables and indicators of compromise (IOCs) have been extracted and cataloged. Risk has been assessed and correlated with known threat actors and historical campaigns.
Detection rules, incident response steps, and D3FEND-aligned mitigation strategies are included so your team can act on this intelligence immediately.
Structured threat data is packaged as a STIX 2.1 bundle and can be visualized as an interactive graph — relationships between actors, malware, techniques, and indicators.
Sigma detection rules are derived from the threat techniques in this article and can be converted for deployment across any major SIEM or EDR platform.