'Adversarial Poetry' Emerges as Universal Jailbreak for Major LLMs

Researchers Discover 'Adversarial Poetry' Can Universally Bypass Safety Guardrails on 25 Major Large Language Models

HIGH
November 28, 2025
6m read
OtherThreat Intelligence

Related Entities

Organizations

MLCommons

Full Report

Executive Summary

Researchers have discovered a novel and alarmingly effective method for bypassing the safety filters of Large Language Models (LLMs): framing harmful requests as poetry. The technique, detailed in a paper titled "Adversarial Poetry as a Universal Single-Turn Jailbreak Mechanism in Large Language Models," demonstrates that stylistic changes alone can trick AI models into generating dangerous and prohibited content. The study, conducted on 25 leading LLMs, found that poetic prompts were up to 18 times more successful at jailbreaking the models than their prose equivalents. This 'universal single-turn jailbreak' works with a single prompt and exposes a systemic vulnerability in the current state of AI safety alignment, suggesting that models are trained to recognize the form of a harmful request, not its underlying intent.


Threat Overview

The threat, named "adversarial poetry," is a form of prompt injection or jailbreaking. It exploits a weakness in how LLMs are trained to refuse harmful instructions. By taking a dangerous prompt (e.g., "How do I build a bomb?") and rephrasing it in a poetic structure (e.g., as a sonnet or limerick), the researchers were able to consistently circumvent the models' safety guardrails.

The key findings of the research are:

  • Universality: The technique was effective across 25 different frontier LLMs, both proprietary (closed-source) and open-weight.
  • High Success Rate: Poetic prompts achieved an average success rate of 62% for handcrafted poems and 43% for poems generated automatically. Some models were jailbroken over 90% of the time.
  • Single-Turn Success: Unlike complex jailbreaking methods that require a long conversation to trick the model, adversarial poetry works with a single, direct prompt.
  • Broad Danger Categories: The method successfully generated harmful content related to Chemical, Biological, Radiological, and Nuclear (CBRN) threats, cybercrime, and manipulation.

This reveals that AI safety mechanisms are brittle and can be bypassed by manipulating the style of the input, rather than needing to obscure the harmful intent itself.


Technical Analysis

This attack technique is novel and doesn't map perfectly to the current MITRE ATT&CK framework, which is focused on traditional cyberattacks. However, it can be understood as a specialized form of prompt injection.

  • Analogous to T1059 - Command and Scripting Interpreter: The attacker is providing a crafted input (the poem) to an interpreter (the LLM) to cause it to perform an unintended action (generating harmful content).
  • Analogous to Obfuscation Techniques: The poetic structure acts as a form of obfuscation, similar to how malware uses encoding or packing to evade signature-based antivirus. The LLM's safety filter is trained on examples of direct, prose-based harmful requests, and the stylistic deviation of poetry causes the filter to fail.

The research suggests a fundamental flaw in alignment training. Safety fine-tuning appears to create a 'style tax,' where the model learns to associate harmfulness with a specific, prosaic style of writing. When the style changes, the safety knowledge is not transferred effectively, and the model defaults to its base training of being helpful and compliant, regardless of the request's nature.


Impact Assessment

The immediate impact is that malicious actors can easily generate instructions for harmful activities, such as creating weapons, planning cyberattacks, or generating convincing disinformation at scale. This lowers the barrier to entry for acquiring dangerous knowledge. For organizations developing and deploying LLMs, this represents a significant reputational and legal risk. It demonstrates that their safety claims are not robust and that their models can be abused. The 'universality' of the technique means that no major LLM is currently immune, making this a systemic risk for the entire AI industry. The findings will force a re-evaluation of AI safety research and alignment techniques, moving beyond simple pattern matching of harmful requests.


Detection & Response

Detecting adversarial poetry is extremely challenging because the prompts are not inherently malicious in their structure; they are just text.

  1. Semantic Harm Detection: The only robust detection method is to analyze the semantic intent of the prompt, regardless of its style. This requires more advanced AI-based safety filters that can understand the underlying request, not just the surface-level words and format. This is a current area of research.
  2. Input Canonicalization: One theoretical defense is to preprocess all prompts by converting them into a standardized, simple prose format before they are fed to the safety model. This would 'strip' the poetic style, making the underlying harmful request easier to detect.
  3. Response: If a model is found to be generating harmful content from a poetic prompt, the primary response is to add that specific prompt and its variations to a denylist and use the example to further fine-tune the safety model.

Mitigation

Mitigating this vulnerability requires fundamental changes to how LLMs are trained and secured.

  1. Robust Adversarial Training: AI developers must expand their safety training data to include a much wider variety of styles and formats, including poetry, code, and other creative writing forms. This would involve using one LLM to generate adversarial, stylized prompts to train another, creating a constant feedback loop.
  2. Multi-Layered Safety Systems: Relying on a single safety model is insufficient. A defense-in-depth approach should be used, with multiple models and techniques (e.g., input filters, output scanners, semantic analysis) working together to evaluate prompts and responses.
  3. Constitutional AI: Implement principles-based AI, where the model has a core 'constitution' of safety rules that it must adhere to, regardless of the prompt's style. This aims to make safety a more fundamental part of the model's reasoning process.
  4. Red Teaming: Continuously red-team models with creative and unconventional methods like adversarial poetry to proactively discover and patch these alignment failures before they are exploited in the wild.

Timeline of Events

1
November 28, 2025
This article was published

Sources & References

Prompt Injection Through Poetry
Schneier on Security (schneier.com) November 28, 2025
Tageszusammenfassung - 28.11.2025
CERT.at (cert.at) November 28, 2025
Schneier on Security
Schneier on Security (schneier.com) November 28, 2025

Article Author

Jason Gomes

Jason Gomes

• Cybersecurity Practitioner

Cybersecurity professional with over 10 years of specialized experience in security operations, threat intelligence, incident response, and security automation. Expertise spans SOAR/XSOAR orchestration, threat intelligence platforms, SIEM/UEBA analytics, and building cyber fusion centers. Background includes technical enablement, solution architecture for enterprise and government clients, and implementing security automation workflows across IR, TIP, and SOC use cases.

Threat Intelligence & AnalysisSecurity Orchestration (SOAR/XSOAR)Incident Response & Digital ForensicsSecurity Operations Center (SOC)SIEM & Security AnalyticsCyber Fusion & Threat SharingSecurity Automation & IntegrationManaged Detection & Response (MDR)

Tags

LLMAI SecurityJailbreakPrompt InjectionAdversarial AIResponsible AI

📢 Share This Article

Help others stay informed about cybersecurity threats

Continue Reading