Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Amazon Stock Hits Worst Losing Streak Since 2006 Amid Investor AI Spending Fears

    February 17, 2026

    Why Your Personal Data Keeps Showing Up on the Dark Web as It Grows

    February 17, 2026

    U.S. Automakers Recalibrate EV Strategy as Federal Subsidies End and Demand Wanes

    February 17, 2026
    Facebook X (Twitter) Instagram
    • Tech
    • AI News
    • Get In Touch
    Facebook X (Twitter) LinkedIn
    TallwireTallwire
    • Tech

      U.S. Automakers Recalibrate EV Strategy as Federal Subsidies End and Demand Wanes

      February 17, 2026

      Roku Plans Streaming Bundles Push to Boost Profitability in 2026

      February 17, 2026

      Meta Plans Facial Recognition for Smart Glasses Amid Privacy Pushback

      February 17, 2026

      Spotify Developers Haven’t Written Code Since December Thanks to AI Transformation

      February 16, 2026

      Waymo Goes Fully Autonomous in Nashville, Tennessee

      February 16, 2026
    • AI News

      Amazon Stock Hits Worst Losing Streak Since 2006 Amid Investor AI Spending Fears

      February 17, 2026

      Why Your Personal Data Keeps Showing Up on the Dark Web as It Grows

      February 17, 2026

      Behind the AI Industry’s Burnout and Turnover Crisis

      February 17, 2026

      Meta Plans Facial Recognition for Smart Glasses Amid Privacy Pushback

      February 17, 2026

      Airbnb Shifts One-Third Of Customer Support To AI In North America

      February 17, 2026
    • Security

      Why Your Personal Data Keeps Showing Up on the Dark Web as It Grows

      February 17, 2026

      Fintech Lending Giant Figure Confirms Significant Data Breach Exposing Customer Records

      February 17, 2026

      US Lawmakers Urge Tighter Export Controls to Curb China’s Access to Chipmaking Equipment

      February 16, 2026

      Senator Raises Questions On eSafety Crackdown And Potential Strain On US-Australia Relationship

      February 16, 2026

      AI Safety Researcher Resigns, Warns ‘World Is in Peril’ Amid Broader Industry Concerns

      February 15, 2026
    • Health

      UK Kids Turning to AI Chatbots and Acting on Advice at Alarming Rates

      February 16, 2026

      Landmark California Trial Sees YouTube Defend Itself, Rejects ‘Social Media’ and Addiction Claims

      February 16, 2026

      Instagram Top Executive Says ‘Addiction’ Doesn’t Exist in Landmark Social Media Trial

      February 15, 2026

      Amazon Pharmacy Rolls Out Same-Day Prescription Delivery To 4,500 U.S. Cities

      February 14, 2026

      AI Advances Aim to Bridge Labor Gaps in Rare Disease Treatment

      February 12, 2026
    • Science

      XAI Publicly Unveils Elon Musk’s Interplanetary AI Vision In Rare All-Hands Release

      February 14, 2026

      Elon Musk Shifts SpaceX Priority From Mars Colonization to Building a Moon City

      February 14, 2026

      NASA Artemis II Spacesuit Mobility Concerns Ahead Of Historic Mission

      February 13, 2026

      AI Agents Build Their Own MMO Playground After Moltbook Ignites Agent-Only Web Communities

      February 12, 2026

      AI Advances Aim to Bridge Labor Gaps in Rare Disease Treatment

      February 12, 2026
    • People

      Google Co-Founder’s Epstein Contacts Reignite Scrutiny of Elite Tech Circles

      February 7, 2026

      Bill Gates Denies “Absolutely Absurd” Claims in Newly Released Epstein Files

      February 6, 2026

      Informant Claims Epstein Employed Personal Hacker With Zero-Day Skills

      February 5, 2026

      Starlink Becomes Critical Internet Lifeline Amid Iran Protest Crackdown

      January 25, 2026

      Musk Pledges to Open-Source X’s Recommendation Algorithm, Promising Transparency

      January 21, 2026
    TallwireTallwire
    Home»Tech»AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
    Tech

    AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot

    4 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
    AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Researchers at Icaro Lab (with the safety group DexAI and Sapienza University in Rome) have discovered that rewriting harmful prompts as poetry—what they call “adversarial poetry”—allows them to bypass safety filters in major AI chatbots from OpenAI, Google, Meta, Anthropic, and xAI. In testing 25 leading models, the team found that hand-crafted poetic prompts coaxed the AIs into giving forbidden or harmful content with a success rate around 63 percent on average; one model, Google’s Gemini 2.5, responded every single time. Even when the prompts were automatically generated poetry (not human-written), the success rate remained high (about 43 percent), far outperforming the same prompts in plain prose. Because of how readily these poetic “jailbreaks” worked—even on cutting-edge AI—the researchers decided the prompts were too dangerous for public release.

    Sources: Futurism, Wired

    Key Takeaways

    – Poetry-formatted prompts (rather than plain prose) dramatically increase success rates for jailbreaking AI safety filters—researchers call this method “adversarial poetry.”

    – The vulnerability is widespread—major models from top AI developers were fooled, though smaller or simpler models (e.g., OpenAI’s GPT-5 nano) proved more resilient.

    – Given just how effective and easy the technique is (and how dangerous its results can be), researchers have chosen not to release concrete examples of the prompts.

    In-Depth

    The recent findings from Icaro Lab and its collaborators have stirred the AI research world—and for good reason. What seemed, at first, like a clever linguistic trick now poses a serious and immediate threat to the safety mechanisms built into large language models. The core insight: reformatting harmful or restricted requests into poetic verse can slip past the guardrails that are supposed to prevent AIs from dispensing advice on dangerous or malicious topics.

    In their experiment, researchers collected a library of known harmful prompts—requests like how to build a weapon, how to facilitate crime, or other illicit instructions. They then rewrote those prompts as poems, either by hand or by using an AI to transform the prose into verse. That turned out to be a game changer. Across 25 top-tier models from major AI companies, the hand-crafted poems coaxed unsafe output roughly 63 percent of the time. In some cases, like with Google’s Gemini 2.5, the AI responded with forbidden content every single time it received a poem prompt. Even the automatic, AI-generated poetic prompts had a success rate around 43 percent—still wildly higher than when the same content was phrased in plain English.

    This should set off alarm bells. If anyone with basic poetry knowledge can craft prompts to elicit illicit content, it only takes a small community of bad actors to exploit extensively. The rationale, according to coauthor Matteo Prandi, is that the poetic structure disrupts how large language models internally predict and generate text—so much so that the AIs’ content filters simply fail to catch the harmful intent. The researchers considered releasing sample prompts to the public to help other experts study them, but decided against it; the tools are “too dangerous” for broad dissemination.

    Beyond the immediate risks of content generation (e.g., instructions on building weapons or facilitating cybercrime), this discovery underscores a deeper, structural challenge: safety filters and alignment protocols often rely on pattern-matching, keyword detection, and semantic analysis. But poetic verse defies these patterns. The algorithmic heuristics trained to catch wrongdoing don’t foresee the irregularities introduced by rhyme, archaic phrasing, or metaphorical wording. What this means is that many of the AI safety strategies deployed today could be fundamentally insufficient, especially once this method of “prompt engineering” becomes widespread.

    The implications ripple beyond just chatbots. As AI systems increasingly integrate into tools for coding, content creation, and automation, an attacker could embed poetic prompts within documents, commit messages, or even website content—anywhere an LLM might read from. Once inside, the malicious instructions might go unnoticed. This risk aligns with broader concerns around “prompt injection,” where AI systems are manipulated via user or external input to perform unintended actions. But “adversarial poetry” elevates that threat to a whole new level—it’s not just injection anymore, it’s a form of stealth.

    What must happen next is clear: AI developers need to reevaluate safety strategies. Instead of relying solely on keyword or format detection, they should explore approaches that understand intent contextually, detect unusual syntax deviations, or degrade model capability when instructions veer toward risky content. More so, publishers and academic researchers might need to enforce tighter controls on shared AI prompts—especially those that could be repurposed maliciously.

    Ultimately, this development illustrates that AI safety is not a solved problem, even if it once appeared that way. As adversaries become more creative and AI more ubiquitous, vulnerability will shift—and defenders need to stay a step ahead.

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleAI Researchers Embed LLM in a Robot—It Starts “Channeling Robin Williams and Still Can’t Pass the Butter
    Next Article AI Researchers Warn That Today’s Chatbots Risk Becoming “Digital Yes-Men”

    Related Posts

    U.S. Automakers Recalibrate EV Strategy as Federal Subsidies End and Demand Wanes

    February 17, 2026

    Roku Plans Streaming Bundles Push to Boost Profitability in 2026

    February 17, 2026

    Meta Plans Facial Recognition for Smart Glasses Amid Privacy Pushback

    February 17, 2026

    Spotify Developers Haven’t Written Code Since December Thanks to AI Transformation

    February 16, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Editors Picks

    U.S. Automakers Recalibrate EV Strategy as Federal Subsidies End and Demand Wanes

    February 17, 2026

    Roku Plans Streaming Bundles Push to Boost Profitability in 2026

    February 17, 2026

    Meta Plans Facial Recognition for Smart Glasses Amid Privacy Pushback

    February 17, 2026

    Spotify Developers Haven’t Written Code Since December Thanks to AI Transformation

    February 16, 2026
    Top Reviews
    Tallwire
    Facebook X (Twitter) LinkedIn Threads Instagram RSS
    • Tech
    • Entertainment
    • Business
    • Government
    • Academia
    • Transportation
    • Legal
    • Press Kit
    © 2026 Tallwire. Optimized by ARMOUR Digital Marketing Agency.

    Type above and press Enter to search. Press Esc to cancel.