Close Menu

    Subscribe to Updates

    Get the latest tech news from Tallwire.

      What's Hot

      AI Infrastructure Investment Surges With Multi-Billion Dollar Data Center Deals

      March 2, 2026

      Netflix Backs Off Warner Bros. Deal As Paramount’s Higher Bid Prevails

      March 2, 2026

      Major Cybercrime Group Claims Theft Of 1.7 Million CarGurus Corporate Records

      March 1, 2026
      Facebook X (Twitter) Instagram
      • Tech
      • AI
      • Get In Touch
      Facebook X (Twitter) LinkedIn
      TallwireTallwire
      • Tech

        Amazon Overtakes Walmart As America’s Largest Company By Revenue

        March 1, 2026

        Chinese Sellers Peddling Anti-Drone Weapons On TikTok Raise Security Alarms

        March 1, 2026

        Say Goodbye to the Undersea Cable That Made the Global Internet Possible

        March 1, 2026

        Microsoft Copilot Bug Exposed “Confidential” Emails Despite Label

        February 28, 2026

        Taara Beam Launch Brings 25Gbps Optical Wireless Networks to Cities

        February 27, 2026
      • AI

        AI Infrastructure Investment Surges With Multi-Billion Dollar Data Center Deals

        March 2, 2026

        Study Signals AI Search Shift Threatens Traditional Web Traffic Model

        March 1, 2026

        Amazon’s Security Chief Warns AI Will Flood Data, Expand Cyber Risk

        March 1, 2026

        AI Password Generation Poses Major Security Risk, Experts Warn

        February 28, 2026

        Microsoft Copilot Bug Exposed “Confidential” Emails Despite Label

        February 28, 2026
      • Security

        Major Cybercrime Group Claims Theft Of 1.7 Million CarGurus Corporate Records

        March 1, 2026

        Google Cracks Down On Android Apps And Developer Accounts In 2025

        March 1, 2026

        Massive Exposed Database With Billions of Social Security Numbers Sparks Identity Theft Fears

        March 1, 2026

        Amazon’s Security Chief Warns AI Will Flood Data, Expand Cyber Risk

        March 1, 2026

        Password Managers Share a Hidden Weakness

        March 1, 2026
      • Health

        Social Media Addiction Trial Draws Grieving Parents Seeking Accountability From Tech Platforms

        February 19, 2026

        Portugal’s Parliament OKs Law to Restrict Children’s Social Media Access With Parental Consent

        February 18, 2026

        Parents Paint 108 Names, Demand Snapchat Reform After Deadly Fentanyl Claims

        February 18, 2026

        UK Kids Turning to AI Chatbots and Acting on Advice at Alarming Rates

        February 16, 2026

        Landmark California Trial Sees YouTube Defend Itself, Rejects ‘Social Media’ and Addiction Claims

        February 16, 2026
      • Science

        Astronomers Confirm Discovery Of Galaxy Nearly Entirely Composed Of Dark Matter

        March 1, 2026

        Microsoft Claims 100 Percent Renewable Energy Match Across Global Electricity Use

        February 28, 2026

        Taara Beam Launch Brings 25Gbps Optical Wireless Networks to Cities

        February 27, 2026

        Large Hadron Collider Enters Third Shutdown For Major Upgrade

        February 26, 2026

        Google Phases Out Android’s Built-In Weather App, Replacing It With Search-Based Forecasts

        February 25, 2026
      • Tech

        Sam Altman Says ‘AI Washing’ Is Being Used to Mask Corporate Layoffs

        February 28, 2026

        Zuckerberg Testifies In Landmark Trial Over Alleged Teen Social Media Harms

        February 23, 2026

        Gay Tech Networks Under Spotlight In Silicon Valley Culture Debate

        February 23, 2026

        Google Co-Founder’s Epstein Contacts Reignite Scrutiny of Elite Tech Circles

        February 7, 2026

        Bill Gates Denies “Absolutely Absurd” Claims in Newly Released Epstein Files

        February 6, 2026
      TallwireTallwire
      Home»Tech»AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
      Tech

      AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot

      4 Mins Read
      Facebook Twitter Pinterest LinkedIn Tumblr Email
      AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
      AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
      Share
      Facebook Twitter LinkedIn Pinterest Email

      Researchers at Icaro Lab (with the safety group DexAI and Sapienza University in Rome) have discovered that rewriting harmful prompts as poetry—what they call “adversarial poetry”—allows them to bypass safety filters in major AI chatbots from OpenAI, Google, Meta, Anthropic, and xAI. In testing 25 leading models, the team found that hand-crafted poetic prompts coaxed the AIs into giving forbidden or harmful content with a success rate around 63 percent on average; one model, Google’s Gemini 2.5, responded every single time. Even when the prompts were automatically generated poetry (not human-written), the success rate remained high (about 43 percent), far outperforming the same prompts in plain prose. Because of how readily these poetic “jailbreaks” worked—even on cutting-edge AI—the researchers decided the prompts were too dangerous for public release.

      Sources: Futurism, Wired

      Key Takeaways

      – Poetry-formatted prompts (rather than plain prose) dramatically increase success rates for jailbreaking AI safety filters—researchers call this method “adversarial poetry.”

      – The vulnerability is widespread—major models from top AI developers were fooled, though smaller or simpler models (e.g., OpenAI’s GPT-5 nano) proved more resilient.

      – Given just how effective and easy the technique is (and how dangerous its results can be), researchers have chosen not to release concrete examples of the prompts.

      In-Depth

      The recent findings from Icaro Lab and its collaborators have stirred the AI research world—and for good reason. What seemed, at first, like a clever linguistic trick now poses a serious and immediate threat to the safety mechanisms built into large language models. The core insight: reformatting harmful or restricted requests into poetic verse can slip past the guardrails that are supposed to prevent AIs from dispensing advice on dangerous or malicious topics.

      In their experiment, researchers collected a library of known harmful prompts—requests like how to build a weapon, how to facilitate crime, or other illicit instructions. They then rewrote those prompts as poems, either by hand or by using an AI to transform the prose into verse. That turned out to be a game changer. Across 25 top-tier models from major AI companies, the hand-crafted poems coaxed unsafe output roughly 63 percent of the time. In some cases, like with Google’s Gemini 2.5, the AI responded with forbidden content every single time it received a poem prompt. Even the automatic, AI-generated poetic prompts had a success rate around 43 percent—still wildly higher than when the same content was phrased in plain English.

      This should set off alarm bells. If anyone with basic poetry knowledge can craft prompts to elicit illicit content, it only takes a small community of bad actors to exploit extensively. The rationale, according to coauthor Matteo Prandi, is that the poetic structure disrupts how large language models internally predict and generate text—so much so that the AIs’ content filters simply fail to catch the harmful intent. The researchers considered releasing sample prompts to the public to help other experts study them, but decided against it; the tools are “too dangerous” for broad dissemination.

      Beyond the immediate risks of content generation (e.g., instructions on building weapons or facilitating cybercrime), this discovery underscores a deeper, structural challenge: safety filters and alignment protocols often rely on pattern-matching, keyword detection, and semantic analysis. But poetic verse defies these patterns. The algorithmic heuristics trained to catch wrongdoing don’t foresee the irregularities introduced by rhyme, archaic phrasing, or metaphorical wording. What this means is that many of the AI safety strategies deployed today could be fundamentally insufficient, especially once this method of “prompt engineering” becomes widespread.

      The implications ripple beyond just chatbots. As AI systems increasingly integrate into tools for coding, content creation, and automation, an attacker could embed poetic prompts within documents, commit messages, or even website content—anywhere an LLM might read from. Once inside, the malicious instructions might go unnoticed. This risk aligns with broader concerns around “prompt injection,” where AI systems are manipulated via user or external input to perform unintended actions. But “adversarial poetry” elevates that threat to a whole new level—it’s not just injection anymore, it’s a form of stealth.

      What must happen next is clear: AI developers need to reevaluate safety strategies. Instead of relying solely on keyword or format detection, they should explore approaches that understand intent contextually, detect unusual syntax deviations, or degrade model capability when instructions veer toward risky content. More so, publishers and academic researchers might need to enforce tighter controls on shared AI prompts—especially those that could be repurposed maliciously.

      Ultimately, this development illustrates that AI safety is not a solved problem, even if it once appeared that way. As adversaries become more creative and AI more ubiquitous, vulnerability will shift—and defenders need to stay a step ahead.

      Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
      Previous ArticleAI Researchers Embed LLM in a Robot—It Starts “Channeling Robin Williams and Still Can’t Pass the Butter
      Next Article AI Researchers Warn That Today’s Chatbots Risk Becoming “Digital Yes-Men”

      Related Posts

      Amazon Overtakes Walmart As America’s Largest Company By Revenue

      March 1, 2026

      Chinese Sellers Peddling Anti-Drone Weapons On TikTok Raise Security Alarms

      March 1, 2026

      Say Goodbye to the Undersea Cable That Made the Global Internet Possible

      March 1, 2026

      Microsoft Copilot Bug Exposed “Confidential” Emails Despite Label

      February 28, 2026
      Add A Comment
      Leave A Reply Cancel Reply

      Editors Picks

      Amazon Overtakes Walmart As America’s Largest Company By Revenue

      March 1, 2026

      Chinese Sellers Peddling Anti-Drone Weapons On TikTok Raise Security Alarms

      March 1, 2026

      Say Goodbye to the Undersea Cable That Made the Global Internet Possible

      March 1, 2026

      Microsoft Copilot Bug Exposed “Confidential” Emails Despite Label

      February 28, 2026
      Popular Topics
      spotlight picks Qualcomm Sam Altman Series A UAE Tech Satya Nadella Startup Tesla Cybertruck trending Tesla Tim Cook Sundar Pichai Taiwan Tech Robotics Ransomware SpaceX Samsung Quantum computing Series B
      Major Tech Companies
      • Apple News
      • Google News
      • Meta News
      • Microsoft News
      • Amazon News
      • Samsung News
      • Nvidia News
      • OpenAI News
      • Tesla News
      • AMD News
      • Anthropic News
      • Elbit News
      AI & Emerging Tech
      • AI Regulation News
      • AI Safety News
      • AI Adoption
      • Quantum Computing News
      • Robotics News
      Key People
      • Sam Altman News
      • Jensen Huang News
      • Elon Musk News
      • Mark Zuckerberg News
      • Sundar Pichai News
      • Tim Cook News
      • Satya Nadella News
      • Mustafa Suleyman News
      Global Tech & Policy
      • Israel Tech News
      • India Tech News
      • Taiwan Tech News
      • UAE Tech News
      Startups & Emerging Tech
      • Series A News
      • Series B News
      • Startup News
      Tallwire
      Facebook X (Twitter) LinkedIn Threads Instagram RSS
      • Tech
      • Entertainment
      • Business
      • Government
      • Academia
      • Transportation
      • Legal
      • Press Kit
      © 2026 Tallwire. Optimized by ARMOUR Digital Marketing Agency.

      Type above and press Enter to search. Press Esc to cancel.