Close Menu

    Subscribe to Updates

    Get the latest tech news from Tallwire.

      What's Hot

      Epic Games Adds Inflation To In-Game Currency

      April 16, 2026

      Starlink Outage Reveals Military Dependence on SpaceX

      April 16, 2026

      The Gaming World as of April 2026

      April 15, 2026
      Facebook X (Twitter) Instagram
      • Tech
      • AI
      • Get In Touch
      Facebook X (Twitter) LinkedIn
      TallwireTallwire
      • Tech

        Starlink Outage Reveals Military Dependence on SpaceX

        April 16, 2026

        The Gaming World as of April 2026

        April 15, 2026

        Amazon Buys Satellite Company Globalstar- It’s About Control of Space-Based Connectivity

        April 15, 2026

        NASA Astronauts Use iPhones to Capture Historic Artemis II Mission Images

        April 8, 2026

        OpenAI Expands Influence With Strategic TBPN Media Acquisition

        April 8, 2026
      • AI

        Anthropic Code Leak Raises Questions About AI Security and Industry Oversight

        April 8, 2026

        The Rise Of Agentic AI Signals A Shift From Tools To Autonomous Digital Actors

        April 8, 2026

        AI Chatbots Draw Scrutiny As Teens Engage In Intimate Roleplay And Emotional Dependency

        April 8, 2026

        Ai-Powered Startup Signals Rise Of One-Person Billion-Dollar Companies

        April 8, 2026

        OpenAI Secures Historic $122 Billion Funding Round at $852 Billion Valuation

        April 7, 2026
      • Security

        Anthropic Code Leak Raises Questions About AI Security and Industry Oversight

        April 8, 2026

        DeFi Platform Drift Halts Operations After Multi-Million Dollar Crypto Hack

        April 7, 2026

        Fake WhatsApp App Exposes Users To Government Spyware Operation

        April 7, 2026

        ICE Deploys Controversial Spyware Tool In Drug Trafficking Investigations

        April 7, 2026

        Telehealth Firm Discloses Breach Amid Rising Digital Health Vulnerabilities

        April 6, 2026
      • Health

        European Crackdown Targets Social Media’s Impact on Children

        April 8, 2026

        AI Chatbots Draw Scrutiny As Teens Engage In Intimate Roleplay And Emotional Dependency

        April 8, 2026

        Australia Moves To Curb Social Media Addiction Among Youth With Expanded Under-16 Ban

        April 5, 2026

        Australia’s eSafety Regulator Warns Big Tech As Teens Circumvent Social Media Restrictions

        April 5, 2026

        Meta Finally Held Accountable For Harming Teens, But Real Reform Remains Uncertain

        April 2, 2026
      • Science

        Starlink Outage Reveals Military Dependence on SpaceX

        April 16, 2026

        Amazon Buys Satellite Company Globalstar- It’s About Control of Space-Based Connectivity

        April 15, 2026

        Artemis II Splashdown Signals A Step Closer to Mass Space Travel

        April 12, 2026

        Peter Thiel’s Bold Ag-Tech Gamble Signals High-Tech Disruption of Traditional Ranching

        April 6, 2026

        White House Tech Advisor David Sacks Steps Down To Lead Presidential Science Advisory

        March 31, 2026
      • Tech

        Starlink Outage Reveals Military Dependence on SpaceX

        April 16, 2026

        Peter Thiel’s Bold Ag-Tech Gamble Signals High-Tech Disruption of Traditional Ranching

        April 6, 2026

        Zuckerberg Quietly Offers Musk Support As Tech Titans Align Around Government Power

        April 4, 2026

        White House Tech Advisor David Sacks Steps Down To Lead Presidential Science Advisory

        March 31, 2026

        Another Billionaire Signals Exit As California’s Taxes Drives Out High-Profile Entrepreneurs

        March 28, 2026
      TallwireTallwire
      Home»Tech»AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
      Tech

      AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot

      Updated:March 21, 20264 Mins Read
      Facebook Twitter Pinterest LinkedIn Tumblr Email
      AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
      AI Researchers Keep “Dangerous” Poetry-Based Prompts Under Wraps, Warn They Could Break Any Chatbot
      Share
      Facebook Twitter LinkedIn Pinterest Email

      Researchers at Icaro Lab (with the safety group DexAI and Sapienza University in Rome) have discovered that rewriting harmful prompts as poetry—what they call “adversarial poetry”—allows them to bypass safety filters in major AI chatbots from OpenAI, Google, Meta, Anthropic, and xAI. In testing 25 leading models, the team found that hand-crafted poetic prompts coaxed the AIs into giving forbidden or harmful content with a success rate around 63 percent on average; one model, Google’s Gemini 2.5, responded every single time. Even when the prompts were automatically generated poetry (not human-written), the success rate remained high (about 43 percent), far outperforming the same prompts in plain prose. Because of how readily these poetic “jailbreaks” worked—even on cutting-edge AI—the researchers decided the prompts were too dangerous for public release.

      Sources: Futurism, Wired

      Key Takeaways

      – Poetry-formatted prompts (rather than plain prose) dramatically increase success rates for jailbreaking AI safety filters—researchers call this method “adversarial poetry.”

      – The vulnerability is widespread—major models from top AI developers were fooled, though smaller or simpler models (e.g., OpenAI’s GPT-5 nano) proved more resilient.

      – Given just how effective and easy the technique is (and how dangerous its results can be), researchers have chosen not to release concrete examples of the prompts.

      In-Depth

      The recent findings from Icaro Lab and its collaborators have stirred the AI research world—and for good reason. What seemed, at first, like a clever linguistic trick now poses a serious and immediate threat to the safety mechanisms built into large language models. The core insight: reformatting harmful or restricted requests into poetic verse can slip past the guardrails that are supposed to prevent AIs from dispensing advice on dangerous or malicious topics.

      In their experiment, researchers collected a library of known harmful prompts—requests like how to build a weapon, how to facilitate crime, or other illicit instructions. They then rewrote those prompts as poems, either by hand or by using an AI to transform the prose into verse. That turned out to be a game changer. Across 25 top-tier models from major AI companies, the hand-crafted poems coaxed unsafe output roughly 63 percent of the time. In some cases, like with Google’s Gemini 2.5, the AI responded with forbidden content every single time it received a poem prompt. Even the automatic, AI-generated poetic prompts had a success rate around 43 percent—still wildly higher than when the same content was phrased in plain English.

      This should set off alarm bells. If anyone with basic poetry knowledge can craft prompts to elicit illicit content, it only takes a small community of bad actors to exploit extensively. The rationale, according to coauthor Matteo Prandi, is that the poetic structure disrupts how large language models internally predict and generate text—so much so that the AIs’ content filters simply fail to catch the harmful intent. The researchers considered releasing sample prompts to the public to help other experts study them, but decided against it; the tools are “too dangerous” for broad dissemination.

      Beyond the immediate risks of content generation (e.g., instructions on building weapons or facilitating cybercrime), this discovery underscores a deeper, structural challenge: safety filters and alignment protocols often rely on pattern-matching, keyword detection, and semantic analysis. But poetic verse defies these patterns. The algorithmic heuristics trained to catch wrongdoing don’t foresee the irregularities introduced by rhyme, archaic phrasing, or metaphorical wording. What this means is that many of the AI safety strategies deployed today could be fundamentally insufficient, especially once this method of “prompt engineering” becomes widespread.

      The implications ripple beyond just chatbots. As AI systems increasingly integrate into tools for coding, content creation, and automation, an attacker could embed poetic prompts within documents, commit messages, or even website content—anywhere an LLM might read from. Once inside, the malicious instructions might go unnoticed. This risk aligns with broader concerns around “prompt injection,” where AI systems are manipulated via user or external input to perform unintended actions. But “adversarial poetry” elevates that threat to a whole new level—it’s not just injection anymore, it’s a form of stealth.

      What must happen next is clear: AI developers need to reevaluate safety strategies. Instead of relying solely on keyword or format detection, they should explore approaches that understand intent contextually, detect unusual syntax deviations, or degrade model capability when instructions veer toward risky content. More so, publishers and academic researchers might need to enforce tighter controls on shared AI prompts—especially those that could be repurposed maliciously.

      Ultimately, this development illustrates that AI safety is not a solved problem, even if it once appeared that way. As adversaries become more creative and AI more ubiquitous, vulnerability will shift—and defenders need to stay a step ahead.

      AI Research AI Safety Anthropic Google Meta OpenAI
      Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
      Previous ArticleAI Researchers Embed LLM in a Robot—It Starts “Channeling Robin Williams and Still Can’t Pass the Butter
      Next Article AI Researchers Warn That Today’s Chatbots Risk Becoming “Digital Yes-Men”

      Related Posts

      Starlink Outage Reveals Military Dependence on SpaceX

      April 16, 2026

      The Gaming World as of April 2026

      April 15, 2026

      Amazon Buys Satellite Company Globalstar- It’s About Control of Space-Based Connectivity

      April 15, 2026

      Anthropic Code Leak Raises Questions About AI Security and Industry Oversight

      April 8, 2026
      Add A Comment
      Leave A Reply Cancel Reply

      Editors Picks

      Starlink Outage Reveals Military Dependence on SpaceX

      April 16, 2026

      The Gaming World as of April 2026

      April 15, 2026

      Amazon Buys Satellite Company Globalstar- It’s About Control of Space-Based Connectivity

      April 15, 2026

      NASA Astronauts Use iPhones to Capture Historic Artemis II Mission Images

      April 8, 2026
      Popular Topics
      Software SpaceX Tesla Cybertruck Sundar Pichai Samsung Series A Taiwan Tech Stocks Startup Satellite trending UAE Tech starlink Satya Nadella Viral Tesla Tim Cook spotlight Series B Space
      Major Tech Companies
      • Apple News
      • Google News
      • Meta News
      • Microsoft News
      • Amazon News
      • Samsung News
      • Nvidia News
      • OpenAI News
      • Tesla News
      • AMD News
      • Anthropic News
      • Elbit News
      AI & Emerging Tech
      • AI Regulation News
      • AI Safety News
      • AI Adoption
      • Quantum Computing News
      • Robotics News
      Key People
      • Sam Altman News
      • Jensen Huang News
      • Elon Musk News
      • Mark Zuckerberg News
      • Sundar Pichai News
      • Tim Cook News
      • Satya Nadella News
      • Mustafa Suleyman News
      Global Tech & Policy
      • Israel Tech News
      • India Tech News
      • Taiwan Tech News
      • UAE Tech News
      Startups & Emerging Tech
      • Series A News
      • Series B News
      • Startup News
      Tallwire
      Facebook X (Twitter) LinkedIn Threads Instagram RSS
      • Tech
      • Entertainment
      • Business
      • Government
      • Academia
      • Transportation
      • Legal
      • Press Kit
      © 2026 Tallwire. Optimized by ARMOUR Digital Marketing Agency.

      Type above and press Enter to search. Press Esc to cancel.