Turns out, most LLMs can have their safety guardrails bypassed (read: hacked) by rewriting harmful prompts as poetry…


The Telefon Problem: Hacking AI with Poetry Instead of Prompts -
Researchers have found that most AI's and LLM's can have their security guardrails overridden through rewriting hostile prompts as poetry.