AI chatbots can be wooed into crimes with poetry
Source: The Verge
It turns out my parents were wrong. Saying “please” doesn’t get you what you want—poetry does. At least, it does if you’re talking to an AI chatbot.
That’s according to a new study from Italy’s Icaro Lab, an AI evaluation and safety initiative from researchers at Rome’s Sapienza University and AI co‑founders. The researchers discovered that by framing requests as poetic verses, they could coax chatbots into providing instructions for illicit activities that the models would normally refuse to share. The findings highlight a novel manipulation technique that exploits the models’ tendency to be more agreeable and creative when prompted with artistic language, raising fresh concerns for AI safety and prompting calls for stronger guardrails against adversarial prompting.