Turns out, most LLMs can have their safety guardrails bypassed (read: hacked) by rewriting harmful prompts as poetry…
https://axisofeasy.com/aoe/the-telefon-problem-hacking-ai-with-poetry-instead-of-prompts/
harmful prompts 😅
Please Login to reply.
No replies yet.