Global Feed Post Login
Replying to Avatar Mark E. Jeftovic

Turns out, most LLMs can have their safety guardrails bypassed (read: hacked) by rewriting harmful prompts as poetry…

https://axisofeasy.com/aoe/the-telefon-problem-hacking-ai-with-poetry-instead-of-prompts/

Avatar
Ike 1mo ago

harmful prompts 😅

Reply to this note

Please Login to reply.

Discussion

No replies yet.