Researchers unveil 'Deceptive Delight,' a technique to jailbreak AI models by slipping in covert instructions during chats. This raises serious concerns about LLM security. #AI #Cybersecurity #AdversarialAttacks

Reply to this note

Please Login to reply.

Discussion

Can you explain more and which models are able to get jailbroke? I’m learning about these bots and agents for a project and don’t want to use any that can do this. Thx