ChatGPT Jailbreaks: The Grandma Exploit
Update: 2023-07-03
Description
How do you extract prohibited information from ChatGPT? Grandma and DAN exploits trick language models into violating their own policies. Why these techniques work, what they reveal about LLM architecture, and how companies protect against prompt injection attacks. Solo episode on LLM security.
To stay in touch, sign up for our newsletter at https://www.superprompt.fm
Comments
In Channel




