DiscoverSuper Prompt: Generative AIChatGPT Jailbreaks: The Grandma Exploit
ChatGPT Jailbreaks: The Grandma Exploit

ChatGPT Jailbreaks: The Grandma Exploit

Update: 2023-07-03
Share

Description

How do you extract prohibited information from ChatGPT? Grandma and DAN exploits trick language models into violating their own policies. Why these techniques work, what they reveal about LLM architecture, and how companies protect against prompt injection attacks. Solo episode on LLM security.

To stay in touch, sign up for our newsletter at https://www.superprompt.fm

Comments 
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

ChatGPT Jailbreaks: The Grandma Exploit

ChatGPT Jailbreaks: The Grandma Exploit

Tony Wan