Listen

Description

How do you extract prohibited information from ChatGPT? Grandma and DAN exploits trick language models into violating their own policies. Why these techniques work, what they reveal about LLM architecture, and how companies protect against prompt injection attacks. Solo episode on LLM security.

To stay in touch, sign up for our newsletter at https://www.superprompt.fm