DiscoverAI Podcast Summaries from Transcripted.ai (VIDEO)AI Guardrails Don't Work — Lenny's Podcast with Sander Schulhoff
AI Guardrails Don't Work — Lenny's Podcast with Sander Schulhoff

AI Guardrails Don't Work — Lenny's Podcast with Sander Schulhoff

Update: 2025-12-21
Share

Description

Hook: Powerful AI plus real-world permissions creates an urgent security problem few teams are prepared for.
This condensed version (original 2 hours → new 4 minutes) of Lenny Rachitsky’s interview with Sander Schulhoff distills why guardrails fail, how jailbreaking and prompt injection differ, and why agents that can act (email, browser, payments) dramatically amplify risk. You’ll learn practical defenses—tight permissioning, least-privilege architectures, logging and observability, and hiring cross-disciplinary security + AI talent—plus why static test metrics and marketing claims about adversarial robustness are misleading. Sander emphasizes concrete steps for deployment, adaptive evaluation for robustness, and avoiding risky publishable jailbreaks that help attackers. Topics covered include adversarial robustness, prompt injection, AI security, agent permissioning, and misinformation risk. Listen now to get the key ideas in minutes.
Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

AI Guardrails Don't Work — Lenny's Podcast with Sander Schulhoff

AI Guardrails Don't Work — Lenny's Podcast with Sander Schulhoff

Transcripted.ai