How Attackers Trick AI: Lessons from Gandalf’s Creator

How Attackers Trick AI: Lessons from Gandalf’s Creator

Update: 2025-03-18
Share

Description

🔒 How Secure is AI? Gandalf’s Creator Exposes the Risks 🔥

AI security is under attack, and hackers are finding new ways to manipulate AI systems. In this episode, Guy Podjarny sits down with Mateo Rojas-Carulla, co-founder of Lakera and creator of Gandalf, to break down the biggest threats facing AI today—from prompt injections and jailbreaks to data poisoning and agent manipulation.

What You’ll Learn:
- How attackers exploit AI vulnerabilities in real-world applications
- Why AI models struggle to separate instructions from external data
- How Gandalf’s 60M+ attack attempts revealed shocking insights
- What the Dynamic Security Utility Framework (DSEC) means for AI safety
- Why red teaming is critical for preventing AI disasters

Whether you’re a developer, security expert, or just curious about AI risks, this episode is packed with must-know insights on keeping AI safe in an evolving landscape.

💡 Can AI truly be secured? Or will attackers always find a way? Drop your thoughts in the comments! 👇

Watch the episode on YouTube: https://youtu.be/RKCvlJT_r4s

Join the AI Native Dev Community on Discord: https://tessl.co/4ghikjh

Ask us questions: podcast@tessl.io

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

How Attackers Trick AI: Lessons from Gandalf’s Creator

How Attackers Trick AI: Lessons from Gandalf’s Creator

Tessl