DiscoverVanishing GradientsEpisode 43: Tales from 400+ LLM Deployments: Building Reliable AI Agents in Production
Episode 43: Tales from 400+ LLM Deployments: Building Reliable AI Agents in Production

Episode 43: Tales from 400+ LLM Deployments: Building Reliable AI Agents in Production

Update: 2025-01-16
Share

Description

Hugo speaks with Alex Strick van Linschoten, Machine Learning Engineer at ZenML and creator of a comprehensive LLMOps database documenting over 400 deployments. Alex's extensive research into real-world LLM implementations gives him unique insight into what actually works—and what doesn't—when deploying AI agents in production.



In this episode, we dive into:




  • The current state of AI agents in production, from successes to common failure modes

  • Practical lessons learned from analyzing hundreds of real-world LLM deployments

  • How companies like Anthropic, Klarna, and Dropbox are using patterns like ReAct, RAG, and microservices to build reliable systems

  • The evolution of LLM capabilities, from expanding context windows to multimodal applications

  • Why most companies still prefer structured workflows over fully autonomous agents



We also explore real-world case studies of production hurdles, including cascading failures, API misfires, and hallucination challenges. Alex shares concrete strategies for integrating LLMs into your pipelines while maintaining reliability and control.



Whether you're scaling agents or building LLM-powered systems, this episode offers practical insights for navigating the complex landscape of LLMOps in 2025.



LINKS



Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Episode 43: Tales from 400+ LLM Deployments: Building Reliable AI Agents in Production

Episode 43: Tales from 400+ LLM Deployments: Building Reliable AI Agents in Production

Hugo Bowne-Anderson