DiscoverHumans of ReliabilityYou Can’t Fix What You Don’t Measure: Observability in the Age of AI with Conor Bronsdon
You Can’t Fix What You Don’t Measure: Observability in the Age of AI with Conor Bronsdon

You Can’t Fix What You Don’t Measure: Observability in the Age of AI with Conor Bronsdon

Update: 2025-11-05
Share

Description

Only 50% of companies monitor their ML systems. Building observability for AI is not simple: it goes beyond 200 OK pings. In this episode, Sylvain Kalache sits down with Conor Brondsdon (Galileo) to unpack why observability, monitoring, and human feedback are the missing links to make large language model (LLM) reliable in production.

Conor dives into the shift from traditional test-driven development to evaluation-driven development, where metrics like context adherence, completeness, and action advancement replace binary pass-fail checks. He also shares how teams can blend human-in-the-loop feedback, automated guardrails, and small language models to keep AI accurate, compliant, and cost-efficient at scale.

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

You Can’t Fix What You Don’t Measure: Observability in the Age of AI with Conor Bronsdon

You Can’t Fix What You Don’t Measure: Observability in the Age of AI with Conor Bronsdon

Rootly