AI Evals & Discovery

AI Evals & Discovery

Update: 2025-09-23
Share

Description

What you’ll learn in this episode:



  • What “evals” actually mean in the AI/ML world

  • Why evals are more than just quality assurance

  • The difference between golden datasets, synthetic data, and real-world traces

  • How to identify error modes and turn them into evals

  • When to use code-based evals vs. LLM-as-judge evals

  • How discovery practices inform every step of AI product evaluation

  • Why evals require continuous maintenance (and what “criteria drift” means for your product)

  • The relationship between evals, guardrails, and ongoing human oversight


Resources & Links:



Mentioned in the episode:



Coming soon from Teresa:



  • Weekly Monday posts sharing lessons learned while building AI products

  • A new podcast interviewing cross-functional teams about real-world AI product development stories

Comments 
In Channel
AI Evals & Discovery

AI Evals & Discovery

2025-09-2326:45

Building AI Products

Building AI Products

2025-09-1621:16

Curation

Curation

2025-09-0218:20

Summer Break

Summer Break

2025-07-2202:10

Go to the Source

Go to the Source

2025-07-1521:28

AI Prototyping

AI Prototyping

2025-06-1720:36

What is Product Ops?

What is Product Ops?

2025-06-1022:18

Measurable Outcomes

Measurable Outcomes

2025-06-0313:07

Saying No

Saying No

2025-05-1322:09

Big Backlogs

Big Backlogs

2025-04-2919:18

Capacity Planning

Capacity Planning

2025-04-0815:55

loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

AI Evals & Discovery

AI Evals & Discovery

Episodes