DiscoverThe AI Native Dev - from Copilot today to AI Native Software Development tomorrowAI Evaluation and Testing: How to Know When Your Product Works (or Doesn’t)
AI Evaluation and Testing: How to Know When Your Product Works (or Doesn’t)

AI Evaluation and Testing: How to Know When Your Product Works (or Doesn’t)

Update: 2024-12-10
Share

Description

This episode of AI Native Dev, hosted by Simon Maple and Guy Podjarny, features a mashup of conversations with leading figures in the AI industry. Guests include Des Traynor, founder of Intercom, who discusses the paradigm shift generative AI brings to product development. Rishabh Mehrotra, Head of AI at SourceGraph, emphasizes the importance of evaluation processes over model training. Tamar Yehoshua, President of Products and Technology at Glean, shares her experiences in enterprise search and the challenges of using LLMs in data-sensitive environments. Finally, Simon Last, Co-Founder and CTO of Notion, talks about continuous improvement and the iterative processes at Notion. Each guest provides invaluable insights into the evolving landscape of AI-driven products.

Watch the episode on YouTube: https://youtu.be/gZ4sGROvOdQ

Comments 
loading
In Channel
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

AI Evaluation and Testing: How to Know When Your Product Works (or Doesn’t)

AI Evaluation and Testing: How to Know When Your Product Works (or Doesn’t)

Tessl