DiscoverLenny's Podcast: Product | Career | GrowthWhy most AI products fail: Lessons from 50+ AI deployments at OpenAI, Google, and Amazon
Why most AI products fail: Lessons from 50+ AI deployments at OpenAI, Google, and Amazon

Why most AI products fail: Lessons from 50+ AI deployments at OpenAI, Google, and Amazon

Update: 2026-01-11
Share

Digest

This podcast explores the nuances of building AI products, distinguishing them from traditional software by highlighting non-determinism and the agency-control tradeoff. It advocates for a "problem first" approach, starting with minimal AI autonomy and gradually increasing it through continuous calibration and development (CCCD). Key challenges like reliability and the importance of user experience are discussed, alongside success factors such as strong leadership and a culture of empowerment. The role of evaluations (evals) is clarified as part of a feedback loop, not a standalone solution. The future of AI is predicted to involve more proactive, multi-modal experiences, with essential skills for builders including design, judgment, and taste. The concept of "pain is the new moat" emphasizes that enduring the iterative development process creates a competitive advantage. Ultimately, customer obsession and deep workflow understanding are presented as the core differentiators for successful AI product companies.

Outlines

00:00:00
Introduction to AI Product Development: Core Differences and Challenges

The podcast introduces the fundamental differences between building AI products and traditional software, highlighting non-determinism and the agency-control tradeoff. It also touches upon the current state of AI product development in 2025, noting reduced skepticism but persistent execution challenges due to the field's immaturity.

00:11:45
The "Problem First" Approach and Gradual Agency in AI

It's recommended to start AI product development with a "problem first" mindset, beginning with minimal autonomy and high human control. This approach is illustrated with examples like customer support, where AI progresses from suggestions to autonomous resolution, emphasizing gradual agency and control.

00:16:02
Behavior Calibration, Progression, and Reliability in AI Systems

AI systems require behavior calibration to minimize negative customer experiences while gradually increasing autonomy. Examples of AI product progression, such as coding assistants and marketing tools, are provided. Reliability remains a major challenge, with many current AI products focusing on productivity due to lower autonomy.

00:22:38
Success Factors and the Crucial Role of Evals

Successful AI product development hinges on great leaders, a culture of empowerment, and strong technical prowess focused on understanding workflows. Evals are presented as crucial but misunderstood, forming part of a feedback loop that complements production monitoring.

00:46:00
The CCCD Framework and Advancing AI Development

The Continuous Calibration, Continuous Development (CCCD) framework is presented as an iterative approach to AI product development, emphasizing continuous calibration and development to build trust and improve systems. Advancing in AI development involves minimizing surprises and observing consistent user behavior.

01:01:24
Overhyped vs. Underhyped AI Concepts and Business Problem Focus

Multi-agent systems are considered misunderstood, while coding agents are seen as underrated. The importance of obsessing over the business problem and user pain points is emphasized over rapid building or solely focusing on evals.

01:05:22
The Future of AI: Proactive, Multi-Modal Experiences, and Essential Skills

The next year of AI will likely see proactive agents understanding workflows better and multi-modal experiences. Key skills for AI product builders include nailing design, judgment, and taste, alongside a willingness to experiment and take ownership.

01:12:48
"Pain is the New Moat" and Customer Obsession in AI

The "pain is the new moat" concept highlights that companies succeed by enduring the difficult process of iteration and learning in AI development. The core advice for building AI products is to be obsessed with customers and problems, deeply understanding workflows.

Keywords

Non-determinism in AI


Refers to the unpredictable nature of AI systems, where the same input can produce different outputs. This is a key challenge in AI product development, impacting user experience and system reliability.

Agency-Control Tradeoff


The balance between giving an AI system autonomy (agency) to make decisions and maintaining human control over those decisions. Increasing agency often means relinquishing some control, requiring trust in the AI's reliability.

Problem First Approach


A product development strategy prioritizing a deep understanding of the problem to be solved before focusing on the technical solution. In AI, this means starting with minimal autonomy and human oversight to validate the problem and solution.

Behavior Calibration


The process of fine-tuning an AI system's behavior through iterative development and monitoring. It involves understanding how the system interacts with users and data, and adjusting it to achieve desired outcomes and minimize errors.

Continuous Calibration, Continuous Development (CCCD)


An AI product development lifecycle framework inspired by CI/CD. It emphasizes iterative cycles of development, evaluation, and calibration to continuously improve AI systems and build user trust.

Evals (Evaluations)


A set of tests and metrics used to assess the performance and reliability of AI models and systems. Evals can range from predefined datasets to real-time production monitoring, crucial for identifying and mitigating issues.

Coding Agents


AI agents specifically designed to assist with software development tasks, such as writing code, debugging, and code review. They are considered underrated despite their potential to significantly boost developer productivity.

Pain is the New Moat


A concept suggesting that the difficult, iterative process of developing AI products, involving significant learning and problem-solving, creates a competitive advantage (moat) for companies that successfully navigate it.

Customer Obsession


A business strategy focused on understanding and meeting customer needs and expectations. In AI product development, it means deeply understanding user pain points and workflows to create valuable solutions.

Workflow Understanding


The process of deeply comprehending how users perform tasks and achieve goals within their specific operational context. This understanding is critical for designing effective AI products that integrate seamlessly and provide real value.

Q&A

  • What are the two fundamental differences between building AI products and traditional software?

    The two key differences are non-determinism, meaning both user input and AI output can be unpredictable, and the agency-control tradeoff, where granting more autonomy to AI systems requires relinquishing some human control.

  • What is the recommended approach for starting AI product development?

    The "problem first" approach is recommended, starting with minimal AI agency and high human control. This allows for gradual iteration, building confidence, and a better understanding of the problem being solved before increasing autonomy.

  • How does the Continuous Calibration, Continuous Development (CCCD) framework help in building AI products?

    CCCD provides an iterative process for AI development, focusing on continuous improvement through cycles of development, evaluation, and calibration. This helps manage risks, build user trust, and adapt to unexpected behaviors.

  • What is the significance of "pain is the new moat" in the context of AI development?

    This concept suggests that the challenging, iterative process of learning and problem-solving in AI development creates a unique competitive advantage. Companies that endure this "pain" build valuable knowledge and capabilities that are hard for competitors to replicate.

  • What are the key factors for success in companies building AI products?

    Success relies on great leaders who are hands-on and adaptable, a culture of empowerment and augmentation (not replacement), and strong technical expertise focused on deeply understanding workflows and choosing the right tools for the job.

  • What is the role of "evals" in AI development, and are they sufficient on their own?

    Evals are important for testing and monitoring AI systems, but they are not sufficient alone. They are part of a broader feedback loop that also includes production monitoring and customer feedback, with the specific approach depending on the application.

  • What is considered misunderstood in the current AI landscape?

    Multi-agent systems are seen as misunderstood. The idea that simply connecting multiple agents will lead to utopia is often unrealistic; successful multi-agent systems require careful design, control, and often a supervisory agent.

  • What skills should aspiring AI product builders focus on developing?

    Aspiring builders should focus on developing strong design, judgment, and taste, alongside a willingness to experiment and take ownership. As implementation becomes cheaper, these human-centric skills will be crucial differentiators.

Show Notes

Aishwarya Naresh Reganti and Kiriti Badam have helped build and launch more than 50 enterprise AI products across companies like OpenAI, Google, Amazon, and Databricks. Based on these experiences, they’ve developed a small set of best practices for building and scaling successful AI products. The goal of this conversation is to save you and your team a lot of pain and suffering.

We discuss:

1. Two key ways AI products differ from traditional software, and why that fundamentally changes how they should be built

2. Common patterns and anti-patterns in companies that build strong AI products versus those that struggle

3. A framework they developed from real-world experience to iteratively build AI products that create a flywheel of improvement

4. Why obsessing about customer trust and reliability is an underrated driver of successful AI products

5. Why evals aren’t a cure-all, and the most common misconceptions people have about them

6. The skills that matter most for builders in the AI era

Brought to you by:

Merge—The fastest way to ship 220+ integrations: https://merge.dev/lenny

Strella—The AI-powered customer research platform: https://strella.io/lenny

Brex—The banking solution for startups: https://www.brex.com/product/business-account?ref_code=bmk_dp_brand1H25_ln_new_fs

Transcript: https://www.lennysnewsletter.com/p/what-openai-and-google-engineers-learned

My biggest takeaways (for paid newsletter subscribers): https://www.lennysnewsletter.com/i/183007822/referenced

Get 15% off Aishwarya and Kiriti’s Maven course, Building Agentic AI Applications with a Problem-First Approach, using this link: https://bit.ly/3V5XJFp

Where to find Aishwarya Naresh Reganti:

• LinkedIn: https://www.linkedin.com/in/areganti

• GitHub: https://github.com/aishwaryanr/awesome-generative-ai-guide

• X: https://x.com/aish_reganti

Where to find Kiriti Badam:

• LinkedIn: https://www.linkedin.com/in/sai-kiriti-badam

• X: https://x.com/kiritibadam

Where to find Lenny:

• Newsletter: https://www.lennysnewsletter.com

• X: https://twitter.com/lennysan

• LinkedIn: https://www.linkedin.com/in/lennyrachitsky/

In this episode, we cover:

(00:00 ) Introduction to Aishwarya and Kiriti

(05:03 ) Challenges in AI product development

(07:36 ) Key differences between AI and traditional software

(13:19 ) Building AI products: start small and scale

(15:23 ) The importance of human control in AI systems

(22:38 ) Avoiding prompt injection and jailbreaking

(25:18 ) Patterns for successful AI product development

(33:20 ) The debate on evals and production monitoring

(41:27 ) Codex team’s approach to evals and customer feedback

(45:41 ) Continuous calibration, continuous development (CC/CD) framework

(58:07 ) Emerging patterns and calibration

(01:01:24 ) Overhyped and under-hyped AI concepts

(01:05:17 ) The future of AI

(01:08:41 ) Skills and best practices for building AI products

(01:14:04 ) Lightning round and final thoughts

Referenced:

• LevelUp Labs: https://levelup-labs.ai/

• Why your AI product needs a different development lifecycle: https://www.lennysnewsletter.com/p/why-your-ai-product-needs-a-different

Booking.com: https://www.booking.com

• Research paper on agents in production (by Matei Zaharia’s lab): https://arxiv.org/pdf/2512.04123

• Matei Zaharia’s research on Google Scholar: https://scholar.google.com/citations?user=I1EvjZsAAAAJ&hl=en

• The coming AI security crisis (and what to do about it) | Sander Schulhoff: https://www.lennysnewsletter.com/p/the-coming-ai-security-crisis

• Gajen Kandiah on LinkedIn: https://www.linkedin.com/in/gajenkandiah

• Rackspace: https://www.rackspace.com

• The AI-native startup: 5 products, 7-figure revenue, 100% AI-written code | Dan Shipper (co-founder/CEO of Every): https://www.lennysnewsletter.com/p/inside-every-dan-shipper

• Semantic Diffusion: https://martinfowler.com/bliki/SemanticDiffusion.html

• LMArena: https://lmarena.ai

• Artificial Analysis: https://artificialanalysis.ai/leaderboards/providers

• Why humans are AI’s biggest bottleneck (and what’s coming in 2026) | Alexander Embiricos (OpenAI Codex Product Lead): https://www.lennysnewsletter.com/p/why-humans-are-ais-biggest-bottleneck

• Airline held liable for its chatbot giving passenger bad advice—what this means for travellers: https://www.bbc.com/travel/article/20240222-air-canada-chatbot-misinformation-what-travellers-should-know

• Demis Hassabis on LinkedIn: https://www.linkedin.com/in/demishassabis

• We replaced our sales team with 20 AI agents—here’s what happened | Jason Lemkin (SaaStr): https://www.lennysnewsletter.com/p/we-replaced-our-sales-team-with-20-ai-agents

• Socrates’s quote: https://en.wikipedia.org/wiki/The_unexamined_life_is_not_worth_living

• Noah Smith’s newsletter: https://www.noahpinion.blog

Silicon Valley on HBO Max: https://www.hbomax.com/shows/silicon-valley/b4583939-e39f-4b5c-822d-5b6cc186172d

• Clair Obscur: Expedition 33: https://store.steampowered.com/app/1903340/Clair_Obscur_Expedition_33/

• Wisprflow: https://wisprflow.ai

• Raycast: https://www.raycast.com

• Steve Jobs’s quote: https://www.goodreads.com/quotes/463176-you-can-t-connect-the-dots-looking-forward-you-can-only

Recommended books:

 When Breath Becomes Air: https://www.amazon.com/When-Breath-Becomes-Paul-Kalanithi/dp/081298840X

The Three-Body Problem: https://www.amazon.com/Three-Body-Problem-Cixin-Liu/dp/0765382032

A Fire Upon the Deep: https://www.amazon.com/Fire-Upon-Deep-Zones-Thought/dp/0812515285

Production and marketing by https://penname.co/. For inquiries about sponsoring the podcast, email <a href="mailto:podcast@lennyrachitsky.com" targe

Comments 
In Channel
loading

Table of contents

00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Why most AI products fail: Lessons from 50+ AI deployments at OpenAI, Google, and Amazon

Why most AI products fail: Lessons from 50+ AI deployments at OpenAI, Google, and Amazon

Lenny Rachitsky