AI changed what excellence requires in product management.

AI changed what excellence requires in product management.

AI-powered products are fundamentally different from the software most product teams have spent their careers building. That difference is not about speed or automation. It is about uncertainty, control, and responsibility.

I know.  No Duh! 

Stay with me…

After listening to a recent episode of Lenny’s Podcast featuring Aishwarya Reganti and Kiriti Badam, leaders who have built and deployed dozens of AI products at companies like OpenAI, Google, Amazon, and Databricks, a few things became really clear about where product management is changing the most dramatically.    What follows is a product leader’s interpretation of what this shift means for how we design, roadmap, and lead AI-driven products.  

Why AI breaks the assumptions traditional software was built on

Traditional software is largely deterministic. A user clicks a button and the system responds the same way every time. AI products are not built on that assumption. 

AI systems are probabilistic on both sides of the interaction. Users express intent in natural language, often imprecisely. Models respond with outputs that vary based on phrasing, context, and underlying uncertainty. That means product teams are designing systems where neither inputs nor outputs can be fully predicted in advance.

Think about an AI chatbot used as a support tool. One person types, “Why is my bill higher this month?” Another types, “You charged me wrong.” Another types out a dissertation of their problem, with all caps and screenshots.  All are pointing at the same underlying problem, but they arrive with different language, emotion, and expectations. The model must interpret intent, retrieve relevant information, and respond appropriately, even though the starting signals are ambiguous.  And each time AI interpretation might yield different results that might get interpreted by the user differently.

For product teams, this changes the job. We are no longer designing a single, predictable flow. We are shaping behavior across a range of possible interactions; inputs and outputs.

Why AI autonomy must be earned, not shipped

Another major shift in AI product management is the tradeoff between agency and control. Agency refers to how much autonomy the system has to act on behalf of the user. Control refers to how much oversight or intervention remains with humans.  Most AI failures we see in the wild are not model failures. They are product decisions where too much agency was granted before the system earned trust.

 Using the same AI Chatbot support example, early versions of the experience might suggest responses to a human agent. The human decides what to send. Later versions might draft replies directly to customers. Eventually, the system might issue refunds or resolve cases autonomously.  Each step increases agency and reduces control. Product leaders must decide when that progression is justified, based on evidence, not ambition.  Companies like OpenAI and other AI-native organizations explicitly stage autonomy over time. They start with high human oversight, observe behavior, and only expand autonomy once reliability is demonstrated in real usage.

Why AI demands a new definition of product success

AI products cannot be fully validated upfront. Their real behavior only emerges once real users interact with them. That makes calibration a core product responsibility, not a technical afterthought.  

Calibration is about measuring what actually matters in pursuit of the right outcomes, not just collecting signals because they are easy to track.  For AI-driven products, the outcomes that matter most are trust, accuracy, and confidence.  In practice, strong product teams deliberately track signals that indicate whether trust and accuracy are holding:

  • When users rephrase the same request multiple times, signaling misunderstanding or low confidence
  • When humans override or heavily edit AI output, signaling gaps in judgment
  • When the system responds confidently but incorrectly, signaling risk to trust
  • When customers escalate, abandon, or bypass the AI altogether

These signals are not edge cases. They are leading indicators of outcome failure. Strong teams treat this calibration as outcome-driven discovery in production. They regularly review real interactions, not just aggregate performance metrics, to understand where the system is helping users achieve their goals and where it is undermining them.

This calibration tells you whether the product is getting closer to the outcomes you care about, and where it is falling short.

The learning loop as a product responsibility

Reganti and Badam also describe the importance of implementing the “flywheel.”  AKA Learning loop and intentionally designing them into the product AND process from day 1. On the surface this feels like…”Well, yeaaaah…” But the urgency and importance of a learning loop in AI is exponentially more critical to the success of the product than it is for traditional software.  A majority of the catastrophic bugs are found early and then teams usually move on to new features or other initiatives.  (I’m not saying that’s the right thing, just that it’s not catastrophic to the business.) With AI, user trust erodes quickly, and early signals can snowball.  Best-in-class teams do this through very specific tactics:

  • Override reviews: Product teams regularly review where humans corrected or overruled the AI and ask why.
  • Pattern tagging: Common failure modes are tagged and tracked over time instead of being handled ad hoc.
  • Guardrail evolution: Constraints are tightened or relaxed based on where trust holds or breaks. For example, narrowing the scope of actions the AI can take until confidence improves.
  • Roadmap feedback: Insights from real usage directly influence prioritization, not just model tuning.

Without these loops, teams see the same failures repeatedly. With them, mistakes become rarer, autonomy grows safely, and trust compounds.  These learning loops drive your roadmap and agency decision for the AI.


How can we think about product management in this new reality?

1. Designing products that assume uncertainty instead of hiding it

Strong AI product teams do not try to eliminate uncertainty. They design for it explicitly.  In practice, this means product teams decide where ambiguity is acceptable and where it is not, and then design the experience and tracking accordingly. For example, teams building AI-driven experiences often:

      • Define “safe-to-fail” vs. “must-be-correct” moments in the user journey. An AI-generated draft, suggestion, or explanation can tolerate ambiguity. A refund decision, medical recommendation, or data change cannot.
      • Design clarification into the experience rather than letting the model guess. When inputs are vague, the product is intentionally built to ask follow-up questions or narrow scope instead of confidently producing a low-quality answer.
      • Constrain output when confidence is low, such as summarizing known facts, surfacing options, or deferring action, rather than fabricating certainty.

This shifts product thinking from “How accurate is the model?” to “How should the product behave when the model is unsure?” That is a product decision, not a modeling one.

2. Treating AI autonomy as something the product earns over time

High-performing teams do not ship fully autonomous AI experiences on day one. They treat autonomy as a roadmap decision, not a launch feature. In practice, this shows up as intentional staging of AI responsibility, often moving through phases like:

      • Recommend: the AI suggests an action, response, or classification, but a human decides.
      • Draft: the AI produces a fuller output that a human edits, approves, or rejects.
      • Act: the AI executes actions with limited or no human involvement.

What matters is not the labels, but the discipline:

      • Autonomy increases only after teams see consistent, trustworthy behavior in real usage.
      • Early versions are designed to expose mistakes safely, when humans can intervene.
      • Product teams explicitly decide which actions require human confirmation and which do not.

This turns autonomy into something that is earned through demonstrated reliability, rather than assumed because the model is powerful.

3. Running outcome-driven learning loops, not vanity metrics

The best AI product teams do not measure success by model accuracy alone. They measure outcomes tied to trust and usefulness, then use those signals to guide product decisions.  In practice, this means teams:

      • Track behavioral signals, not just ratings. Rewrites, retries, overrides, ignored suggestions, and manual corrections all signal where the product is failing or succeeding.
      • Review failure patterns regularly, not as one-off bugs. Repeated breakdowns in the same step of a workflow become roadmap inputs, not support tickets.
      • Use outcomes to adjust the product, not just the prompt or model. Sometimes the right fix is tighter constraints, better context, reduced autonomy, or a redesigned interaction.

This is what Reganti and Badam describe as a “flywheel” in the podcast:  Usage creates insight, insight drives product changes, and those changes improve future behavior. 

The key product lesson is this: learning does not happen automatically just because users are interacting with AI. It only happens when teams deliberately decide what outcomes matter and build feedback loops around them.


The Takeaway 

AI does not change the fundamentals of product management. It raises the bar on them.  The teams that succeed are not the ones chasing smarter models. They are the ones that:

  • Design for uncertainty instead of pretending it doesn’t exist
  • Treat autonomy as a product decision that evolves over time
  • Build tight learning loops around real outcomes, not surface-level metrics

In an AI-driven world, the PM’s job is no longer to define exactly what the product will do.
It is to define how the product should behave when things do not go as planned.


 Want help putting this into practice?

At Product Rebels, we work with product leaders and teams navigating exactly this shift: using AI to accelerate delivery without losing product discipline or customer trust.

Is your team experiencing challenges in implementing AI into your product operating model or struggling in establishing the practices that enable the best outcomes from AI product building?  

Schedule 30 minutes with us to learn a little bit about you and explore how we can help. 

 About Product Rebels

Product Rebels helps product leaders bring their teams from good to great. We work with established product organizations that already know the basics of product management but want to operate at a higher level. Our focus is not Product Management 101. It’s helping teams build strong customer foundations and outcome-oriented ways of working that consistently translate into better results; for customers and for the business.

We partner with leaders and teams to change how product work actually happens day to day: how customer insight is gathered and shared, how problems are framed, how tradeoffs are made, and how learning compounds over time. AI is infused throughout these practices as an accelerator, helping teams synthesize learning faster, explore more options, and move with greater confidence without sacrificing judgment or customer connection.

The result is product teams that don’t just ship more, they build the right things, make better decisions under pressure, and deliver meaningful, sustained impact.


 Sources and further reading

Lenny’s Podcast: What OpenAI and Google engineers learned building AI products
https://www.lennysnewsletter.com/p/what-openai-and-google-engineers-learned

Ready To Transform Your Product Team?