K
KnowMBAAdvisory
AI StrategyAdvanced7 min read

AI Pricing Experiments

AI pricing experiments test how to price AI products themselves and how to use AI to test pricing on non-AI products. The two are different sports. For pricing AI products: the canonical pattern is OpenAI's tier experimentation โ€” Free, Plus ($20), Pro ($200), Enterprise. Each tier tests willingness-to-pay against feature differentiation. For using AI to optimize pricing on other products: the pattern is Adobe-style ML personalization, where prices are tested per segment with bandits or A/B tests against a holdout. In both cases the trap is changing pricing without measurement infrastructure to detect cannibalization.

Also known asDynamic Pricing for AIAI Tier TestingML Price Optimization

The Trap

The trap is launching a new AI tier (e.g., 'Pro at $200/mo') without a measurement plan to detect cannibalization from the existing tier. OpenAI's $200 Pro plan was an experiment to extract surplus from power users โ€” but if 30% of $200 buyers were prior $20 Plus subscribers churning UP, that's cannibalization to net positive; if 30% of NEW $20 Plus signups were diverted DOWN from Pro because of confusion, that's cannibalization to net negative. Without segment-level cohort tracking, you don't know which.

What to Do

Before any AI pricing change, instrument three measurement layers: (1) Cohort revenue per visitor โ€” total revenue / visitors who saw the new pricing, (2) Tier mix shift โ€” % of new signups by tier vs. baseline, (3) Net new MRR โ€” accounting for cannibalization. Run the change as a 50/50 split for 4 weeks minimum. Real signal usually emerges at week 3 because of consideration cycles. Don't trust week 1 numbers โ€” they over-index on power users.

Formula

Net Pricing Lift = (Revenue per Visitor_new โˆ’ Revenue per Visitor_baseline) ร— Visitor Volume

In Practice

OpenAI launched ChatGPT Pro at $200/month in December 2024 as a tier experiment, sitting alongside Plus at $20. Sam Altman publicly stated they were losing money on Pro because power users consumed more inference than expected โ€” demonstrating the value of price experimentation: even at a 10x price multiple, willingness-to-pay among heavy users exceeded marginal cost in ways the company hadn't fully modeled. The experiment generated learning that retroactively reshaped capacity planning.

Pro Tips

  • 01

    Test price up before testing price down. It's almost always reversible โ€” you can drop a price next quarter โ€” but raising prices that you've previously dropped destroys trust.

  • 02

    Bandit algorithms (Thompson Sampling, contextual bandits) are appropriate for personalized pricing on commodity SKUs but inappropriate for SaaS subscriptions because the customer notices the variation. Use bandits for one-time purchases; use A/B tests for subscription tiers.

  • 03

    Always grandfather existing customers. A 'price up' that backfills to existing customers triggers churn that wipes out the new pricing's gains. Adobe learned this in 2012 with the Creative Cloud transition โ€” the customer outrage was a multi-year headwind.

Myth vs Reality

Myth

โ€œAI lets you optimize pricing in real time per customerโ€

Reality

It can โ€” but most contexts don't reward it. SaaS customers compare notes; B2B customers procurement-audit your pricing. Real-time personalized pricing is appropriate for hotels, airlines, and ride-share โ€” almost nothing else.

Myth

โ€œA pricing test has clear winners after a weekโ€

Reality

Pricing has long consideration cycles. A 1-week test over-weights impulsive buyers. Minimum 4 weeks for SaaS; 8+ weeks for enterprise. Anything shorter is noise.

Try it

Run the numbers.

Pressure-test the concept against your own knowledge โ€” answer the challenge or try the live scenario.

๐Ÿงช

Knowledge Check

You launch a $200 'Pro' tier alongside your existing $20 'Plus' tier. After 30 days: 200 Pro signups, but Plus signups dropped from 5,000/mo to 4,200/mo. What's the most likely net MRR impact?

Real-world cases

Companies that lived this.

Verified narratives with the numbers that prove (or break) the concept.

๐Ÿง 

OpenAI

December 2024

mixed

OpenAI launched ChatGPT Pro at $200/month, a 10x premium over the existing $20 Plus tier. Sam Altman publicly admitted on X that the company was losing money on Pro because power users consumed more inference than the price covered โ€” a candid signal that the tier was an experiment in extracting surplus from the heaviest users, not a settled pricing decision. The experiment yielded reusable knowledge: capacity planning changed, and subsequent product tiers (Sora, Operator) priced into the Pro bundle reflected the learning.

Pro Tier Price

$200/mo

Plus Tier Price

$20/mo (10x lower)

Disclosed Outcome

Losing money โ€” tier underpriced for heavy users

Pricing experiments are valuable even when they 'fail' financially short-term โ€” they reveal willingness-to-pay segments and consumption patterns that are otherwise invisible.

Source โ†—
๐ŸŽจ

Adobe (Sensei AI in Creative Cloud)

2017-present

success

Adobe uses ML (Sensei) to test promotional pricing and bundle composition across Creative Cloud customer segments. Personalized retention offers, free-month promos for at-risk users, and bundle upgrades are tested against held-out controls before being rolled out broadly. The discipline came from the painful 2012 Creative Cloud transition, when a forced subscription move triggered widespread customer backlash โ€” Adobe rebuilt its pricing testing infrastructure to never repeat that.

Pricing Test Cadence

Continuous

Methodology

Personalized offers + holdout

AI-driven pricing personalization works best for retention offers, not net-new acquisition pricing. Customers don't compare retention offers; they do compare list prices.

Source โ†—

Related concepts

Keep connecting.

The concepts that orbit this one โ€” each one sharpens the others.

Beyond the concept

Turn AI Pricing Experiments into a live operating decision.

Use this concept as the framing layer, then move into a diagnostic if it maps directly to a current bottleneck.

Typical response time: 24h ยท No retainer required

Turn AI Pricing Experiments into a live operating decision.

Use AI Pricing Experiments as the framing layer, then move into diagnostics or advisory if this maps directly to a current business bottleneck.