Framework

Not all features should go live immediately. Many should be A/B tested first.

Prioritization for experiments differs from feature prioritization:

DimensionFeature PrioritizationExperiment Prioritization
MetricRevenue/engagement impactLearning value
QuestionWhat has biggest impact?What's most uncertain?
GoalShip the best featuresLearn the most

High-uncertainty, medium-impact feature should be tested. High-certainty feature should be shipped.

Actionable Steps

1. Categorize Each Idea

  • High confidence → Ship directly
  • Medium confidence → A/B test first
  • Low confidence → Prototype/learn first

2. Prioritize Tests by Expected Learning Value

Test the idea that, if wrong, would most change your strategy.

3. Run Test Quickly, Get Signal Fast

The faster you learn, the more experiments you can run.

Key Takeaways

  • Testing priorities differ from shipping priorities. Test uncertain ideas, ship confident ideas.
  • Maximize learning velocity. Faster tests = more learning in same time frame.
  • Test the idea that, if wrong, would most change your strategy. Learning value > potential upside.

The Experiment Prioritization Problem

You have 30 experiment ideas. Your team can run 5 this quarter. Which 5?

Common mistakes:

  1. "Test the feature with highest upside" (ignores certainty)

    • If you're 80% confident it works, testing is low-value
    • You should ship directly
  2. "Test the feature with highest downside" (too conservative)

    • Never test because there's always downside risk
    • Paralysis by analysis
  3. "Test random ideas" (no prioritization framework)

    • 15 small experiments, learn nothing
    • Should have run 5 focused experiments

Best approach: Prioritize by learning value per dollar/time spent.


The ICE-V Framework for Experiment Prioritization

ICE = Impact × Confidence × Ease V = Value of information (what do we learn if wrong?)

Step 1: Score Each Experiment on ICE

Impact (1-5 scale):

  • How much will this improve the metric if it works?
  • 1 = +0.5% (minimal)
  • 5 = +20%+ (transformational)

Example: "Add social sharing button"

  • Potential impact on engagement: +3% (small but positive)
  • Impact score: 3

Confidence (1-5 scale):

  • How confident are you this will work?
  • 1 = Total speculation
  • 5 = We're 90%+ confident it works

Example: "Add social sharing button"

  • Similar products have it, users asked for it
  • Confidence score: 4

Ease (1-5 scale):

  • How easy/fast is this to test?
  • 1 = Months of work
  • 5 = Hours of work

Example: "Add social sharing button"

  • Design: 4 hours
  • Engineering: 8 hours
  • Testing infrastructure: Already set up
  • Ease score: 5

ICE Score = 3 × 4 × 5 = 60

Step 2: Score Value of Information (V)

Value of information = magnitude of strategy change if you learn you were wrong

Question: "If this test fails, how much does it change your product strategy?"

Examples:

Test: "Add social sharing button"

  • If it fails: "OK, sharing isn't important to users. No strategy change."
  • V score: 1 (low learning value)
  • Recommendation: Don't test. Your confidence is high (4/5). Ship it.

Test: "Change pricing from per-user to per-company"

  • If it fails: "We misunderstood buyer behavior. Strategy is wrong."
  • If it works: "This unlocks enterprise market."
  • V score: 5 (high learning value)
  • Recommendation: Test this, even if effort is high.

Test: "Build mobile app vs. responsive web"

  • If mobile wins: "We build native, shift resources"
  • If web wins: "We go web-only, avoid native maintenance"
  • V score: 5 (high learning value)
  • Recommendation: Test this despite high effort.

Step 3: Prioritization Matrix

Plot each experiment on:

  • X-axis: ICE Score (left = low, right = high)
  • Y-axis: Value of Information (bottom = low, top = high)
QuadrantActionExample
High ICE, High VDO FIRST ✓Pricing model change, core flow optimization
High ICE, Low VSHIP DIRECTLY (skip test)Add social sharing, copy tweaks
Low ICE, High VDO AFTER (risky but learn a lot)New product category, novel interaction
Low ICE, Low VDEPRIORITIZEMinor UI tweaks, niche feature requests

Real-World Case Study: Experiment Prioritization

Company: Mid-Market SaaS (50K users)

Q2 Experiment Backlog: 30 ideas

ExperimentImpactConfEaseICEVQuadrantPriority
Pricing model (per-user → per-org)521105High V, Low ICEDO 2ND
Onboarding checklist345602Quad IISHIP
Dark mode154201Quad IVSKIP
New export formats233181Quad IVSKIP
AI-powered suggestions41285High V, Low ICEDO 1ST*
Help sidebar position245401Quad IISHIP
Slack integration334363Quad IIIDO 3RD
New dashboard layout234242Quad IVSKIP

*Note: AI-powered suggestions has low ICE because it's unproven tech. But V=5 (if it works, changes strategy). Moved to DO 1ST.

Prioritization Decision:

Run these 5 experiments this quarter (in order):

  1. AI-powered suggestions (High V learning, risky)
  2. Pricing model change (High V learning, need more confidence first)
  3. Slack integration (Moderate learning, medium effort)
  4. Onboarding checklist (Ship directly, don't test—high confidence)
  5. Help sidebar (Ship directly, don't test—high confidence)

Results (Q2):

ExperimentOutcomeLearning
AI suggestionsWorks! +15% engagementStrategy insight: AI is value driver. Invest here.
Pricing modelFails. Only 15% adoptionLearning: Users want per-user flexibility. New pricing model needed.
Slack integrationWorks. +20% retentionDevelopers love workflows. Expand integration ecosystem.
Onboarding+8% onboarding completion✓ Shipped, worked as expected
Help sidebar+3% support ticket reduction✓ Shipped, minor win

Strategic decisions based on learning:

  • Allocate 30% Q3 engineering to AI features (was 5%)
  • Redesign pricing to keep per-user flexibility (was going to change)
  • Expand Slack integration (prioritize over other integrations)

Without experiment prioritization: Would have tested all 30 ideas, learned less, and taken longer to act on insights.


Anti-Patterns in Experiment Prioritization

Anti-Pattern 1: "Test everything, no matter how confident"

The problem: You test "Add social sharing button" (90% confident it works). Result: 4 weeks testing confirms the obvious.

The fix: If confidence > 75%, ship directly. Use testing for uncertain ideas only.

Anti-Pattern 2: "Run too many experiments in parallel"

The problem: Running 15 small experiments simultaneously. Result: No statistical power. No clear winners. Wasted effort.

The fix: Run 3-5 focused experiments. Get clear signal. Ship winning ones. Then iterate.

Anti-Pattern 3: "Optimize for speed, not learning"

The problem: "We need results fast, so let's test surface-level changes (button color, copy)" Result: Fast answers, no strategic insight.

The fix: Balance speed with learning value. Spend 6 weeks testing pricing model (high V) vs. 1 week testing button color (low V).


The Economics of Experiment Prioritization

Scenario: $100K testing budget for 20 people × 5 weeks

Bad prioritization:

  • 20 small experiments (low learning value each)
  • Cost: $100K total
  • Learning: Fractional
  • ROI: Low

Good prioritization:

  • 5 focused experiments (high learning value each)
  • Cost: $100K total
  • Learning: High (strategic insights)
  • ROI: Identifies 2 features worth $1M+ in revenue potential

Math: If 1 experiment uncovers a $2M+ feature opportunity, ROI is 20:1.


PMSynapse Connection

The gap: Most PMs run experiments but don't tie them to strategy or prioritization. PMSynapse connects experiments to outcomes. For each proposed experiment: "What hypothesis are we testing? What strategy changes if we learn it's wrong? Does this have high learning value?" By forcing this thinking, PMs shift from "run lots of tests" to "run the right tests."


Key Takeaways (Expanded)

  • Don't test high-confidence ideas. If you're 80%+ confident, ship directly. Use testing for uncertain bets.

  • Prioritize by learning value, not upside potential. A $100K opportunity that you're 90% confident about beats a $1M opportunity that you're 20% confident about.

  • Run 3-5 focused experiments, not 20 small ones. You'll learn more in the same time frame.

  • Balance speed with learning. Don't optimize purely for fast results at the expense of strategic insight.

  • Track experiment outcomes and use them to inform strategy. If pricing model test fails, that's a strategic insight worth millions.

Experiment Prioritization: Which A/B Tests to Run First

Article Type

SPOKE Article — Links back to pillar: /product-prioritization-frameworks-guide

Target Word Count

2,500–3,500 words

Writing Guidance

Cover: the ICE framework for experiments, learning-based prioritization, and how to estimate the value of information from each experiment. Soft-pitch: PMSynapse supports hypothesis-driven product development and helps track experiment outcomes.

Required Structure

1. The Hook (Empathy & Pain)

Open with an extremely relatable, specific scenario from PM life that connects to this topic. Use one of the PRD personas (Priya the Junior PM, Marcus the Mid-Level PM, Anika the VP of Product, or Raj the Freelance PM) where appropriate.

2. The Trap (Why Standard Advice Fails)

Explain why generic advice or common frameworks don't address the real complexity of this problem. Be specific about what breaks down in practice.

3. The Mental Model Shift

Introduce a new framework, perspective, or reframe that changes how the reader thinks about this topic. This should be genuinely insightful, not recycled advice.

4. Actionable Steps (3-5)

Provide concrete actions the reader can take tomorrow morning. Each step should be specific enough to execute without further research.

5. The Prodinja Angle (Soft-Pitch)

Conclude with how PMSynapse's autonomous PM Shadow capability connects to this topic. Keep it natural — no hard sell.

6. Key Takeaways

3-5 bullet points summarizing the article's core insights.

Internal Linking Requirements

  • Link to parent pillar: /blog/product-prioritization-frameworks-guide
  • Link to 3-5 related spoke articles within the same pillar cluster
  • Link to at least 1 article from a different pillar cluster for cross-pollination

SEO Checklist

  • Primary keyword appears in H1, first paragraph, and at least 2 H2s
  • Meta title under 60 characters
  • Meta description under 155 characters and includes primary keyword
  • At least 3 external citations/references
  • All images have descriptive alt text
  • Table or framework visual included