Product Management· 6 min read · April 10, 2026

How to Prioritize a Product Backlog Using RICE Scoring: 2026 Guide

A practical guide to prioritizing a product backlog using RICE scoring, covering how to estimate Reach, Impact, Confidence, and Effort accurately, avoid common scoring biases, and run a calibration session that makes RICE work across teams.

How to prioritize a product backlog using RICE scoring requires defining Reach, Impact, Confidence, and Effort in terms specific to your product before scoring anything — because teams that apply RICE without agreed definitions produce scores that reflect individual assumptions rather than shared evidence, making the prioritization worse than a gut-feel list.

RICE is a prioritization framework developed at Intercom. Score = (Reach × Impact × Confidence) / Effort. Used correctly, it surfaces which backlog items produce the most value per unit of engineering investment. Used incorrectly, it provides false precision on bad assumptions.

The RICE Formula

RICE Score = (Reach × Impact × Confidence) / Effort

Reach:      Number of users affected per time period
Impact:     Effect on the target metric per user (0.25 to 3)
Confidence: How confident are we in our estimates (10-100%)
Effort:     Person-months of work required

How to Estimate Each Dimension Accurately

Reach — Who and How Many?

Reach measures how many users will be affected by this item in a defined time period (typically per quarter).

Common mistakes:

  • Using total user base instead of affected users (a feature for enterprise admins affects 2% of users, not 100%)
  • Using aspirational numbers rather than current segment sizes

Better approach: Pull the actual count from your analytics. "The user segment who would encounter this feature has 4,200 monthly active users" is more reliable than "we have 50,000 total users."

Impact — How Much Does It Move the Needle?

Intercom's Impact scale: 3 = massive, 2 = high, 1 = medium, 0.5 = low, 0.25 = minimal.

Common mistake: Assigning 3 to everything. If everything is massive impact, you've lost the discrimination power of the scale.

Calibration rule: In any given scoring session, no more than 20% of items should receive a 3. If more than 20% score 3, force a forced-rank comparison: "Of these items that all scored 3, which truly has the most impact? Assign that one 3 and move the others to 2."

According to Shreyas Doshi on Lenny's Podcast, Impact is the dimension where PM intuition is most unreliable — PMs consistently overestimate the impact of features on user behavior, especially for features that solve problems the PM personally finds frustrating rather than problems identified through customer research.

Confidence — How Strong Is Your Evidence?

  • 100%: Hard data (A/B test result, customer research with 8+ interviews, analytics showing clear drop-off)
  • 80%: Soft data (qualitative interviews with 3–5 users, sales call observations)
  • 50%: Assumption ("We believe users want this because...")
  • 20%: Speculation ("We think this might help")

The confidence multiplier is the integrity check of the RICE framework. A high Reach × Impact score with 20% confidence should surface an experiment, not a full build.

Effort — True Cost, Not Ideal Case

Effort should be estimated in person-months including:

  • Engineering development
  • Design work
  • QA testing
  • PM spec and review time
  • Post-launch monitoring

Use your team's historical velocity for similar-sized items rather than bottom-up estimation, which consistently underestimates by 30–50%.

According to Gibson Biddle on Lenny's Podcast, effort estimation is where RICE breaks down most often — teams use best-case estimates rather than historical averages, producing RICE scores that favor complex features over simple ones because the effort input is systematically underestimated for large items.

Running a RICE Scoring Session

Step 1: Define Dimensions Together

Before scoring, align the team on:

  • What "Reach" means in your context (per month? per quarter?)
  • What the Impact scale points represent for your specific product
  • What constitutes 100%, 80%, 50%, 20% confidence in your evidence standards

This alignment prevents one PM scoring 3 because a feature is "strategically important" while another scores 2 because the user research is sparse.

Step 2: Score Independently, Then Compare

Each PM or stakeholder scores independently before the group discussion. This prevents anchoring bias — the first score stated in a group setting disproportionately influences all subsequent scores.

Step 3: Investigate Outliers

When scores diverge significantly (one person gives 3, another gives 0.5 for Impact), the divergence is information — it means different assumptions about the evidence or the user behavior. Investigate rather than average.

Step 4: Apply the Sanity Check

After scoring, check whether the ranked list matches your team's intuition about which items matter most. If the list seems wrong, don't override the scores — investigate which dimension is carrying bias.

According to Lenny Rachitsky's writing on product prioritization, RICE is most valuable not as a ranking machine but as a structured conversation tool — the process of discussing why scores diverge surfaces the assumptions and evidence gaps that, once resolved, align the team far more than any framework output.

FAQ

Q: What is RICE scoring in product management? A: A prioritization framework where RICE Score equals Reach times Impact times Confidence divided by Effort. It ranks backlog items by expected value per unit of engineering investment.

Q: What is the biggest mistake teams make with RICE scoring? A: Applying RICE without agreed definitions for each dimension, producing scores that reflect individual assumptions rather than shared evidence. A RICE session without pre-aligned definitions is worse than gut-feel prioritization.

Q: How do you estimate Impact in RICE scoring? A: Use Intercom's scale: 3 massive, 2 high, 1 medium, 0.5 low, 0.25 minimal. Enforce that no more than 20% of items receive a 3 to maintain discrimination power. Use forced-rank comparison when too many items tie at 3.

Q: What does Confidence measure in RICE scoring? A: How strong your evidence is for the Reach and Impact estimates. 100% means hard data from A/B tests or extensive research. 50% means assumptions. 20% means speculation. Confidence is the integrity check that prevents building on guesswork.

Q: How do you prevent RICE scoring from becoming a political exercise? A: Score independently before group discussion to prevent anchoring, investigate score outliers rather than averaging them, and require customer evidence for Confidence above 50%.

HowTo: Prioritize a Product Backlog Using RICE Scoring

  1. Define all four RICE dimensions in team-specific terms before scoring: what Reach period you are using, what the Impact scale points mean for your product, what evidence thresholds correspond to each Confidence percentage
  2. Pull actual user segment sizes from analytics for Reach estimates rather than using the total user base or aspirational projections
  3. Score each backlog item independently before group discussion to prevent anchoring bias from the first score stated
  4. Investigate score divergences rather than averaging them — disagreements about Impact or Confidence reveal different assumptions about evidence or user behavior that should be resolved
  5. Apply the 20 percent rule for Impact scores: if more than 20 percent of items receive the maximum score, run a forced-rank comparison to restore discrimination between genuinely massive and merely high-impact items
  6. Run the sanity check after scoring by comparing the RICE-ranked list against team intuition — if the list seems wrong, investigate which dimension carries systematic bias rather than overriding the scores
lenny-podcast-insights

Practice what you just learned

PM Streak gives you daily 3-minute lessons with streaks, XP, and a leaderboard.

Start your streak — it's free

Related Articles