How to prioritize a product backlog using RICE scoring requires defining Reach, Impact, Confidence, and Effort in terms specific to your product before scoring anything — because teams that apply RICE without agreed definitions produce scores that reflect individual assumptions rather than shared evidence, making the prioritization worse than a gut-feel list.
RICE is a prioritization framework developed at Intercom. Score = (Reach × Impact × Confidence) / Effort. Used correctly, it surfaces which backlog items produce the most value per unit of engineering investment. Used incorrectly, it provides false precision on bad assumptions.
The RICE Formula
RICE Score = (Reach × Impact × Confidence) / Effort
Reach: Number of users affected per time period
Impact: Effect on the target metric per user (0.25 to 3)
Confidence: How confident are we in our estimates (10-100%)
Effort: Person-months of work required
How to Estimate Each Dimension Accurately
Reach — Who and How Many?
Reach measures how many users will be affected by this item in a defined time period (typically per quarter).
Common mistakes:
- Using total user base instead of affected users (a feature for enterprise admins affects 2% of users, not 100%)
- Using aspirational numbers rather than current segment sizes
Better approach: Pull the actual count from your analytics. "The user segment who would encounter this feature has 4,200 monthly active users" is more reliable than "we have 50,000 total users."
Impact — How Much Does It Move the Needle?
Intercom's Impact scale: 3 = massive, 2 = high, 1 = medium, 0.5 = low, 0.25 = minimal.
Common mistake: Assigning 3 to everything. If everything is massive impact, you've lost the discrimination power of the scale.
Calibration rule: In any given scoring session, no more than 20% of items should receive a 3. If more than 20% score 3, force a forced-rank comparison: "Of these items that all scored 3, which truly has the most impact? Assign that one 3 and move the others to 2."
According to Shreyas Doshi on Lenny's Podcast, Impact is the dimension where PM intuition is most unreliable — PMs consistently overestimate the impact of features on user behavior, especially for features that solve problems the PM personally finds frustrating rather than problems identified through customer research.
Confidence — How Strong Is Your Evidence?
- 100%: Hard data (A/B test result, customer research with 8+ interviews, analytics showing clear drop-off)
- 80%: Soft data (qualitative interviews with 3–5 users, sales call observations)
- 50%: Assumption ("We believe users want this because...")
- 20%: Speculation ("We think this might help")
The confidence multiplier is the integrity check of the RICE framework. A high Reach × Impact score with 20% confidence should surface an experiment, not a full build.
Effort — True Cost, Not Ideal Case
Effort should be estimated in person-months including:
- Engineering development
- Design work
- QA testing
- PM spec and review time
- Post-launch monitoring
Use your team's historical velocity for similar-sized items rather than bottom-up estimation, which consistently underestimates by 30–50%.
According to Gibson Biddle on Lenny's Podcast, effort estimation is where RICE breaks down most often — teams use best-case estimates rather than historical averages, producing RICE scores that favor complex features over simple ones because the effort input is systematically underestimated for large items.
Running a RICE Scoring Session
Step 1: Define Dimensions Together
Before scoring, align the team on:
- What "Reach" means in your context (per month? per quarter?)
- What the Impact scale points represent for your specific product
- What constitutes 100%, 80%, 50%, 20% confidence in your evidence standards
This alignment prevents one PM scoring 3 because a feature is "strategically important" while another scores 2 because the user research is sparse.
Step 2: Score Independently, Then Compare
Each PM or stakeholder scores independently before the group discussion. This prevents anchoring bias — the first score stated in a group setting disproportionately influences all subsequent scores.
Step 3: Investigate Outliers
When scores diverge significantly (one person gives 3, another gives 0.5 for Impact), the divergence is information — it means different assumptions about the evidence or the user behavior. Investigate rather than average.
Step 4: Apply the Sanity Check
After scoring, check whether the ranked list matches your team's intuition about which items matter most. If the list seems wrong, don't override the scores — investigate which dimension is carrying bias.
According to Lenny Rachitsky's writing on product prioritization, RICE is most valuable not as a ranking machine but as a structured conversation tool — the process of discussing why scores diverge surfaces the assumptions and evidence gaps that, once resolved, align the team far more than any framework output.
FAQ
Q: What is RICE scoring in product management? A: A prioritization framework where RICE Score equals Reach times Impact times Confidence divided by Effort. It ranks backlog items by expected value per unit of engineering investment.
Q: What is the biggest mistake teams make with RICE scoring? A: Applying RICE without agreed definitions for each dimension, producing scores that reflect individual assumptions rather than shared evidence. A RICE session without pre-aligned definitions is worse than gut-feel prioritization.
Q: How do you estimate Impact in RICE scoring? A: Use Intercom's scale: 3 massive, 2 high, 1 medium, 0.5 low, 0.25 minimal. Enforce that no more than 20% of items receive a 3 to maintain discrimination power. Use forced-rank comparison when too many items tie at 3.
Q: What does Confidence measure in RICE scoring? A: How strong your evidence is for the Reach and Impact estimates. 100% means hard data from A/B tests or extensive research. 50% means assumptions. 20% means speculation. Confidence is the integrity check that prevents building on guesswork.
Q: How do you prevent RICE scoring from becoming a political exercise? A: Score independently before group discussion to prevent anchoring, investigate score outliers rather than averaging them, and require customer evidence for Confidence above 50%.
HowTo: Prioritize a Product Backlog Using RICE Scoring
- Define all four RICE dimensions in team-specific terms before scoring: what Reach period you are using, what the Impact scale points mean for your product, what evidence thresholds correspond to each Confidence percentage
- Pull actual user segment sizes from analytics for Reach estimates rather than using the total user base or aspirational projections
- Score each backlog item independently before group discussion to prevent anchoring bias from the first score stated
- Investigate score divergences rather than averaging them — disagreements about Impact or Confidence reveal different assumptions about evidence or user behavior that should be resolved
- Apply the 20 percent rule for Impact scores: if more than 20 percent of items receive the maximum score, run a forced-rank comparison to restore discrimination between genuinely massive and merely high-impact items
- Run the sanity check after scoring by comparing the RICE-ranked list against team intuition — if the list seems wrong, investigate which dimension carries systematic bias rather than overriding the scores