How to define product success metrics for a new feature requires answering three questions before any code is written: what user behavior change will prove the feature is working, what business outcome does that behavior change produce, and what guardrail metrics will tell you if you are creating harm in the process of achieving the primary goal?
Feature success metrics defined after launch are not success metrics — they are retrospective reporting. The discipline of defining metrics before building forces the team to be specific about what they are trying to achieve, which often reveals that the feature is solving the wrong problem.
The Metric Hierarchy for a New Feature
H3: Level 1 — The Primary Success Metric
One metric that directly measures the behavior change the feature is designed to create.
Rules:
- It must move within 14–30 days of the feature launching (if it takes 90 days to see movement, you cannot course-correct)
- It must be caused by the feature, not correlated (if the metric moves for other reasons, you cannot attribute it)
- It must be measurable today (before you define it as a success metric, confirm it can be pulled from your analytics system)
Examples by feature type:
| Feature | Primary success metric | |---------|----------------------| | Bulk export | % of eligible users who export within 14 days | | Team invitation flow redesign | 14-day team invitation rate | | Mobile push notifications | 7-day notification opt-in rate | | AI-generated summaries | Summary usage rate per user per week |
H3: Level 2 — The Business Impact Metric
One metric that links the feature's user behavior change to a business outcome. This is often a lagging indicator — it takes longer to move, but it's what the feature ultimately exists to improve.
Examples:
- Feature drives activation → business impact: 30-day retention rate
- Feature drives expansion → business impact: seats added per account within 60 days
- Feature reduces friction → business impact: support ticket volume per user
The business impact metric should be in the feature's PRD as the ultimate justification for building it.
According to Lenny Rachitsky's writing on feature metrics, the most common metric failure is defining a primary metric that is too lagging — 90-day retention is a business impact metric, not a primary success metric for a specific feature, because you cannot distinguish the feature's contribution from dozens of other changes that happened in those 90 days.
H3: Level 3 — Guardrail Metrics
2–3 metrics that must not degrade as a result of the feature. Guardrails prevent the team from hitting the primary metric at the expense of unintended harm.
Common guardrails:
- Feature does not increase app load time above baseline
- Feature does not increase support ticket volume by more than X%
- Feature does not decrease NPS below baseline
- Feature does not reduce usage of related core features (cannibalization check)
The guardrail metrics should be defined by asking: "What could go wrong with this feature that would harm the user or business without showing up in the primary metric?"
According to Shreyas Doshi on Lenny's Podcast, the guardrail metric most commonly missed is feature cannibalization — a new feature that creates a better path to an outcome may reduce usage of an existing feature that was essential to retention, and without a guardrail on the existing feature's usage, the PM doesn't know they've broken something until churn increases months later.
Setting Targets
H3: The Baseline-Plus-Improvement Approach
For most features, set the target as baseline performance plus a specific improvement:
- "Current bulk export rate among eligible users: 0% (feature doesn't exist). Target: 25% of eligible users export within 14 days of launch."
- "Current team invitation rate: 18% within 14 days. Target: 28% within 14 days after redesign."
The improvement target should be based on evidence: comparable feature launches at your company, industry benchmarks, or the expected impact from user research.
H3: The Minimum Viable Success Standard
Define what "not worth the investment" looks like before launch:
"If fewer than 15% of eligible users export within 14 days, we will consider this feature unsuccessful and evaluate whether to iterate or sunset it."
This minimum viable success standard creates accountability and prevents the team from declaring victory on a metric that is technically positive but commercially insignificant.
Instrumentation Requirements
H3: Define Events Before Building
Before engineering starts building, specify the analytics events that must be instrumented:
- Feature view event (user sees the feature)
- Feature first use event (user completes the primary action)
- Feature repeated use event (user uses the feature a second time within N days)
- Error events (user encounters an error in the feature flow)
No feature should ship without these events defined and verified in a staging environment. A feature that ships without instrumentation cannot be measured, which means success cannot be confirmed.
According to Gibson Biddle on Lenny's Podcast, the product teams that make the best feature investment decisions are the ones that treat instrumentation as a first-class requirement alongside the feature itself — every feature without instrumentation is a black box that can't be learned from, and over time these black boxes accumulate into a product with no learnable feedback loops.
FAQ
Q: How do you define success metrics for a new product feature? A: Define a primary success metric that measures the direct behavior change within 14–30 days, a business impact metric that links the behavior change to a business outcome, and 2–3 guardrail metrics that must not degrade.
Q: What is the difference between a feature success metric and a business impact metric? A: A feature success metric measures direct user behavior change caused by the feature within 14–30 days. A business impact metric measures the downstream business outcome — typically a lagging indicator like 30-day retention — that the feature ultimately supports.
Q: What are guardrail metrics in feature development? A: Metrics that must not degrade as a result of a feature launch — such as support ticket volume, NPS, load time, or usage of related core features. They prevent the team from hitting the primary metric at the expense of unintended harm.
Q: How do you set a feature success target before launch? A: Use baseline performance plus expected improvement based on comparable launches, industry benchmarks, or user research evidence. Define a minimum viable success standard — the result below which the feature is considered unsuccessful.
Q: When should you define feature success metrics? A: Before any code is written. Metrics defined after launch are retrospective reporting, not success criteria. Pre-launch definition forces specificity about what the feature is trying to achieve.
HowTo: Define Product Success Metrics for a New Feature
- Define the primary success metric — the specific user behavior change the feature is designed to create, measurable within 14 to 30 days of launch, before any code is written
- Define the business impact metric — the downstream business outcome the feature ultimately supports, such as 30-day retention or expansion revenue, as the long-term justification for the investment
- Define 2 to 3 guardrail metrics that must not degrade — support volume, NPS, load time, and cannibalization of related features — and set threshold values for each
- Set targets for the primary metric using baseline-plus-improvement approach based on comparable launches, benchmarks, or user research
- Define a minimum viable success standard — the result below which the feature is considered unsuccessful and will be iterated or sunsetted
- Specify the analytics events to instrument before engineering starts building and verify them in staging before launch — no feature ships without instrumented events