The evidence-based approach to feature prioritization

The evidence-based approach to feature prioritization

Stop building features based on gut feel. Start building features based on data.

Tucker Schreiber·February 16, 2026·6 min read

Your VP of Sales just dropped three enterprise deals in your lap. Each wants a different feature. You have one sprint.

Your designer is pushing for a UX overhaul she's been sketching for months. Your CEO saw a competitor demo and wants "something like that." And buried in your backlog is a feature request from forty-seven support tickets that nobody has championed because it's unsexy.

Which one do you build?

If you're like most PMs, the answer comes down to who argued loudest in the last meeting. RICE scores are theatre — everyone knows you can massage the numbers to justify whatever you already decided. Stack ranking is a popularity contest. And "data-driven" usually means "I found one metric that supports my gut feeling."

Feature prioritization is the highest-leverage thing you do as a PM. Get it right and you ship things that move numbers. Get it wrong and you burn months on features that collect dust. There's a better way than vibes.

Three ways gut-feel prioritization breaks

The HIPPO problem

HIPPO: Highest Paid Person's Opinion. CEO says "we need AI features." The team builds it — not because evidence supports it, but because authority overrides analysis.

This isn't about bad leaders. It's about absent systems. No shared evidence base means the most confident person wins by default. This is how you end up with feature-rich products that solve nobody's actual problem.

Survivorship bias

The customers who email you, tweet at you, and hop on calls are a self-selected group. They are your most engaged users. They are not representative.

Building from vocal feedback means you optimize for power users while ignoring the silent majority who churned last month without a word. Survivorship bias is dangerous because it feels like customer-centricity. You are listening to customers. Just the wrong ones.

The recency trap

Last week's support tickets feel more urgent than last quarter's churn analysis. One angry customer on a morning call can redirect an entire sprint.

Without a system that aggregates evidence over time, you chase whatever's freshest instead of whatever's most important.

The four types of product evidence

Qualitative evidence

Customer interviews are the richest source. A good interview reveals not just what users want, but why — the underlying jobs, frustrations, and workflows. If you're not running regular interviews, the guide on synthesizing customer interviews covers the full process.

Support tickets and feedback are unsolicited signal. They arrive on their own schedule, reflecting moments of genuine friction. The weakness: volume. You need a system to aggregate hundreds of tickets into patterns, not react to individual complaints.

Sales call notes and lost-deal reasons reveal what prospects expected but didn't find. Gold for prioritizing features that expand your addressable market.

Quantitative evidence

Product usage data shows what users actually do, not what they say they do. Adoption rates, funnel drop-offs, time-on-task — objective signal about where your product delivers and where it fails.

Business metrics — retention curves, conversion rates, NPS — provide outcome-level evidence. A feature request might sound compelling in an interview, but if the problem affects 2% of users contributing 1% of revenue, the math changes.

Combining both

Neither is sufficient alone. Interviews tell you why something is broken. Usage data tells you how many people are affected. The best decisions come from convergent evidence — themes that show up in both conversations and behavioral data.

Framework: from raw evidence to ranked priorities

Step 1: Map evidence to themes

Raw evidence — transcripts, tickets, usage logs — isn't useful for prioritization in its unprocessed form. Extract discrete signals, cluster them into themes.

A theme is a recurring pattern across multiple sources. "Users struggle to find the export function" is a theme. "User #4782 couldn't find the export button" is a data point.

Three things matter:

  • Extract structured signals from each source. Be concrete. "Onboarding is confusing" is too vague. "Users abandon onboarding at workspace creation because they don't understand team vs. personal" is useful.
  • Cluster by similarity. Two users complaining about different aspects of the same workflow belong to the same theme, even if they use different words.
  • Preserve attribution. Every theme links back to source evidence. You'll need this when a stakeholder asks "where did this come from?" A theme with twelve attributed sources beats one plucked from thin air.

This mapping step is where most teams break down. It's tedious, subjective, and error-prone across dozens of sources. Tools like Mimir automate extraction and clustering, but whether you use AI or a spreadsheet, the output is the same: themes backed by traceable evidence.

Step 2: Assess frequency and severity

Score each theme on two dimensions:

  • Frequency: How many distinct sources mention it? (Not mentions — one user ranting ten times is still one source.)
  • Severity: How much does it affect the user's ability to get value? Scale: cosmetic (annoying but workable), moderate (requires workaround), severe (blocks core workflow), critical (causes churn).

Plot them on a frequency-severity matrix. Upper-right quadrant = highest priority.

Step 3: Estimate impact and effort

Frequency and severity tell you which problems matter. Impact and effort tell you which solutions to pursue.

For each high-priority theme, define a potential solution and estimate:

  • Impact: Be specific. "Reducing onboarding drop-off from 40% to 20% adds ~500 activated users/month." Tie to a metric.
  • Effort: T-shirt sizes. Fake-precise hour estimates are fiction at this stage.

High-impact, low-effort = quick wins. High-impact, high-effort = strategic bets needing exec buy-in. Low-impact anything = cut it.

Step 4: Validate with converging evidence

Before committing, check for evidence convergence:

  • Customers mention it in interviews? (qualitative)
  • Usage data shows a drop-off? (quantitative)
  • Support team flags it as recurring? (operational)
  • Solving it connects to a metric you're actively trying to move? (strategic)

All four? Ship it yesterday. Only one? Treat it with skepticism, no matter how passionate that one source is.

Presenting to stakeholders

Lead with the problem

Executives are more persuaded by well-defined problems than proposed solutions.

Weak: "We should rebuild the onboarding flow."

Strong: "34 of 50 interviewees mention onboarding confusion. Usage data shows 41% abandon at step 3. 30-day retention is 23%, and onboarding is the #1 reason in churn surveys."

The solution follows once the problem is established.

Show your evidence trail

Every recommendation should be traceable. When a stakeholder asks "where did this come from?", point to specific interviews, tickets, and data points.

Simple evidence summary per recommendation:

  • Theme: One-sentence problem description
  • Sources: Count by type (e.g., "12 interviews, 47 tickets, 3 usage cohorts")
  • Key quotes: Two or three representative customer statements
  • Metric impact: Which metric, what projected improvement
  • Effort: T-shirt size with brief rationale

Anticipate the HIPPO

If a senior stakeholder's pet feature ranks low in your evidence, don't ignore it. Address it directly. Show where it sits in the matrix. Acknowledge whatever supporting evidence exists. Let the comparison speak.

Most leaders, when presented with clear evidence, adjust. The ones who won't were never going to be persuaded by data — and now you have a paper trail.

Making it stick

Start with your next prioritization decision — even if it's choosing between two features for next sprint:

  1. Gather evidence. Interviews, feedback, tickets, usage data you already have.
  2. Map to themes. Extract signals, cluster, preserve attribution.
  3. Score frequency and severity. Be honest about what the evidence says, not what you want it to say.
  4. Estimate impact and effort. Tie impact to metrics. Rough effort sizing.
  5. Present with evidence. Lead with the problem. Show sources.

The framework sharpens with repetition — themes get more precise, severity assessments get more calibrated, and stakeholders start expecting evidence-backed recommendations instead of opinion-backed ones.

Teams that prioritize based on evidence don't just ship better features. They ship fewer features — and each one matters more. For how this connects to roadmap planning, see fixing your roadmap with customer evidence.

Related articles

Ready to make evidence-based product decisions?

Paste customer feedback into Mimir and get ranked recommendations in 60 seconds.

Try Mimir free