PM Metrics Cheat Sheet: A Comprehensive Guide

TL;DR

This PM Metrics Cheat Sheet gives product managers at tech companies a field-tested framework for selecting, implementing, and communicating the right metrics across product lifecycle stages. Unlike generic lists, it reflects actual patterns from hiring committee debates, executive reviews, and cross-functional alignment challenges at top tech firms. You’ll learn which metrics move the needle in promotion packets, how to avoid common measurement pitfalls, and what leaders actually look for in metric hygiene.

Who This Is For

This guide is for product managers in tech — especially mid-level PMs (L4–L6 at FAANG-tier) — who are preparing for promotion cycles, leading new product initiatives, or navigating stakeholder alignment on success criteria. It’s also valuable for early-career PMs at high-growth startups trying to professionalize their measurement approach. If your roadmap lacks clear KPIs, your stakeholder debates turn circular, or your A/B tests get questioned in debriefs, this cheat sheet addresses the real-world gaps that aren’t covered in textbooks.

What are the core categories of PM metrics that matter in real product decisions?

Every effective product leader segments metrics into five buckets: engagement, conversion, retention, satisfaction, and business impact. These aren’t theoretical — in a Q3 2023 platform initiative debrief, the hiring manager pushed back because the PM only reported daily active users (DAU) without showing cohort-based retention or funnel drop-off at key steps. Engagement metrics like session duration or feature adoption rate help answer “Are people using it?” Conversion metrics (e.g., sign-up to paid, click-to-purchase) reveal friction points. Retention — measured as week-1 or month-1 return rate — is the single strongest predictor of long-term product health; I’ve seen multiple promotion packets stalled because retention wasn’t tracked. Satisfaction includes NPS, CES, and support ticket volume — often overlooked until escalations happen. Business impact metrics like ARPU, LTV:CAC, or contribution margin are required when presenting to finance or execs. At Amazon, every PRFAQ must include at least one business metric; missing it delays review cycles.

Which metrics should you track during discovery versus launch?

During discovery, track leading indicators like intent-to-purchase scores, concept testing completion rates, or early sign-ups via waitlists — not lagging outcomes. In a 2022 AI assistant project at a major cloud provider, the team initially proposed measuring revenue impact during discovery, which delayed approval because it was premature. Instead, they pivoted to tracking % of users who said they’d use the tool daily in concept testing — a validated proxy. Post-launch, shift to lagging metrics: conversion rate, DAU/MAU, and retention curves. For a mobile app launch in Q2 2023, the team used Day 7 retention >40% as a go/no-go signal for scaling; that threshold was based on historical data from similar products. Use North Star metrics only after validating product-market fit; otherwise, they create false confidence. Early-stage products benefit more from diagnostic metrics (e.g., onboarding completion time) than vanity metrics like total downloads.

How do you align stakeholders on metrics before an experiment?

Pre-align by co-defining primary, secondary, and guardrail metrics in the experiment design doc — and get signatures. In a 2021 pricing test, the growth team launched a change that increased conversion by 12% but reduced average order value by 18%. The experiment was rolled back because AOV wasn’t listed as a guardrail metric upfront, causing friction with the revenue team. Primary metrics should be directional and tied to the hypothesis (e.g., “We believe simplifying the checkout flow will increase purchase completion by 10%”). Secondary metrics capture broader impacts (e.g., support tickets, session duration). Guardrail metrics protect against unintended consequences — common ones include churn rate, error rates, and customer satisfaction scores. At Google, every experiment requires at least one guardrail metric. If the bar isn’t set before launch, the data becomes negotiable in post-mortems. In a recent HC meeting, a PM’s impact was downgraded because their win claim ignored a 15% spike in help desk contacts.

What’s the difference between input and outcome metrics, and why does it matter?

Outcome metrics measure business or user behavior change (e.g., retention increased by 8%), while input metrics track effort or output (e.g., shipped 5 features). Hiring committees consistently downgrade PMs who conflate the two. In a 2022 promotion review, a candidate claimed “delivered 100% of roadmap” as a success — but the committee questioned impact because no outcome metric was tied to the delivery. Outcome metrics answer “Did it work?” Input metrics answer “Did we do it?” Executives care about outcomes; engineering managers care about inputs. A strong PM narrative links inputs to outcomes: “We shipped three onboarding improvements (input), resulting in a 14% increase in Day 3 activation (outcome).” At Meta, promotion packets require a “metric story” that traces actions to outcomes. When candidates only list shipped features, the packet fails the “so what?” test.

Interview Stages / Process

At FAANG-level companies, PM candidates are evaluated on metric literacy across three stages:

  1. Phone Screen (30–45 mins)
    Focus: Can you name and define key metrics? Expect questions like “How would you measure success for a social feed?”
    Timeline: 1 week from application.
    Evaluation: Interviewers look for structured thinking — e.g., breaking down into engagement, retention, and satisfaction. Saying “I’d track likes” without context gets a no-hire. Better answer: “Primary metric: time-in-feed. Secondary: follow rate. Guardrail: comment quality flagged by moderation.”

  2. Onsite Round 1 – Product Sense (60 mins)
    Focus: Design a product and define success metrics.
    Example: “Design a feature to increase driver retention for a rideshare app.”
    Evaluation: Hiring committees reject candidates who pick vanity metrics (e.g., total rides). Strong candidates segment drivers (new vs. experienced), propose cohort-based retention (e.g., % of drivers active in weeks 2–4), and suggest driver NPS as a satisfaction proxy. In Q4 2022, 60% of no-hire decisions in this round were due to poor metric selection.

  3. Onsite Round 2 – Execution (60 mins)
    Focus: Diagnose a metric drop or analyze an A/B test.
    Example: “DAU dropped 15% last week. How do you investigate?”
    Evaluation: Strong candidates segment the data (platform, geography, user tier) and check system health before jumping to product causes. One candidate in a 2023 debrief was praised for asking, “Was there an app store rejection or CDN outage?” before blaming UX changes.

  4. Onsite Round 3 – Leadership & Drive (45 mins)
    Focus: Tell a story of past impact using metrics.
    Evaluation: Committees look for causality — did the PM’s action cause the change? A candidate who said “revenue grew 20% during my project” was questioned until they isolated the feature’s contribution using holdback analysis. Top performers quantify impact: “Our recommendation engine drove 7% of total conversion, validated by a 3-point lift in a 10%-sized holdback.”

  5. Debrief & Hiring Committee (HC) Review (1–2 weeks post-onsite)
    Focus: Consensus on hire/no-hire, level, and compensation.
    Outcome: Metric clarity directly affects leveling. A candidate with ambiguous impact is often leveled down. In one HC, a PM was offered L5 instead of L6 because their metric story relied on team-wide growth, not attributable impact.

Common Questions & Answers

Q: How do you choose a North Star Metric?

A North Star should reflect core user value and correlate with long-term business health. For Slack, it’s weekly messages sent; for Airbnb, nights booked. Avoid metrics that can be gamed — e.g., “total uploads” on a content platform led to spam in one case. Instead, use “high-quality uploads” defined by engagement thresholds. In a 2021 strategy offsite, the product leadership at a fintech company shifted from “accounts opened” to “3rd transaction completed” because the latter predicted retention.

Q: What’s a good retention rate?

It depends on the product type. For enterprise SaaS, 90%+ monthly retention is expected. For consumer apps, 40–50% Day 28 retention is strong. In a benchmarking exercise across 12 apps in 2022, median Day 7 retention was 32%. Below 20% indicates serious onboarding issues. One team improved from 18% to 41% by reducing required steps in sign-up — a change validated through funnel analysis.

Q: How do you measure feature adoption?

Use adoption rate: # of active users of the feature / total eligible users. Set a baseline — e.g., “We expect 60% of daily users to use the new search bar within 2 weeks.” Pair it with depth of use: e.g., average queries per user. In a 2023 launch, a PM defined “success” as 50% adoption with at least 2 uses per week. They hit 62%, but qualitative feedback revealed users didn’t understand the feature — showing why behavioral metrics need context.

Q: How do you handle conflicting metrics?

Trade-offs are normal. In a 2022 experiment, a UX change increased conversion by 9% but reduced average session duration by 12%. The team accepted the trade-off because conversion was the primary goal. Document the decision rationale: “We prioritized conversion over engagement because this is a transactional flow.” Avoid declaring “win” if guardrails are breached — one PM’s offer was rescinded after they ignored a 20% increase in support load.

Q: What’s the most underrated metric?

Time-to-value (TTV): how fast a user gets their first “aha” moment. At Dropbox, reducing TTV from 7 days to 48 hours doubled long-term retention. In a 2020 study of 8 apps, products that delivered value in under 24 hours had 2.5x higher retention at Day 30. One PM instrumented a “first successful upload” event and used it to trigger onboarding nudges — lifting retention by 11%.

Preparation Checklist

  1. Define your product’s North Star and align it with company OKRs.
  2. Map your user journey and identify 1–2 key metrics per stage (e.g., activation, adoption, retention).
  3. For every initiative, specify primary, secondary, and guardrail metrics in writing — get stakeholder sign-off.
  4. Instrument events to capture behavioral data (e.g., feature usage, funnel steps) before launch.
  5. Establish baselines before running experiments — historical trends prevent false signals.
  6. Practice explaining metric trade-offs: “We accept lower session duration for higher conversion because…”
  7. Use cohort analysis to isolate impact — avoid aggregate data traps.
  8. Run a weekly health check on core metrics: track anomalies, segment data, and flag risks early.
  9. Benchmark your metrics against industry standards (e.g., 40%+ Day 7 retention for mobile apps).
  10. Document metric definitions in a shared glossary to prevent misalignment.

Mistakes to Avoid

  1. Measuring everything, proving nothing
    One PM tracked 37 metrics in their dashboard — during the review, the VP asked, “Which one matters?” The PM couldn’t answer. Limit primary metrics to 1–2 per goal. In a 2022 HC, a candidate was dinged for reporting “improved 8 metrics” without prioritizing impact.

  2. Confusing correlation with causation
    A team saw a 20% DAU increase after launching a new feature — but failed to run an A/B test. Later analysis showed the spike coincided with a marketing campaign. Always isolate variables. At Netflix, every feature launch includes a holdback group to measure true incremental impact.

  3. Ignoring metric decay
    Metrics lose validity over time. “Clicks” worked as a proxy for engagement in 2015, but today users scroll without intent. One product team at LinkedIn kept using “profile views” as a success metric — until data showed 70% came from automated bots. Re-evaluate metric relevance quarterly.

  4. Presenting clean data without context
    A PM once showed a 15% increase in conversion with no segmentation. When asked, they couldn’t say if it was driven by new or existing users. Always show breakdowns: by region, device, user tier. In a 2023 review, a candidate impressed the committee by revealing that a metric lift was concentrated in one under-served market — leading to a strategic shift.

FAQ

What’s the most important metric for a new product?

Time-to-value (TTV) is the most critical metric for new products. If users don’t experience value quickly, they churn. For a B2B tool, TTV might be “first successful integration in under 2 hours”; for a consumer app, “first meaningful interaction within 24 hours.” At Notion, reducing TTV by simplifying onboarding increased 7-day retention by 18%. Focus on shortening the path from sign-up to “aha” moment — this has more impact than broad awareness metrics.

How do you measure product-led growth?

Track three core metrics: activation rate, expansion rate, and net revenue retention (NRR). Activation is the % of users who hit a value milestone (e.g., invite a teammate, run a report). Expansion measures upsell within existing accounts — look for >10% quarterly user growth per account. NRR above 120% indicates strong organic growth. In a 2022 review, a SaaS PM used these metrics to justify doubling the self-serve team — NRR had grown from 105% to 130% in six months.

Should you track NPS for every product?

No — NPS is useful only when you have volume and actionability. For low-touch products, NPS surveys can feel intrusive and yield low response rates (<5%). One PM at a mobile bank saw NPS drop but couldn’t trace it to any product change — the dip was due to external brand sentiment. Use NPS selectively: when you can close the loop with detractors, or when benchmarking against competitors. For most internal tools, CES (Customer Effort Score) is more actionable.

How do you set metric targets?

Base targets on historical trends, benchmarks, or incremental opportunity. For example, if current conversion is 5%, a 10% lift (to 5.5%) is reasonable; doubling it to 10% requires extraordinary changes. Use rule of three: worst case, expected, best case. In a 2021 roadmap review, a PM proposed a “100% increase in engagement” — the committee rejected it as unrealistic without modeling. Strong targets are ambitious but grounded in data.

What’s the difference between DAU and MAU?

DAU (Daily Active Users) counts unique users per day; MAU (Monthly Active Users) counts unique users over 30 days. The DAU/MAU ratio indicates engagement — a ratio above 0.2 (20%) is healthy for most apps; above 0.5 is exceptional. For example, Instagram’s DAU/MAU is ~0.65. But don’t rely on the ratio alone — one app had a high ratio but declining MAU, signaling shrinking reach. Always analyze both metrics together.

How do you prove your impact in a promotion packet?

Quantify attributable impact using controlled experiments or holdback analysis. Say “my feature drove a 5% increase in conversion” only if you can isolate it. One PM used a 5% holdback group to show their recommendation engine contributed 8% of total revenue — that number became the centerpiece of their L6 promotion. Avoid team-wide metrics unless you led the initiative. Committees want to know: what changed because of you?

Related Reading

The book is also available on Amazon Kindle.

Need the companion prep toolkit? The PM Interview Prep System includes frameworks, mock interview trackers, and a 30-day preparation plan.


About the Author

Johnny Mai is a Product Leader at a Fortune 500 tech company with experience shipping AI and robotics products. He has conducted 200+ PM interviews and helped hundreds of candidates land offers at top tech companies.