Part 2: Defining the Right Metrics: Leading vs. Lagging KPIs

Part 2: Defining the Right Metrics: Leading vs. Lagging KPIs
Photo by Nick Hillier / Unsplash
TL;DR: Metrics guide your team’s decisions, but the wrong ones can mislead. This post breaks down how to define the right KPIs and structure them effectively for live-service games.

Series: Data as a North Star – Part 2 of 5

In live-service games, strong metrics help teams prioritize, evaluate impact, and stay aligned. But poor metrics—or unclear definitions—can derail learning and lead to false conclusions.

That’s why choosing the right metrics (and the right way to read them) is as important as the feature itself. And above all, you need specificity. Many teams get this wrong by choosing broad KPIs too early or skipping the process of mapping metrics to the smallest behavior they’re actually trying to shift.

Go Granular: Metrics Begin With Behavior

Metrics are not just abstract numbers — they are reflections of player behavior. If you want to influence outcomes, you need to start by defining the smallest, most specific action you expect to change.

Instead of asking, “Will this improve retention?”, ask:

  • Will players open more chests?
  • Will they complete the tutorial faster?
  • Will they replay this battle node more often?

The smaller and more behaviorally grounded your metric is, the easier it is to validate your hypothesis and generate actionable insights.

🎯 Pro tip: Start by identifying the one action that, if it changed, would show your feature is working.

What is Metric Mapping?

Metric mapping is the process of identifying what success looks like, how you’ll measure it, and what could go wrong — before launch. It connects your hypothesis to real behavior.

A good metric map includes:

  • Primary Metric: The smallest, specific behavior you're trying to shift.
  • Guardrails: Metrics that ensure you’re not causing unintended harm (e.g., crash rate, churn).
  • Evaluation Window: Timeframe over which results should be evaluated (e.g., D1–D3).
  • Expected Direction & Magnitude: Are you expecting a lift or a drop? By how much?
✅ Example:Primary: % of players completing all 3 event battlesGuardrail: Average session length (ensure sessions aren’t bloated)Window: D2–D4Expectation: +20% vs. control

Understanding Metrics, Dimensions, and Cuts

Metrics only become useful when paired with the right context. That’s where dimensions and cuts come in.

Term Definition Example
Metric A quantitative measurement D1 retention, ARPDAU
Dimension A lens that segments the data Country, platform, player cohort
Cut A specific view of a metric using dimensions D1 retention for Tier 1 Android users
Aggregation Time-based grouping applied to the metric Daily, weekly, cumulative, cohort
🎯 Example: "% of players opening second chest by D3 in Tier 1 iOS" is a meaningful, actionable cut that directly reflects engagement.

Snapshot vs. Cumulative vs. Cohorted Views

Type Definition Best For
Daily Snapshot of behavior on a specific day Short-term spikes, early feedback
Weekly Aggregated behavior over 7 days Mid-range trend analysis
Monthly Aggregated behavior over a 30-day cycle Long-term shifts or seasonality
Cumulative Running total over time Measuring full impact (e.g. revenue)
Cohorted Groups players by start time Tracking retention or funnel behavior
🧠 Tip: Daily spikes can be misleading. Use weekly, monthly, and cohort views to understand longer-term patterns and behavior.

Leading vs. Lagging Metrics

One of the most common mistakes in game teams is relying only on lagging metrics — like revenue or retention — which only move after behavior changes.

Type Definition Example
Leading KPI Early signals tied to specific behaviors Tutorial completion, retry attempts, chest open rate
Lagging KPI High-level outcomes influenced by many behaviors D7 retention, ARPDAU, churn
🎯 Use case: If your goal is to improve D7 retention, leading metrics like mid-session exits or event re-engagement are better early signals.

Great product teams identify the right leading indicators that predict movement in their lagging KPIs—and they root those indicators in granular, measurable player actions.


How Do You Know If You Chose the Right Metric?

Ask yourself:

  • Does this metric reflect the specific behavior I’m trying to influence?
  • Will it move before my broader KPIs (retention, revenue) change?
  • Is it sensitive enough to show impact with small sample sizes?
  • Can I explain the logic to a stakeholder in one sentence?

If you answer “no” to any of those, your metric may be too broad—or disconnected from your hypothesis.


Mini Case Study: When the Metric Was Wrong

🎮 Scenario: A team launched a new PvE event with a hypothesis that it would improve engagement by encouraging players to play more battles during off-peak hours.Original Metric: D1 retentionResult: No significant change, feature deemed unsuccessful.
🔍 What went wrong?

D1 retention didn’t directly reflect the behavior the team hoped to influence—participation in event battles. When they drilled into metrics like "average battles played in event per user" and "event participation rate," they saw a measurable lift. That increase also contributed to longer session lengths and higher D3 retention.

✅ Lesson: Always connect your metric to the core player behavior you're trying to shift. Lagging metrics like retention may mask early wins if you’re not granular enough in how you define success.

Final Thought

Metrics should help you learn, not just report. When you define them with granular behavior, layer in smart dimensions, and apply them in context, they become powerful tools for building games that grow and evolve with your players.

If Part 1 was about forming the right hypothesis, this post helps you measure the right outcome. Coming up next in Part 3: How to use A/B testing to validate those outcomes in the wild.


Coming Up Next:

Part 3 – A/B Testing Best Practices for Live Service Games