What is Metric Ownership and Governance?
Metric ownership and governance is the set of roles, rules, and routines that keep product metrics consistent, trusted, and change-managed. Think of a metric as a product: it has an owner, a definition, a data contract, a version, an SLA, and a lifecycle.
Why this matters
- Executive decisions rely on stable definitions (e.g., what counts as an “active user”).
- Feature teams need timely, accurate metrics to run experiments and measure impact.
- Without governance, duplicated metrics, hidden filters, and silent changes cause confusion and bad decisions.
Real tasks you’ll perform as a Product Analyst:
- Define or update a metric contract (owner, definition, data source, latency, version, and SLA).
- Lead a change process when logic must change (e.g., new signup flow).
- Set up quality checks and alerting for critical dashboards.
- Run a metric review to retire duplicates and align teams on one source of truth.
Concept explained simply + Mental model
Mental model: “Metric as a product.”
- Owner: accountable for correctness and clarity.
- Contract: what it measures, where it comes from, update cadence, allowed values, filters, and caveats.
- Version: breaking changes create a new version; old versions are deprecated with a timeline.
- Lifecycle: propose → review → adopt → monitor → change/deprecate.
RACI template (quick reference)
- Responsible: Metric Owner (usually a Product Analyst or Analytics Lead).
- Accountable: Product area lead or Head of Analytics.
- Consulted: Data Engineering, Product Manager, Legal/Privacy as needed.
- Informed: All metric consumers (squads, leadership, finance).
Core components of metric governance
1) Clear definition and documentation
- Name + short description
- Precise formula (filters, time windows, inclusions/exclusions)
- Business rules (e.g., only activated users; exclude staff/internal)
- Data sources and tables; transformation logic
- Latency/refresh cadence (e.g., daily by 07:00)
- Owner and contact
- Known caveats and validation checks
2) Data contract
- Schema: fields, types, allowed values
- SLAs: delivery time, completeness thresholds
- Quality expectations: null rates, duplication rules
- Change policy: how to propose, approve, and schedule schema or logic changes
3) Naming and versioning
- Stable names; use suffixes for breaking changes (e.g., signup_conversion_v2)
- Minor changes documented; breaking changes create new version
- Deprecation timeline and migration plan
4) Change management
- Proposal → review → approval → communication → migration → monitoring
- Side-by-side period to compare old/new versions
- Backfill strategy documented
5) SLAs, quality checks, and alerting
- Timeliness: by X time daily/weekly
- Completeness: threshold for rows/events
- Anomaly checks: spikes/drops, nulls, duplicates
- Alert channels and on-call rotation
6) Access, privacy, and compliance
- Audience: who can see and who can edit
- PII handling and retention rules
- Audit trail of changes
7) Review cadence and sunsetting
- Quarterly reviews for relevance and duplication
- Sunset unused metrics with a clear archive note
Worked examples
Example 1: Weekly Active Users (WAU)
- Name: weekly_active_users_v1
- Definition: Unique users with at least 1 qualifying session in the last 7 days (rolling). Exclude internal/staff.
- Data sources: events.app_sessions (user_id, ts, is_internal)
- Logic: distinct user_id where is_internal = false and ts between today-6 and today
- Latency: daily by 06:30; SLA 99.5%
- Owner: Product Analytics – Core Product
- Quality checks: day-over-day change within +/- 20% unless release flag
Example 2: Signup Conversion Rate (v2 change)
- Old: signups/landing_page_visits
- Issue: new multi-step flow; old denominator overcounts bots
- New (v2): signups/unique_verified_visitors with bot filter and 30-min sessionization
- Plan: run v1 and v2 in parallel for 2 weeks; communicate expected level shift; backfill v2 for 90 days
Example 3: Feature Adoption
- Name: feature_x_adoption_rate_v1
- Definition: users who used feature X at least once in D7 after activation divided by users activated in that period
- Data contract: requires events.feature_x_use with user_id, ts, and feature_flag_state
- Alert: if denominator < expected by 30% after an experiment, page the analyst on-call
Example 4: Deprecating a duplicate metric
- Problem: two teams built different “DAU” metrics
- Resolution: select one canonical definition; re-label dashboards; archive the duplicate with a redirect note and sunset date
How to implement governance in your team
- Inventory: list top 20 metrics used by leaders and squads.
- Assign owners: one per metric; publish contacts.
- Create contracts: fill a one-page template for each metric.
- Set SLAs and checks: timeliness, completeness, anomaly thresholds, alerts.
- Versioning rules: define when to create v2 and how to sunset v1.
- Change board: lightweight weekly review (15–30 minutes) with owner + data eng + PM.
- Quarterly review: prune, merge duplicates, update caveats.
Exercises (practice here, then complete the tasks below)
These mirror the Exercises section on this page. Write your answers in a doc or notes. Then compare with the provided solutions.
Exercise 1: Draft a metric contract
Create a contract for Weekly Active Users (WAU). Include: owner, purpose, exact definition, formula/SQL sketch, filters, data sources, latency/SLA, quality checks, versioning, change process, caveats.
- Checklist:
- Owner and contact present
- Formula includes inclusions/exclusions
- Latency and SLA quantified
- Quality checks and alerts defined
- Versioning and change steps clear
Exercise 2: Plan a breaking change
Your company adds email verification before activation. This affects “Activation Rate.” Propose a change plan with: v2 definition, side-by-side period, backfill, communication, dashboard migration, and deprecation date.
- Checklist:
- Rationale for change
- Parallel run timeline
- Backfill scope and limits
- Communication audience and channel
- Sunset/deprecation date
Common mistakes and how to self-check
- Silent redefinitions: Always version breaking changes and announce them.
- Vague formulas: Write explicit filters and time windows; avoid “etc.”
- No owner: Assign a single accountable owner.
- No SLA: Define refresh time and quality thresholds.
- Too many near-duplicate metrics: Prefer a single canonical metric.
- Skipping side-by-side comparisons: Always run old vs new to quantify level shifts.
Self-check: Could a new teammate reproduce the metric from your doc without asking you questions? If not, your contract needs more detail.
Practical projects
- Build a metric catalog for your product area (10–15 metrics) with owners, contracts, and SLAs.
- Run a metric consolidation: identify 3 duplicates and merge into one canonical version with deprecation notes.
- Implement a weekly change review: track proposals, decisions, and announce updates to consumers.
Who this is for
- Product Analysts who publish or maintain metrics and dashboards.
- PMs and Data Engineers collaborating on metric definitions and pipelines.
Prerequisites
- Basic SQL and data modeling familiarity.
- Understanding of your product’s core funnels and events.
Learning path
- Before: Event tracking design, SQL for analytics, core product metrics.
- Now: Metric ownership and governance (this lesson).
- Next: Experimentation metrics, north star and input metrics, dashboard QA.
Next steps
- Complete the two exercises and compare with the solutions.
- Take the Quick Test at the end. Tests are available to everyone; if you log in, your progress will be saved.
- Pick one practical project to implement this week.
Mini challenge
Find one high-impact metric in your team that lacks a clear owner or SLA. Assign an owner, draft a one-page contract, and set a 2-week review date.
Glossary
- Metric contract: A concise document that defines a metric’s logic, ownership, data sources, SLAs, and change process.
- Breaking change: A change that makes the new value not directly comparable with the old value.
- SLA: Service Level Agreement for timeliness and quality (e.g., daily by 07:00 at 99.5%).
- Side-by-side: Running old and new versions concurrently to measure differences before migration.