luvv to helpDiscover the Best Free Online Tools
Topic 6 of 8

Data Quality SLAs And Ownership

Learn Data Quality SLAs And Ownership for free with explanations, exercises, and a quick test (for Analytics Engineer).

Published: December 23, 2025 | Updated: December 23, 2025

Why this matters

As an Analytics Engineer, your work powers dashboards, decisions, and downstream automations. Clear Data Quality SLAs (service level agreements) and ownership make quality predictable. Without them, you face silent data drift, unclear escalation, and blame loops. With them, the right people respond quickly, stakeholders trust data, and you protect business outcomes.

  • Real tasks you will do: define freshness targets for critical models, agree response times for incidents, set ownership boundaries across producer/transform/consumer teams, and document runbooks and exception windows.
  • Impact: fewer broken dashboards, faster fixes, and a shared language for quality trade-offs.

Concept explained simply

Think of data products like public transport. Riders expect buses to arrive on time (freshness), have room available (completeness), and go to the right stops (accuracy/consistency). An SLA is the timetable plus escalation if a bus is late.

  • SLI (Indicator): how you measure. Example: minutes from source extract to model availability.
  • SLO (Objective): the target. Example: 95% of days the model is ready by 06:30 UTC.
  • SLA (Agreement): the commitment and what happens if you miss. Example: on-call within 15 minutes, stakeholder notice, root cause analysis within 24 hours.
Mental model

Use the 3Ms: Measure, Mean it, Mobilize.

  • Measure: choose 1–2 precise SLIs per dataset (freshness, completeness, accuracy, validity, uniqueness, consistency).
  • Mean it: set realistic SLOs with an error budget (allowed failures over time).
  • Mobilize: define who acts, how fast, and how you communicate.

Key components of a Data Quality SLA

  • Scope: which datasets/models, which business use cases (e.g., Finance close, Daily Ops).
  • SLIs: clear, machine-measurable metrics (e.g., max lag in minutes; % non-null keys; reconciliation % vs source).
  • SLOs: target levels and measurement window (e.g., 99% of runs per calendar month).
  • Error budget: how many allowed misses before corrective actions.
  • Incident severity and response: severity levels, response and resolution times.
  • Ownership (RACI): Responsible, Accountable, Consulted, Informed per dataset and pipeline.
  • Monitoring and evidence: where metrics live and how they alert.
  • Exceptions and change management: temporary waivers with end dates; process for updating the SLA.
  • Review cadence: quarterly review with producers and consumers.
Common SLIs by dimension
  • Freshness: max source-to-model lag; time of latest successful load.
  • Completeness: % of expected rows; % non-null primary keys; row count delta vs baseline.
  • Accuracy/Reconciliation: % variance vs source-of-truth; sample audit pass rate.
  • Validity: % values passing rules (e.g., date within range, enums).
  • Uniqueness: % duplicate primary keys.
  • Consistency: schema compatibility checks; referential integrity pass rate.

Ownership model

Ownership clarifies who fixes what. Use a simple RACI per dataset and per pipeline step.

  • Producers (source systems): own upstream correctness and schema changes.
  • Analytics Engineering: owns transformations, tests, and data models exposed to consumers.
  • Data Stewards/Business Owners: define business rules, criticality, and accept SLOs.
  • BI/Consumers: report issues, validate business fit, and help prioritize.
RACI example pattern
  • Finance_MART.daily_bookings freshness incident: Responsible = AE on-call; Accountable = AE manager; Consulted = Finance steward; Informed = Source app team.
  • Schema change breaking ingestion: Responsible = Source team; Accountable = Source product owner; Consulted = AE; Informed = BI leads.

Worked examples

Example 1: Daily finance model freshness

  • SLI: minutes from last successful extract to model availability.
  • SLO: 99% of business days model ready by 06:30 UTC.
  • Error budget: up to 2 misses/month.
  • Severity: Sev-2 if 06:30–07:00; Sev-1 after 07:00 on business days.
  • Ownership: AE on-call Responsible; Finance steward Consulted.
  • Escalation: page on-call; status note to Finance within 15 minutes; RCA in 24 hours.

Example 2: Event stream completeness

  • SLI: % events landed within 60 minutes of emission.
  • SLO: 99.5% per week.
  • Error budget: 0.5% weekly.
  • Sev policy: Sev-2 if weekly projection shows budget breach; Sev-1 if slice affecting checkout events > 0.7% for > 2 hours.
  • Ownership: Platform data ingestion team Accountable; AE Consulted for downstream model lag.

Example 3: Accuracy via reconciliation

  • SLI: absolute % variance of revenue vs billing system for prior day.
  • SLO: |variance| ≤ 0.5% for 95% of days.
  • Controls: automated recon report; sample manual audit monthly.
  • Action: if |variance| > 0.5% then downgrade dashboards and notify Finance.
Example 4: Schema change contract
  • Rule: producers announce breaking changes 7 days ahead; analytics confirms migration plan within 2 days.
  • Exception: emergency fix allowed with 24-hour waiver and explicit risk note.

Templates and snippets

Minimal SLA template
Dataset: [name]
Criticality: [High/Medium/Low]
Business use: [who uses it, for what]
SLIs: [metric name + definition]
SLOs: [target + window]
Error budget: [allowed misses]
Severity & response: [Sev-1/2/3: response & resolution times]
Ownership (RACI): [Responsible, Accountable, Consulted, Informed]
Monitoring: [where metrics/alerts live]
Exceptions: [process, approver, end date]
Review: [cadence]
Incident runbook skeleton
  • Detect: alert fires, capture context (job id, model, last good timestamp).
  • Stabilize: stop dependent downstream jobs if needed.
  • Diagnose: source availability, ingestion, transform, publish layers.
  • Communicate: post status with ETA and business impact.
  • Fix: patch or rollback; re-run with backfill.
  • Recover: verify SLIs, re-enable schedules.
  • Learn: RCA and prevention action by next business day.
Severity guide
  • Sev-1: critical decision blocked or regulatory impact. Response 15 min, resolution 2 hours or workaround.
  • Sev-2: major dashboard degraded. Response 30 min, resolution same day.
  • Sev-3: minor data delay. Response 4 hours, resolution 2 days.

Hands-on exercises

These mirror the exercises below. Try them now. The quick test is open to everyone; saving your progress is available to logged-in users.

  1. Draft an SLA: Pick a dataset and fill the minimal SLA template. Include one freshness SLI and one completeness SLI.
  2. Map ownership (RACI): For one pipeline, assign R, A, C, I for normal ops and for incidents.
  3. Write a runbook: For a missed freshness SLO, write 6–8 bullet steps from alert to recovery.
  • Checklist: SLIs are measurable; thresholds are realistic; error budget defined; response times clear; owners named; exception window rule exists.

Common mistakes and self-check

  • Vague metrics: “data is up to date.” Self-check: can a script compute it?
  • Targets without baseline. Self-check: did you analyze 2–4 weeks of historical performance?
  • No error budget. Self-check: do you know how many misses trigger corrective action?
  • Alerting to everyone or no one. Self-check: is there exactly one Responsible on-call?
  • Ignoring schema changes. Self-check: do you have a notice period and compatibility checks?
  • One-size-fits-all severity. Self-check: does severity depend on business impact time?

Practical projects

  • Create SLAs for three datasets of different criticality and implement automated SLI checks and alerts.
  • Build a reconciliation job comparing a key metric to its source-of-truth and gate downstream publishes when it fails.
  • Run an incident drill: simulate a missed freshness SLO, execute the runbook, and document learnings.

Who this is for

  • Analytics Engineers and BI Developers who publish models used for decisions.
  • Data Stewards and Product Analysts who define data expectations.

Prerequisites

  • Basic understanding of your data platform’s scheduling and monitoring.
  • Ability to write simple tests/queries to compute SLIs (e.g., row counts, time lags).

Learning path

  • First: define SLIs for one critical dataset.
  • Next: agree SLOs and error budget with stakeholders.
  • Then: implement alerts and a runbook, and practice an incident drill.
  • Finally: scale with templates, RACI, and a quarterly SLA review.

Mini challenge

Your marketing attribution model often completes by 06:20, but twice per month finishes at 06:50 due to upstream API limits. Finance wants 06:30 hard cutoffs; Marketing prefers occasional delays to avoid partial data. Draft an SLO and error budget that balances both, set severity thresholds, and outline who approves temporary exceptions.

Next steps

  • Apply the SLA template to one dataset this week.
  • Schedule a 30-minute review with the dataset’s business owner.
  • Take the quick test below to confirm understanding.

Practice Exercises

3 exercises to complete

Instructions

Choose a dataset used for daily decisions (e.g., Finance_MART.daily_bookings). Fill the fields below:

  • SLI 1 (freshness): definition and unit
  • SLI 2 (completeness): definition and threshold
  • SLOs: target and window
  • Error budget: allowed misses
  • Severity and response times
  • Ownership (RACI)
  • Exception rule and review cadence
Expected Output
A concise SLA (6–10 lines) with measurable SLIs, targets, clear ownership, and an exception process.

Data Quality SLAs And Ownership — Quick Test

Test your knowledge with 8 questions. Pass with 70% or higher.

8 questions70% to pass

Have questions about Data Quality SLAs And Ownership?

AI Assistant

Ask questions about this tool