luvv to helpDiscover the Best Free Online Tools
Topic 12 of 12

Refresh and Scheduling

Learn Refresh and Scheduling for free with explanations, exercises, and a quick test (for Data Analyst).

Published: December 20, 2025 | Updated: December 20, 2025

Why this matters

Dashboards are only useful if the data is fresh, reliable, and updated on a predictable schedule. As a Data Analyst, you will set up refresh plans, choose the right data access mode, coordinate with data sources, and monitor jobs so stakeholders can trust what they see.

  • Plan refresh times to match when source systems update.
  • Pick between import, direct/live connection, incremental refresh, or streaming for the right latency and cost.
  • Configure credentials, gateways, and retries so refreshes run without manual work.
  • Monitor failures, alert the right people, and prevent repeated issues.
  • Control costs and load on source systems by scheduling smartly.

Concept explained simply

Refresh and scheduling is the discipline of deciding when and how your dashboard pulls new data, then reliably executing that plan.

Mental model: Water tank and faucets

Imagine your dashboard is a faucet. The storage behind it is a water tank (imported data cache). You can refill the tank on a schedule (scheduled refresh), refill only the newest water (incremental refresh), connect directly to the reservoir (direct/live), or use a real-time pipe (streaming). Your job is to ensure the faucet delivers the right temperature (freshness) with steady pressure (reliability), without flooding the system (overloading sources) or running dry (stale data).

Key concepts and terms

  • Data freshness (latency): Time between a source change and the dashboard showing it.
  • Import vs Direct/Live: Import copies data on refresh; Direct/Live queries the source each view.
  • Incremental refresh: Only refreshes recent partitions (e.g., last 30 days), speeding up jobs.
  • Streaming: Pushes events in near real-time for operational metrics.
  • Gateway/connector: Secure path from BI service to on-prem or private sources.
  • Refresh window: Time period when jobs are allowed to run.
  • Dependencies: One dataset/job depends on another completing first.
  • Retry and alerting: Automatic re-attempts and notifications on failure.
  • Query folding: BI tool pushes filters and aggregations to the source for efficiency.

Worked examples

Example 1: Daily retail sales dashboard (incremental import)
  1. Context: Data warehouse completes its nightly load around 02:10 local. Stakeholders want fresh data by 06:00.
  2. Choice: Import with incremental refresh (last 60 days), full historical partitions stay cached.
  3. Schedule: Run at 02:30 with 2 retries (10 min apart). Second run at 03:30 as a safety net.
  4. Gateway: Use on-prem gateway if warehouse is private; validate credentials weekly.
  5. Alerts: Notify #data-ops channel and dashboard owner on failure.
  6. Outcome: Data is ready by 04:00, meeting the 06:00 SLA with buffer.
Example 2: Marketing dashboard (warehouse nightly + API hourly)
  1. Context: Warehouse table updates nightly at 01:00. Ad API allows 200 calls/hour; data is delayed by the vendor by ~15 minutes.
  2. Choice: Mixed: Import incremental for warehouse tables; hourly import for API extracts with conservative pagination.
  3. Schedule: Warehouse refresh at 01:20. API refresh every hour at :20 past the hour. Final semantic model refresh at :30 past to align tables.
  4. Load control: Stagger pulls to avoid rate-limit spikes.
  5. Monitoring: Separate alerts for API failures with guidance to check rate-limit headers.
Example 3: Executive board report (weekly snapshot)
  1. Context: Executives need a weekly snapshot every Monday 08:00 with a frozen view for the week.
  2. Choice: Import full snapshot weekly; archive previous version as a new partition.
  3. Schedule: Mondays 05:00, with a validation step (row counts vs last week) before publishing.
  4. Governance: Tag dataset with snapshot date; document rollover process.

Step-by-step: design a refresh plan

  1. Define freshness SLA: How fresh must each KPI be (e.g., under 2 hours, daily by 06:00)?
  2. Map source timings: When do upstream systems finish loading? Add buffers for late runs.
  3. Choose access mode per table: Import (fast visuals), Direct/Live (lowest latency), Incremental (scales), Streaming (near real-time).
  4. Partition strategy: Decide number of days to refresh and how far back to keep historical partitions.
  5. Schedule windows: Place jobs after upstream completion; avoid peak app hours if sources are shared.
  6. Credentials and gateways: Set, test, and calendar reminders to rotate secrets before expiry.
  7. Retries and backoff: Configure 2–3 retries with delays; fail fast on auth errors.
  8. Alerts and ownership: Assign an owner; route alerts to a monitored channel.
  9. Test and validate: Dry-run on a smaller dataset; compare row counts and key metrics vs source.
  10. Document: Write the plan in the dataset description: sources, schedule, SLA, owner, and rollback steps.
Mini tasks
  • Write a one-line SLA for your dashboard.
  • List each source and the time it finishes loading.
  • Pick a refresh mode for each table and justify it in one sentence.

Practice exercises

Try these realistic scenarios. Solutions are provided, but attempt them first.

Exercise 1: Design a safe daily schedule

A retail analytics dashboard uses:

  • Data warehouse fact tables updated at 02:00 local with occasional 10-minute delays.
  • CRM API with a 5,000 calls/day limit; you pull 2 endpoints.
  • Business users in UTC+1 need fresh data by 07:00 local on weekdays.

Task: Propose a refresh plan that includes refresh times (UTC and local), mode per table, dependencies, retries, and alerting.

Exercise 2: Troubleshoot a failing refresh

Symptoms:

  • Refresh failed for the last 2 runs with timeout errors on a large dimension table.
  • Recent change: Added 12 months to the incremental window and a new calculated column.
  • Gateway shows increased CPU during refresh window.

Task: Identify likely root cause and propose step-by-step fixes and prevention.

Pre-flight checklist before you schedule

  • I know the required freshness for each KPI.
  • I documented source completion times and buffers.
  • Chosen access mode is justified (import/direct/incremental/streaming).
  • Credentials and gateway tests pass.
  • Alerts go to a monitored channel with an owner.
  • I tested a dry-run and validated row counts.
  • Retries, backoff, and maintenance windows are set.

Common mistakes and how to self-check

  • Mistake: Scheduling before sources finish loading. Self-check: Compare refresh start to upstream completion plus buffer; shift if too early.
  • Mistake: Full refresh for large histories. Self-check: Can you enable incremental refresh or partition pruning?
  • Mistake: Ignoring query folding. Self-check: Ensure filters push down; remove transformations that break folding.
  • Mistake: Credential expiry surprises. Self-check: Calendar rotate dates; test with a non-admin account.
  • Mistake: Overlapping heavy jobs. Self-check: Stagger schedules; cap concurrency.
  • Mistake: No alerts or unclear ownership. Self-check: Verify on-failure alerts and documented on-call person.

Practical projects

  • Project 1: Build a dataset with two sources (warehouse + API). Implement incremental refresh for warehouse and hourly refresh for API; write a 5-line runbook.
  • Project 2: Convert a full daily refresh to incremental with 90-day rolling window. Measure runtime and cost improvements before vs after.
  • Project 3: Set up a dashboard health panel: last refresh time, duration, rows read, and a simple traffic-light indicator for SLA status.

Mini challenge

Product analytics team needs < 2-hour freshness during work hours, but the source database is shared with an app and must not be overloaded. Propose a mode and schedule that meets freshness with minimal impact.

Hint

Consider Direct/Live for low-latency KPIs that are indexed, plus a lightweight cache or off-peak incremental import for heavier aggregates.

Who this is for

  • Data Analysts who build or maintain BI dashboards.
  • Analytics Engineers defining data pipelines and refresh strategies.
  • Team leads responsible for reporting SLAs.

Prerequisites

  • Basic SQL: filters, joins, simple aggregations.
  • Familiarity with at least one BI tool (e.g., how to publish datasets and set refresh).
  • Understanding of your data sources and when they update.

Learning path

  1. Master dataset design for efficient queries (star schema, proper data types).
  2. Learn access modes: import vs direct/live vs streaming and when to use each.
  3. Implement incremental refresh with sensible partitioning and folding.
  4. Set up monitoring, alerts, and documentation for refresh SLAs.

Next steps

  • Apply the checklist to one of your live dashboards and adjust schedules.
  • Add alerts and write a short runbook with common failure playbooks.
  • Iterate on partition windows to balance speed and completeness.

Quick test note: The quick test is available to everyone. If you are logged in, your progress will be saved.

Practice Exercises

2 exercises to complete

Instructions

You own a retail analytics dashboard. Sources:

  • Data warehouse fact tables updated at 02:00 local with occasional 10-minute delays.
  • CRM API with a 5,000 calls/day limit; you pull 2 endpoints, each ~1,200 calls per refresh.
  • Business users in UTC+1 need fresh data by 07:00 local on weekdays.

Task: Propose a refresh plan that includes:

  • Specific refresh times in both UTC and local time.
  • Mode for each table (import, incremental, direct/live) and why.
  • Dependencies and staggering to avoid rate limits.
  • Retries, alerting, and who owns failures.
Expected Output
A short plan stating schedule times (UTC and local), modes per table (e.g., incremental for facts), dependency order, retry count and spacing, and alert routing.

Refresh and Scheduling — Quick Test

Test your knowledge with 10 questions. Pass with 70% or higher.

10 questions70% to pass

Have questions about Refresh and Scheduling?

AI Assistant

Ask questions about this tool