Why dashboard design matters for BI Analysts
Great dashboards turn data into decisions. As a BI Analyst, you translate business questions into a visual system that leaders can trust in minutes, not meetings. Good design means: the right KPI hierarchy, clear layout, focused filters, smooth drilldowns, and alerting that drives action. This skill unlocks faster decisions, fewer ad-hoc requests, and measurable business impact.
What you will be able to do
- Define decision-focused questions and the target audience.
- Build a KPI hierarchy and information architecture that scales.
- Design layouts that highlight what matters first.
- Implement filters, drilldowns, and cross-navigation without confusion.
- Set up alerting logic for proactive monitoring.
- Run usability tests and iterate based on feedback.
Who this is for
- BI Analysts who need to build or overhaul executive, operational, or product dashboards.
- Data-savvy professionals who want to deliver decision-ready analytics.
Prerequisites
- Basic SQL (SELECT, WHERE, GROUP BY, JOIN).
- Comfort with a BI tool (e.g., Power BI, Looker, Tableau) concepts: measures, dimensions, filters.
- Understanding of core business KPIs in your domain (revenue, churn, conversion, SLAs).
Learning path (practical roadmap)
Clarify the business question, decision cadence, and who will use the dashboard. Capture success criteria.
Mini task
Write one sentence: "This dashboard helps [role] decide [action] by tracking [KPI] weekly."
Define the North Star KPI, supporting KPIs, and diagnostic metrics. Group them into logical sections.
Mini task
Draw a 3-level tree: North Star → Drivers → Diagnostics.
Use a simple grid: top row for outcomes, middle for drivers, bottom for diagnostics. Apply pre-attentive cues (size, color).
Mini task
Sketch a 12-column grid and place tiles with size and color annotations.
Keep global filters minimal; add contextual drilldowns and drill-through pages for deeper analysis.
Mini task
Choose max 3 global filters; list 2 drill-through paths.
Define clear pathways between dashboards using consistent labels and parameters.
Mini task
Document source → destination and parameters passed.
Turn KPIs into monitors: who gets alerted, thresholds, and escalation cadence.
Mini task
Write a rule: "Alert if [metric] < [threshold] for [X periods]."
Run 15–30 min tests: can users find answers in under 10 seconds for top KPIs?
Mini task
Prepare 3 tasks and record time-to-answer and errors.
Prioritize feedback; fix high-impact usability issues first. Version the dashboard.
Mini task
Create a change log with date, change, and outcome.
Worked examples
Example 1: Define the KPI hierarchy for a Sales Executive dashboard
Goal: Help VP Sales decide where to focus weekly.
- North Star KPI: Monthly Recurring Revenue (MRR).
- Drivers: Pipeline coverage, Win rate, Average deal size, Sales cycle.
- Diagnostics: Stage conversion, Activity volume, Regional performance.
-- Driver: Pipeline Coverage (this month)
SELECT
SUM(amount) AS pipeline,
3 * SUM(target_mrr) AS target_pipeline,
SUM(amount) / NULLIF(3 * SUM(target_mrr),0) AS coverage_ratio
FROM deals
WHERE close_month = date_trunc('month', CURRENT_DATE);
Layout: Top = MRR vs target sparkline + YoY. Middle = drivers (4 tiles). Bottom = diagnostics table with conditional formatting.
Example 2: Visual prioritization and thresholds
Use color only for status; use size/position for importance.
- MRR card: large, top-left, bold number.
- Status colors: green ≥ 100% target, amber 90–99%, red < 90%.
CASE
WHEN mrr_ratio >= 1.0 THEN 'green'
WHEN mrr_ratio BETWEEN 0.9 AND 1.0 THEN 'amber'
ELSE 'red'
END AS status_color
Example 3: Filters and drill-through
Minimal global filters: Time period, Region, Segment. Drill-through to Account Detail.
-- Respect global filters
SELECT region, segment, SUM(mrr) AS mrr
FROM accounts
WHERE period = :period
AND (:region IS NULL OR region = :region)
AND (:segment IS NULL OR segment = :segment)
GROUP BY 1,2;
Drill-through passes account_id to detail view showing trends, contacts, and activities.
Example 4: Cross-dashboard navigation with parameters
From "Executive Sales" to "Rep Performance" with preset filters: region, manager, month. Keep label consistency: "View rep performance" on both ends. Ensure parameters map 1:1 in destination filters.
Example 5: Alerting and monitoring
Use rolling windows to reduce noise.
-- Alert if 3-day rolling churn rate > 1.2x 30-day baseline
WITH churn AS (
SELECT dt, churned_mrr / NULLIF(start_mrr,0) AS churn_rate
FROM daily_mrr
),
roll AS (
SELECT
dt,
AVG(churn_rate) OVER (ORDER BY dt ROWS BETWEEN 2 PRECEDING AND CURRENT ROW) AS r3,
AVG(churn_rate) OVER (ORDER BY dt ROWS BETWEEN 29 PRECEDING AND CURRENT ROW) AS r30
FROM churn
)
SELECT dt, r3, r30, (r3 > 1.2 * r30) AS alert
FROM roll
ORDER BY dt DESC
LIMIT 1;
Rule: Trigger only if condition holds 2 consecutive days and notify owner with next best action.
Example 6: Quick usability test script
Participants: 3–5 stakeholders.
- Task 1: "Find current MRR vs target and say green/amber/red." Target < 10 seconds.
- Task 2: "Which region underperformed this month?" Target < 30 seconds.
- Task 3: "Drill to a specific account with churn risk." Target < 60 seconds.
Record: time-to-answer, misclicks, comments. Fix issues causing >20% failure or >2 misclicks.
Design drills (hands-on)
- Rewrite 3 vague business requests into clear decision questions with a user role and cadence.
- Build a KPI tree with 1 North Star, 3–5 drivers, 5–10 diagnostics.
- Sketch a 3-row layout (outcomes, drivers, diagnostics) for two different audiences.
- Choose max 3 global filters for your dashboard; move the rest to contextual filters.
- Define 2 drill-through paths and list parameters passed.
- Write 2 alert rules with thresholds and rolling windows.
- Run a 20-minute usability test with at least 3 users; log time-to-answer.
Common mistakes and debugging tips
Mistake: Dashboard is a data dump
Fix: Start with decisions and users. Hide diagnostics by default; elevate outcomes and drivers.
Mistake: Too many global filters
Fix: Limit to 2–3. Move the rest into contextual panels or drill-through pages.
Mistake: Color overuse
Fix: Use color only for status/anomalies. Use size, position, and whitespace for priority.
Mistake: Drill paths are unclear
Fix: Use consistent action labels (e.g., "View details"). Keep level changes predictable (Summary → Segment → Entity).
Mistake: No alert hysteresis
Fix: Add rolling windows and consecutive-trigger rules to reduce noise.
Debugging tip: KPI doesn’t match other reports
Confirm metric definition, filter context, and time grain. Recompute in SQL with explicit filters to isolate discrepancies.
Mini project: Executive Sales Performance Dashboard
Goal: Help VP Sales decide weekly where to focus to hit MRR targets.
- Define: One-sentence purpose, audience, cadence, success criteria.
- Data prep: Build a model with Deals, Accounts, Targets, Activities.
- KPI hierarchy: MRR → Pipeline coverage, Win rate, Avg deal size, Sales cycle → Stage conversion, Activity volume.
- Layout: 3-row design; top cards for MRR, target ratio, sparkline; middle driver tiles; bottom diagnostic table.
- Filters: Global = Time, Region, Segment. Contextual = Owner, Stage.
- Drill-through: Region → Account details.
- Alerting: 3-day vs 30-day churn alert; pipeline coverage < 3x target alert.
- Usability test: 3 tasks, measure time-to-answer; apply fixes.
Acceptance criteria
- Key KPIs visible without scrolling on a laptop screen.
- No more than 3 global filters.
- At least 2 drill-through paths work with correct parameter context.
- Two alert rules documented with thresholds and owners.
- All test tasks solvable within the target times.
Subskills
- Defining Business Questions And Audience — Turn vague asks into decision-focused questions and match them to roles and cadences.
- KPI Hierarchy And Information Architecture — Build KPI trees and group content into logical, scalable sections.
- Layout And Visual Prioritization — Use grids, spacing, and pre-attentive attributes to guide attention.
- Filters And Drilldowns Design — Keep global filters minimal; add contextual filters and predictable drill paths.
- Navigation And Cross Linking Between Dashboards — Define clear pathways and pass parameters consistently.
- Alerting And Monitoring Use Cases — Translate KPIs into reliable alerts with rolling windows and owners.
- Usability Testing With Stakeholders — Run short, task-based tests; measure time-to-answer and misclicks.
- Iteration Based On Feedback — Prioritize and ship improvements in small, verifiable changes.
Next steps
- Work through the subskills to build mastery.
- Complete the mini project to integrate everything.
- Take the skill exam below. Anyone can take it; only logged-in users have their progress saved.