
A “data analysis exam” can mean three different things: a pre-employment data analyst skills assessment, a certification-style exam, or a course final. Most pages ranking today blur these together—and then stop at generic competency lists. This package is built to help you simulate real exam conditions, identify skill gaps quickly, and turn results into a focused study plan.
This data analysis exam prep is designed for professionals preparing for time-boxed screening tests (common on platforms like CodeSignal-style analytics screens, HackerRank-style SQL/stat questions, and SHL-like mixed formats), as well as managers who want a job-relevant evaluation that is structured and scoreable. You’ll practice recurring question archetypes that show up repeatedly: top‑N per group, rolling metrics, funnel conversion, cohort retention, A/B test interpretation, and data-quality triage.
Unlike many vendor landing pages, this guide includes realistic sample questions with the reasoning employers often look for, a transparent scoring model with rubrics for open-ended recommendations, and level-based interpretations that map to typical role expectations (Junior Analyst → Product/BI Analyst → Senior/Lead).
The goal isn’t just to “get the right answer”—it’s to demonstrate sound analytical judgment under constraints.Use this page in two ways: (1) as a structured mock exam and review workflow, or (2) as a diagnostic that shows what to practice next—SQL patterns, Excel modeling, statistical reasoning, or stakeholder-ready communication—so you can improve efficiently.
Before you practice, clarify which exam you’re actually facing. Most frustration comes from preparing for the wrong format.
Typical focus: SQL + interpretation + stats fundamentals + business judgment (sometimes Excel).
Typical format: 45–90 minutes, auto-graded sections plus 1–2 open-ended prompts.
What’s typically reviewed: correctness, efficiency, and the quality of the reasoning (not just calculations).
Typical focus: broader coverage, more definitions and governance, sometimes tool-specific.
Typical format: 90–180 minutes, many multiple-choice questions, strict proctoring.
What’s typically reviewed: breadth and standard terminology.
Typical focus: theory + methods, often heavier stats and formal proofs/derivations.
Typical format: varies; may include long-form responses.
What’s typically reviewed: methodological correctness and explanation.
If you don’t know which one it is: assume pre-employment and prepare for (1) SQL under time pressure, (2) interpretation and metric choice, (3) experimentation basics.
This assessment focuses on skill areas that show up frequently in analytics roles and in analytics screening tests. It’s designed as a practical, job-relevant work sample with a structured scoring approach.
SQL & Data Retrieval (30%)
Joins, aggregation, HAVING, subqueries/CTEs
Window functions (rank/top-N, rolling metrics)
Date/time bucketing, NULL handling, deduplication
Sanity checks: grain, join explosion, filter logic
Spreadsheet Analysis (15%)
Cleaning, text/date functions
XLOOKUP/INDEX-MATCH, SUMIFS/COUNTIFS
PivotTables, basic chart selection
Statistics & Experimentation (20%)
Distributions, sampling bias, confidence intervals
Hypothesis tests, p-values vs practical significance
A/B testing interpretation, guardrail metrics
Correlation vs causation, basic regression intuition
Business Interpretation & Communication (25%)
KPI definitions, metric selection
Funnel/cohort reasoning, anomaly investigation
Writing: assumptions → method → insight → recommendation
Data Quality, Governance & Ethics (10%)
Data validation checks, reconciliation
Access control and appropriate use
Documenting assumptions and limitations
Why this approach is useful: A structured, rubric-based assessment can make review more consistent and transparent than unstructured evaluation alone.
Most exams recycle a small number of patterns. We teach and test those patterns explicitly.
Top-N per group (e.g., top products per category)
Rolling metrics (7-day moving average)
Funnel conversion (view → add-to-cart → purchase)
Cohort retention (week 0, 1, 2…)
Anomaly/root-cause (drop in conversion)
Deduplication (latest record per entity)
Join grain control (avoiding join explosions)
Segmentation (new vs returning, region, channel)
A/B interpretation (lift, confidence, guardrails)
Metric definition traps (active users, revenue, churn)
Data cleaning triage (missing values, outliers)
Visualization critique (misleading axes, wrong chart)
Auto-gradable items: SQL outputs, calculations, MCQ.
Rubric-graded items: written recommendations and critiques.
Partial credit: awarded for correct setup even if the final number is off.
You can run this as a 60–75 minute mock.
SectionTimeQuestion typesWeight
A. SQL25–30 min4 questions30%
B. Excel/Sheets10–12 min2 questions15%
C. Stats/Experimentation15–18 min2 questions20%
D. Business Case (written)12–15 min1 scenario25%
E. Data Quality & Ethics5 min1 scenario10%
Allowed tools (typical in pre-employment tests): basic calculator, scratch paper; sometimes a SQL editor/sandbox; usually no internet. If you’re unsure, practice without external lookup.
Assume an e-commerce product with tables:
You may assume timestamps are UTC.
Prompt: Compute daily revenue and daily purchasing users for the last 14 days.
Expected approach (high level):
Answer format: columns: order_date, revenue, purchasing_users
Prompt: For each country, return the top 3 coupon codes by total revenue in the last 30 days (exclude NULL coupon codes).
Key skills tested:
Prompt: For users created in the last 60 days, compute the funnel conversion rates:
Trap to avoid:
Answer format: one row with three rates.
Prompt: Create a 7-day rolling average of daily revenue for the last 30 days.
Key skills tested:
Prompt: You receive a column order_id that sometimes includes leading/trailing spaces and a revenue column imported as text with commas (e.g., "1,234.50").
Key skills tested: TRIM/CLEAN, VALUE/SUBSTITUTE, SUMIFS.
Prompt: A dashboard shows a line chart of revenue with the y-axis starting at 950,000 (not zero) and claims “Revenue skyrocketed 40%.”
Key skills tested: visualization integrity, communication.
Prompt: In an A/B test, Variant B shows a +1.2% absolute lift in purchase conversion. The 95% CI for lift is [-0.3%, +2.7%].
Key skills tested: statistical significance vs uncertainty, decision framing.
Prompt: You tested 10 UI variants and picked the one with the lowest p-value (0.04). Revenue increased slightly, but customer support tickets rose 8%.
Key skills tested: multiple testing, guardrails, practical significance.
Scenario: Yesterday conversion dropped from 3.1% to 2.4%. Marketing says traffic quality didn’t change. Engineering deployed a checkout change.
Prompt: Write a short analysis plan and recommendation (8–12 sentences). Include:
Key skills tested: structured thinking, triage, stakeholder communication.
Prompt: You’re asked to join employee performance ratings to badge-swipe data to “predict who’s likely to quit” and share results with managers.
Key skills tested: appropriate use, privacy, dignity/oversight, governance.
Total: 100 points. Use this scoring to self-grade or to support a more consistent review process.
Q1: 7 pts
Q2: 8 pts
Q3: 8 pts
Q4: 7 pts
SQL rubric (per question):
Q5: 9 pts (cleaning + SUMIFS correctness)
Q6: 6 pts (critique + better alternative)
Q7: 10 pts
Q8: 10 pts
Stats rubric:
Write-up rubric (0–5 each):
Rubric:
If you use this in hiring, treat results as one input:
What it means: You can follow tutorials, but you’re not yet reliable under exam pressure. Common issues: join explosions, incorrect grains, shaky experiment interpretation.
Action plan (next 2 weeks):
Role fit today: internship/entry-level only with strong support.
What it means: You can solve standard problems, but may struggle with advanced windows, ambiguous prompts, or clear recommendations.
Action plan (7–14 days):
Role fit today: Junior Data Analyst, Reporting Analyst.
What it means: You’re reliable across domains and can communicate tradeoffs. You likely perform well in many pre-employment screens.
Action plan (7 days):
Role fit today: Product Analyst, BI Analyst, Analytics Engineer (light) depending on tooling.
What it means: You show strong analytical judgment, anticipate pitfalls, and produce stakeholder-ready recommendations.
Action plan (ongoing):
Role fit today: Senior Analyst, Analytics Lead, Audit/Finance Analytics specialist.
SQL core (week 1): joins, aggregation, distinct counting, date buckets
SQL patterns (week 2): window functions basics, dedup (ROW_NUMBER)
Stats basics: CI intuition, p-values, practical significance
Practice routine: 30-minute timed drills; review mistakes with a checklist
Build an archetype playbook:
top-N per group template
rolling average template
funnel users template
cohort retention template
Start writing 1-page analyses: define metric, method, result, decision.
Time-boxed mocks weekly.
Add “sanity checks” to every query:
row counts before/after joins
sum of revenue vs baseline
distribution checks for outliers
Improve executive communication: headline + 3 bullets + risk/next step.
Own ambiguity: propose instrumentation improvements, define guardrails.
Mentor others: create reusable SQL snippets and documentation.
Tie analysis to decisions and value (cost, revenue, risk reduction).
Passing signal for many screens: ~70+ with no critical weaknesses (e.g., cannot interpret CI).
SQL is a common filter: Many hiring funnels overweight SQL because it’s easy to auto-grade and shows up frequently in day-to-day work.
Use job-relevant work samples where possible.
Keep scoring consistent with a rubric.
Pair results with structured, role-relevant interviews to support more consistent decision-making.
Practice: LeetCode SQL (easy→medium), StrataScratch (analytics-focused)
Reference topics to master:
window functions: ROW_NUMBER, RANK, LAG/LEAD, rolling frames
date handling: DATE_TRUNC, EXTRACT, interval arithmetic
NULL behavior: COALESCE, COUNT vs COUNT(*)
Skills to drill:
XLOOKUP (or INDEX/MATCH), SUMIFS/COUNTIFS
PivotTables, calculated fields
Text/date functions: TRIM, SUBSTITUTE, DATEVALUE
Core concepts:
confidence intervals, statistical vs practical significance
randomization checks, sample ratio mismatch
multiple testing and guardrails
Framework to use in write-ups:
Claim → Evidence → Impact → Next step
Keep a visible “Assumptions & Limitations” section.
Target roles emphasizing reporting and stakeholder support while you upskill (Reporting Analyst, Junior BI).
Build one portfolio project per archetype (funnel + cohort + A/B interpretation) using a public dataset.
Apply to Product Analyst / BI Analyst roles.
Prepare a 2-minute narrative for each project:
problem, metric, method, result, decision, business value.
Publish a case study that includes:
data quality checks
segmentation strategy
decision recommendation + monitoring plan
In interviews, proactively discuss risks: bias, instrumentation, governance.
First pass: answer what you can in <90 seconds per question.
Second pass: tackle multi-step SQL/windows.
Final 5 minutes: sanity checks and formatting.
Did I define the grain (per user, per order, per day)?
Did any join increase row counts unexpectedly?
Are NULLs handled intentionally?
Are dates bucketed correctly (off-by-one is a top failure mode)?
Do results look plausible vs baseline?
SQL joins/aggregation/window functions, Excel cleaning and pivoting, basic stats and A/B testing interpretation, plus business case reasoning and chart critique.
It’s hard if you rely on memorization. It’s manageable if you practice the recurring archetypes under time pressure and use a validation checklist.
Clear metric definitions, correct methods, explicit assumptions, sanity checks, and a recommendation with next steps—not just numbers.
For many pre-employment assessments, aim for 70+ with no domain at “critical weakness” (especially SQL grain control and stats interpretation).