Resolve Simpson’s paradox in A/B email test
Company: LinkedIn
Role: Data Scientist
Category: Analytics & Experimentation
Difficulty: easy
Interview Round: Technical Screen
A marketing team tests a new email campaign.
They run an experiment for **two weeks** in **two cities (SF and NY)** comparing **Email A vs Email B**.
They observe:
- In **each city (and/or each week)**, **B has a higher conversion rate than A**.
- But when they **combine all data**, **A has a higher overall conversion rate than B**.
## Questions
1. Explain how this can happen (Simpson’s paradox) and list the minimum conditions needed.
2. How would you determine whether **B is truly better than A**?
3. What metrics would you use (primary, diagnostic, guardrails), and what confounders would you worry about (e.g., city baseline differences, time-of-day/timezone effects, imbalance in allocation)?
4. Can you compute a confidence interval (CI) for the treatment effect? If yes, how (conceptually and/or with formulas)?
5. If the dataset is imbalanced across cities/weeks, what would you recommend operationally (reweighting, stratified analysis, rerun, blocking)?
Quick Answer: This question evaluates understanding of Simpson's paradox, causal inference, A/B testing and experimental design within the Analytics & Experimentation domain, covering metric selection, confounding, stratification and statistical estimation.