This question evaluates a data scientist's competency in experimental design, statistical inference for multiple comparisons, sequential testing and adaptive allocation methods (e.g., bandits), and principled metric selection and guardrails for product experiments.
You are the data scientist for a consumer health product running ~100 concurrent A/B tests under tight timelines and limited traffic. Leadership wants to understand how you design experiments, control error rates across many tests, adapt decisions as data accumulates, and choose robust product metrics.
Consider multiple-testing corrections (Bonferroni, Holm, FDR), sequential testing, power vs. speed trade-offs, and principled metric selection (north-star and guardrails).
Login required