PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Analytics & Experimentation/Microsoft

Design evaluation when A/B test is impossible

Last updated: Mar 29, 2026

Quick Overview

This question evaluates a data scientist's competency in causal inference, observational experiment design, metric selection, and production monitoring within the Analytics & Experimentation domain.

  • easy
  • Microsoft
  • Analytics & Experimentation
  • Data Scientist

Design evaluation when A/B test is impossible

Company: Microsoft

Role: Data Scientist

Category: Analytics & Experimentation

Difficulty: easy

Interview Round: Technical Screen

## Scenario You want to evaluate whether a product or model change (e.g., a new ranking strategy, pricing rule, or UI change) improves business outcomes. However, **you cannot run a standard randomized A/B test** due to one or more constraints: - Legal/compliance restrictions (cannot randomize users) - Platform limitation (no experimentation framework) - Strong network effects / interference (user outcomes affect each other) - Rollout must be global (no holdout allowed) - Treatment is self-selected (users opt in) ## Questions 1. **What metrics** would you choose? - Propose a **primary metric** and at least 2 **diagnostic** and 2 **guardrail** metrics. - Explain tradeoffs (e.g., short-term vs long-term, sensitivity vs robustness). 2. **How would you estimate the counterfactual** (what would have happened without the change)? - Propose multiple causal inference approaches (at least 3), e.g. matching/weighting, difference-in-differences, synthetic control, regression discontinuity, instrumental variables, uplift modeling, etc. - For each approach, state key **assumptions**, what data you need, and how you would validate/pressure-test the assumptions. 3. **How would you handle common pitfalls**? - Confounding / selection bias - Seasonality and time trends - Delayed effects / novelty effects - Spillovers/interference - Missing data and metric instrumentation changes 4. Product monitoring follow-up: - Suppose after launch, the **core KPI drops sharply on a single day**. Outline a structured investigation plan to determine whether it’s (a) a real product issue, (b) a logging/pipeline issue, or (c) an external shock. - Include what slices you would check first and what “sanity checks” you would run.

Quick Answer: This question evaluates a data scientist's competency in causal inference, observational experiment design, metric selection, and production monitoring within the Analytics & Experimentation domain.

Related Interview Questions

  • Design Testing Without A/B Experiments - Microsoft (medium)
  • How would you estimate impact without A/B? - Microsoft (medium)
  • Design and analyze email deliverability experiment - Microsoft (hard)
  • Identify research to improve business - Microsoft (medium)
Microsoft logo
Microsoft
Feb 9, 2026, 11:59 AM
Data Scientist
Technical Screen
Analytics & Experimentation
1
0
Loading...

Scenario

You want to evaluate whether a product or model change (e.g., a new ranking strategy, pricing rule, or UI change) improves business outcomes.

However, you cannot run a standard randomized A/B test due to one or more constraints:

  • Legal/compliance restrictions (cannot randomize users)
  • Platform limitation (no experimentation framework)
  • Strong network effects / interference (user outcomes affect each other)
  • Rollout must be global (no holdout allowed)
  • Treatment is self-selected (users opt in)

Questions

  1. What metrics would you choose?
    • Propose a primary metric and at least 2 diagnostic and 2 guardrail metrics.
    • Explain tradeoffs (e.g., short-term vs long-term, sensitivity vs robustness).
  2. How would you estimate the counterfactual (what would have happened without the change)?
    • Propose multiple causal inference approaches (at least 3), e.g. matching/weighting, difference-in-differences, synthetic control, regression discontinuity, instrumental variables, uplift modeling, etc.
    • For each approach, state key assumptions , what data you need, and how you would validate/pressure-test the assumptions.
  3. How would you handle common pitfalls ?
    • Confounding / selection bias
    • Seasonality and time trends
    • Delayed effects / novelty effects
    • Spillovers/interference
    • Missing data and metric instrumentation changes
  4. Product monitoring follow-up:
    • Suppose after launch, the core KPI drops sharply on a single day . Outline a structured investigation plan to determine whether it’s (a) a real product issue, (b) a logging/pipeline issue, or (c) an external shock.
    • Include what slices you would check first and what “sanity checks” you would run.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Analytics & Experimentation•More Microsoft•More Data Scientist•Microsoft Data Scientist•Microsoft Analytics & Experimentation•Data Scientist Analytics & Experimentation
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.