PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Analytics & Experimentation/Uber

Design an A/B test for promo-targeting models

Last updated: Mar 29, 2026

Quick Overview

This question evaluates a data scientist's competency in experiment design, causal inference, metric engineering, power/sample-size calculation, and operational experimentation concerns such as logging, throttling, guardrails, and budget control, and is categorized under Analytics & Experimentation.

  • hard
  • Uber
  • Analytics & Experimentation
  • Data Scientist

Design an A/B test for promo-targeting models

Company: Uber

Role: Data Scientist

Category: Analytics & Experimentation

Difficulty: hard

Interview Round: Technical Screen

Design a controlled experiment to decide whether M1 beats M0 at assigning $5 promos. Requirements: 1) Randomize users into two arms; within each arm, rank by that arm’s model and send to the arm‑specific top‑K per day so both arms have equal expected spend and contact frequency. 2) Define the primary metric as incremental profit per eligible user (redeem uplift × expected GMV − $5) with guardrails (opt‑outs, uninstalls, support contacts). 3) Use triggered analysis (eligible, not recently contacted) and intent‑to‑treat as the estimand; log exposure, assignment, eligibility, and redemption events. 4) Compute required sample size given baseline redemption 3%, MDE +0.5 pp absolute, α=0.05 two‑sided, power 0.8; show the formula, include variance inflation for day‑level clustering, and plan for delayed outcomes. 5) Specify analysis: SRM checks, CUPED using pre‑period outcomes, and sequential monitoring with an alpha‑spending plan; report point estimates and CIs with robust (clustered) variance. 6) Address interference/saturation (e.g., users influencing each other or channel limits); if present, propose cluster randomization or switchback and explain trade‑offs. 7) Detail how you will keep budgets equal despite drift (e.g., per‑arm thresholding, rebalancing), handle throttling and suppression lists, and ensure no spillover between arms.

Quick Answer: This question evaluates a data scientist's competency in experiment design, causal inference, metric engineering, power/sample-size calculation, and operational experimentation concerns such as logging, throttling, guardrails, and budget control, and is categorized under Analytics & Experimentation.

Related Interview Questions

  • Design a Maps Address Search Bar - Uber
  • Evaluate a cold-start rating launch - Uber (medium)
  • Design Pricing Model Experiment - Uber (medium)
  • Evaluate marketplace interventions - Uber (medium)
  • Evaluate UberEATS priority delivery and membership - Uber (medium)
Uber logo
Uber
Oct 13, 2025, 9:49 PM
Data Scientist
Technical Screen
Analytics & Experimentation
6
0

Experiment Design: Compare Two Ranking Models (M1 vs M0) for $5 Promotions

Context

You have two models, M0 (current) and M1 (new), that rank users for a daily $5 promotion. The goal is to run a controlled experiment to decide whether M1 produces higher business value than M0 while keeping spend and contact frequency comparable across arms.

Assume:

  • Users can be eligible on a given day if they meet business rules and are not recently contacted (cool-down enforced).
  • Promotions are sent daily to a limited subset (top-K) within each arm based on that arm’s model ranking.
  • A $5 promo cost is incurred only upon redemption; the business value of a redemption is proportional to GMV (or margin).

Task

Design a controlled experiment that meets the following requirements:

  1. Randomization and Allocation
  • Randomize users into two arms (M0 and M1).
  • Within each arm, rank eligible users by that arm’s model and send to the arm-specific top-K per day so both arms have equal expected spend and contact frequency.
  1. Primary Metric and Guardrails
  • Define the primary metric as incremental profit per eligible user: (redeem uplift × expected GMV − $5), with guardrails on opt-outs, uninstalls, and support contacts.
  1. Triggering and Logging
  • Use triggered analysis: include user-days that are eligible and not recently contacted.
  • Use intent-to-treat (ITT) as the estimand.
  • Log exposure, assignment, eligibility, and redemption events.
  1. Power and Sample Size
  • Compute required sample size given: baseline redemption = 3%, MDE = +0.5 percentage points (absolute), α = 0.05 two-sided, power = 0.8.
  • Show the formula, include variance inflation for day-level clustering, and plan for delayed outcomes.
  1. Analysis Plan
  • Include SRM checks, CUPED using pre-period outcomes, and sequential monitoring with an alpha-spending plan.
  • Report point estimates and confidence intervals with robust (clustered) variance.
  1. Interference and Saturation
  • Address user-to-user interference and channel saturation limits; if present, propose cluster randomization or switchback and explain trade-offs.
  1. Budget Control and Operational Details
  • Keep budgets equal despite drift (e.g., per-arm thresholding and rebalancing), handle throttling and suppression lists, and ensure no spillover between arms.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Analytics & Experimentation•More Uber•More Data Scientist•Uber Data Scientist•Uber Analytics & Experimentation•Data Scientist Analytics & Experimentation
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.