
You are given large-scale, logged observational data from an always-on promotion. Each record contains features X (user/context), a binary treatment A (promotion shown or not), and an outcome Y (e.g., revenue in the next 7 days). Treatment assignment was not randomized and may depend on X.
Goal: Learn heterogeneous treatment effects (HTE) to target users and maximize incremental revenue with a safe, fair, and evaluable deployment plan.
Assumptions (state and justify any deviations):
Propose a causal ML approach that addresses the following:
(a) Choose among T-/S-/X-/DR-learners for HTE, and justify your choice using bias–variance and ignorability considerations.
(b) Specify modeling choices for nuisance functions and HTE, and how you will use cross-fitting to reduce overfitting and nuisance bias.
(c) Describe how you will handle positivity/overlap violations and extreme propensities.
(d) Define evaluation metrics (policy value estimation, uplift/Qini, PEHE proxy) and explain how to implement off-policy evaluation with IPW and doubly robust estimators.
(e) Propose fairness constraints (e.g., demographic parity of treatment) and explain how they affect the learned policy.
(f) Outline how you would A/B test the learned policy safely before full deployment.
Login required