Rapidly Improving Recall Under Class Imbalance (One-Day Plan)
Context
You inherit a binary fraud detection model with severe class imbalance (positive rate ≈ 2%). Evaluation on a temporally separated validation set shows:
-
ROC AUC = 0.61
-
Precision at 90% recall = 0.05 (very low precision at high recall, consistent with extreme imbalance)
-
Operations constraint: only 0.5% of traffic can be reviewed (fixed review capacity)
Goal: In one day, meaningfully improve recall at the same review capacity.
Tasks
-
Diagnosis: Describe how you would quickly distinguish underfitting versus overfitting using learning curves, calibration plots, PR vs ROC analysis at fixed capacity, and leakage/drift checks.
-
Interventions: Propose three changes you can implement in a day (e.g., class-weighted loss, monotonic gradient boosting with categorical encoders, threshold moving using cost-sensitive utility), and justify why each helps.
-
Thresholding for Utility: Show how to choose a decision threshold that maximizes expected utility given:
-
False Positive (FP) cost = $2
-
False Negative (FN) cost = $50
-
Review capacity = 0.5% of traffic
Provide the utility (or cost) formula and outline the selection procedure on validation data.
-
Monitoring: List the minimal logging/monitoring to add at deployment to detect drift and data quality issues within a week.