This question evaluates understanding of confusion matrix components, mapping Type I/Type II errors to false positives/negatives, selection and interpretation of classification metrics (accuracy, precision/recall, F1, ROC-AUC, PR-AUC, calibration), threshold choice under asymmetric costs, and identification of practical pitfalls like class imbalance, data leakage, shifting base rates, and calibration issues. It is commonly asked in Statistics & Math interviews for Data Scientist roles to assess the ability to translate model performance into business impact, testing both conceptual understanding and practical application of evaluation and cost-sensitive decision-making.
You built a binary classifier (e.g., fraud detection, churn risk, medical screening, spam).
You are given a confusion matrix on a validation set: