Model Selection and Validation for a New Feature Launch
You are selecting and validating predictive models (supervised learning) for a new product feature. The goal is to estimate generalization performance, tune hyperparameters, and avoid overfitting or data leakage.
Questions
-
Explain k-fold cross-validation and why it helps control overfitting.
-
How do you decide on the value of k and what trade-offs exist?
-
What is nested cross-validation and when is it preferred?