PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Machine Learning/Capital One

Model flight delays with EDA and explanation

Last updated: Mar 29, 2026

Quick Overview

This question evaluates a data scientist's competency in time-aware predictive modeling, exploratory data analysis for leakage and target drift detection, temporal cross-validation design, feature engineering and handling of rare categories and class imbalance, model selection (linear and tree-based), explainability and robustness testing, and deployment/experiment specification within the Machine Learning / Data Science domain. It is commonly asked because it probes practical application of production-ready ML workflows on temporally ordered data—assessing conceptual understanding of data leakage, drift, and validation alongside practical skills for metric selection, thresholding, inference contracts and operational reliability, so the level of abstraction spans both practical application and conceptual understanding.

  • hard
  • Capital One
  • Machine Learning
  • Data Scientist

Model flight delays with EDA and explanation

Company: Capital One

Role: Data Scientist

Category: Machine Learning

Difficulty: hard

Interview Round: Technical Screen

You are building a model to predict whether a domestic flight will arrive 15+ minutes late at wheels-down time, using only information available by scheduled departure. You receive a 50M-row table with columns: flight_date (YYYY-MM-DD), carrier, dep_airport, arr_airport, sched_dep_time (HH:MM local), dep_delay_min, arr_delay_min, distance_miles, weather_dep_* (temp, precip, vis), weather_arr_* (temp, precip, vis), holiday_flag, aircraft_tail, route_id. Label: late15 = 1 if arr_delay_min >= 15 else 0. Tasks: - EDA: list the exact checks/plots you would run to detect leakage, target drift, and rare-category issues; name at least 3 concrete leakage risks in these columns and how to mitigate each (e.g., removing/lagging features, using only pre-departure weather, excluding realized delays). - Validation: design a time-based cross-validation that respects seasonality and avoids look-ahead. Specify precise train/validation/test date windows and why you chose them. - Modeling: propose two candidates (one linear, one tree-based), feature engineering (cyclical encodings for time-of-day, airport- and carrier-level rolling aggregates, weather joins), handling class imbalance, primary metric(s), and how you would choose and calibrate a decision threshold for operational use. - Explainability & robustness: describe how you'd use SHAP/partial dependence safely with time-ordered data and how you'd test stability across airports/carriers (include at least two specific stress tests such as out-of-sample storms or new routes). - Deployment: define an inference contract (latency/SLAs, feature freshness, failure modes), and outline one A/B test to verify operational value (e.g., proactive rebooking or gate assignment); include success metrics and guardrails.

Quick Answer: This question evaluates a data scientist's competency in time-aware predictive modeling, exploratory data analysis for leakage and target drift detection, temporal cross-validation design, feature engineering and handling of rare categories and class imbalance, model selection (linear and tree-based), explainability and robustness testing, and deployment/experiment specification within the Machine Learning / Data Science domain. It is commonly asked because it probes practical application of production-ready ML workflows on temporally ordered data—assessing conceptual understanding of data leakage, drift, and validation alongside practical skills for metric selection, thresholding, inference contracts and operational reliability, so the level of abstraction spans both practical application and conceptual understanding.

Related Interview Questions

  • Deep-dive XGBoost handling and overfitting - Capital One (medium)
  • Build House Price Model Responsibly - Capital One (easy)
  • Design robber detection from surveillance video - Capital One (easy)
  • How would you design delay and watchlist models? - Capital One (medium)
  • Explain core ML concepts and lifecycle - Capital One (medium)
Capital One logo
Capital One
Oct 13, 2025, 9:49 PM
Data Scientist
Technical Screen
Machine Learning
3
0

Predicting 15+ Minute Arrival Delays at Scheduled-Departure Time

You are building a binary classifier that predicts whether a domestic flight will arrive 15+ minutes late (late15 = 1 if arr_delay_min ≥ 15, else 0), using only information available by scheduled departure time.

You receive a 50M-row table with these columns (one row per scheduled flight):

  • flight_date (YYYY-MM-DD)
  • carrier
  • dep_airport, arr_airport
  • sched_dep_time (HH:MM local)
  • dep_delay_min, arr_delay_min
  • distance_miles
  • weather_dep_* (temp, precip, vis)
  • weather_arr_* (temp, precip, vis)
  • holiday_flag
  • aircraft_tail
  • route_id
  • Label: late15

Assume we must restrict features to those known by scheduled departure and align any aggregates/forecasts accordingly.

Tasks

  1. EDA and Leakage/Drift Checks
  • List the exact checks/plots to detect:
    • Data leakage
    • Target drift over time
    • Rare-category issues
  • Name at least 3 concrete leakage risks present in these columns and how to mitigate each (e.g., remove/lag features, use only pre-departure weather, exclude realized delays).
  1. Validation Design
  • Propose a time-based cross-validation scheme that respects seasonality and avoids look-ahead.
  • Specify precise train/validation/test date windows and justify them.
  1. Modeling Plan
  • Propose two candidate models: one linear and one tree-based.
  • Feature engineering: cyclical encodings for time-of-day, airport/carrier rolling aggregates, and weather joins.
  • Handling class imbalance.
  • Primary metric(s).
  • How to choose and calibrate a decision threshold for operational use.
  1. Explainability & Robustness
  • How you would use SHAP/partial dependence safely with time-ordered data.
  • How you would test stability across airports/carriers, including at least two specific stress tests (e.g., out-of-sample storms, new routes).
  1. Deployment
  • Define an inference contract (latency/SLAs, feature freshness, failure modes).
  • Outline one A/B test to verify operational value (e.g., proactive rebooking or gate assignment), including success metrics and guardrails.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Machine Learning•More Capital One•More Data Scientist•Capital One Data Scientist•Capital One Machine Learning•Data Scientist Machine Learning
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.