PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Machine Learning/OpenAI

Train a classifier and analyze dataset

Last updated: Apr 28, 2026

Quick Overview

This question evaluates proficiency in end-to-end supervised machine learning workflows—covering data validation, exploratory data analysis, preprocessing, model training and selection, evaluation metrics and calibration, fairness assessments, and reproducibility—and tests competencies in applied machine learning engineering, statistical reasoning, and model governance within the Machine Learning domain. It is commonly asked in technical interviews because it verifies practical implementation skills and judgment for deploying reliable classifiers, aligning metrics to business goals and diagnosing performance and fairness across slices, with emphasis on practical application informed by conceptual understanding.

  • hard
  • OpenAI
  • Machine Learning
  • Machine Learning Engineer

Train a classifier and analyze dataset

Company: OpenAI

Role: Machine Learning Engineer

Category: Machine Learning

Difficulty: hard

Interview Round: Technical Screen

Given a labeled dataset, implement an end-to-end classifier training and dataset analysis workflow in Python. Perform exploratory data analysis (schema validation, missing values, target leakage checks, class imbalance, feature distributions). Create proper splits (stratified or time-based), define a baseline, and train at least two models (e.g., logistic regression and gradient boosting) with cross-validation and hyperparameter tuning. Handle imbalance (class weights or resampling), preprocessing (scaling, encoding), and calibration if needed. Choose metrics aligned with the business goal (ROC-AUC, PR-AUC, F1, cost-sensitive metrics), conduct error analysis (confusion slices, feature importances/SHAP), check fairness across key groups, and produce a concise report with the recommended model, expected performance, and code artifacts for reproducible deployment.

Quick Answer: This question evaluates proficiency in end-to-end supervised machine learning workflows—covering data validation, exploratory data analysis, preprocessing, model training and selection, evaluation metrics and calibration, fairness assessments, and reproducibility—and tests competencies in applied machine learning engineering, statistical reasoning, and model governance within the Machine Learning domain. It is commonly asked in technical interviews because it verifies practical implementation skills and judgment for deploying reliable classifiers, aligning metrics to business goals and diagnosing performance and fairness across slices, with emphasis on practical application informed by conceptual understanding.

Related Interview Questions

  • Implement Backprop for a Tiny Network - OpenAI (hard)
  • Filter Bad Human Annotations - OpenAI (medium)
  • Compute Matrix Prefix Products And Gradients - OpenAI (hard)
  • Improve Training With Noisy Annotators - OpenAI (hard)
  • Debug a Broken Transformer - OpenAI (medium)
OpenAI logo
OpenAI
Sep 6, 2025, 12:00 AM
Machine Learning Engineer
Technical Screen
Machine Learning
44
0

End-to-End Binary Classifier Workflow (EDA → Modeling → Fairness → Report)

You are given a labeled tabular dataset and asked to implement a reproducible, end-to-end workflow in Python to analyze the data and train a classifier suitable for deployment.

Assumptions (adapt as needed):

  • Input: a CSV file with a binary target column (e.g., target ∈ {0,1}).
  • Optional columns: a timestamp column for time-based splits; group columns for fairness checks; an ID column to drop.
  • Output: code, metrics, saved model artifact, and a concise text report with a recommended model and expected performance.

Requirements:

  1. Data access and schema validation
    • Load data; verify required columns exist; basic type checks and duplicate rows/IDs.
    • Summarize numeric/categorical feature counts and missingness.
  2. Exploratory data analysis (EDA)
    • Missing values: counts, percentages, imputation plan.
    • Target leakage checks: suspicious feature names, extremely high target correlation/MI.
    • Class imbalance: distribution and imbalance ratio.
    • Feature distributions: univariate summaries (hist/value counts) and basic outlier flags.
  3. Splitting strategy
    • If timestamp present: time-based split (train/validation/test by chronological order).
    • Else: stratified split to preserve class ratio.
  4. Baselines
    • Majority-class and simple model baseline (e.g., Logistic Regression with minimal tuning).
  5. Preprocessing
    • Numeric: impute (median), scale (standard).
    • Categorical: impute (most frequent), one-hot encode (handle_unknown=ignore); consider rare-category handling.
  6. Imbalance handling
    • Use class weights and/or sample weighting; optionally resampling (SMOTE/undersampling) if justified.
  7. Model training
    • Train at least two model families (e.g., Logistic Regression and Gradient Boosting).
    • Use cross-validation with hyperparameter tuning (RandomizedSearchCV or equivalent).
  8. Metrics aligned to business goal
    • Compute ROC-AUC and PR-AUC; report F1/precision/recall at a chosen threshold.
    • If provided, include cost-sensitive evaluation using FP/FN costs.
  9. Calibration (if needed)
    • Assess calibration; calibrate probabilities (Platt or isotonic) if poorly calibrated.
  10. Error analysis
    • Confusion matrix and per-slice analysis (e.g., by key categorical/numeric bins).
    • Feature importances (tree-based) and/or coefficients (linear); optionally SHAP.
  11. Fairness checks
    • Report metrics by key groups; highlight disparities (e.g., demographic parity, equal opportunity).
  12. Reproducibility and report
    • Save the fitted pipeline, metrics JSON, and environment info.
    • Produce a concise recommendation: chosen model, expected performance, and deployment notes.

Deliverables:

  • Python code implementing the above.
  • Saved model artifact and metrics.
  • Short written recommendation with expected performance and guardrails.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Machine Learning•More OpenAI•More Machine Learning Engineer•OpenAI Machine Learning Engineer•OpenAI Machine Learning•Machine Learning Engineer Machine Learning
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.