PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Machine Learning/Amazon

Explain core ML concepts and metrics

Last updated: Mar 29, 2026

Quick Overview

This question evaluates understanding of core machine learning concepts and competencies such as the bias–variance tradeoff, regularization, evaluation metrics for imbalanced classification (accuracy, precision, recall, F1, PR-AUC, ROC-AUC), logistic regression probabilities, ensemble methods, tree hyperparameters, and output activations (sigmoid vs softmax). It is commonly asked in technical interviews for Machine Learning and Data Scientist roles to assess reasoning about model behavior, metric selection, trade-offs and interpretability, testing both conceptual understanding and practical application of model evaluation and tuning.

  • easy
  • Amazon
  • Machine Learning
  • Data Scientist

Explain core ML concepts and metrics

Company: Amazon

Role: Data Scientist

Category: Machine Learning

Difficulty: easy

Interview Round: Technical Screen

You are interviewing for a **Data Scientist** role. Answer the following ML fundamentals questions clearly and concisely. ### Concepts 1. Explain the **bias–variance tradeoff**. How does it relate to **overfitting vs. underfitting**? 2. What are common forms of **regularization** (e.g., L1/L2, early stopping)? What problem does regularization solve? ### Imbalanced classification 3. For **imbalanced datasets**, which evaluation metrics do you prefer and why? Compare **accuracy, precision, recall, F1, PR-AUC, ROC-AUC**. 4. Define **precision** and **recall** and provide their formulas. When would you optimize for one over the other? ### Logistic regression / probabilities 5. In **logistic regression**, what is the model’s **raw output** before converting to a probability? How is it mapped to a probability? ### ROC-AUC interpretation 6. What does it mean if **ROC-AUC = 0.8**? Provide an intuitive interpretation and at least one caveat. ### Models 7. What are **ensemble models** and why do they often outperform a single model? 8. For **tree-based models** (decision trees / random forests / gradient boosting), name key **hyperparameters** and describe how they affect bias/variance. ### Output activations 9. Compare **sigmoid vs. softmax**: when do you use each, and how do their outputs differ?

Quick Answer: This question evaluates understanding of core machine learning concepts and competencies such as the bias–variance tradeoff, regularization, evaluation metrics for imbalanced classification (accuracy, precision, recall, F1, PR-AUC, ROC-AUC), logistic regression probabilities, ensemble methods, tree hyperparameters, and output activations (sigmoid vs softmax). It is commonly asked in technical interviews for Machine Learning and Data Scientist roles to assess reasoning about model behavior, metric selection, trade-offs and interpretability, testing both conceptual understanding and practical application of model evaluation and tuning.

Related Interview Questions

  • Explain Core ML Interview Concepts - Amazon (hard)
  • Evaluate NLP Classification Models - Amazon (easy)
  • Explain overfitting, regularization, and LLM techniques - Amazon (medium)
  • Explain NLP/RL concepts used in LLM agents - Amazon (hard)
  • Design and evaluate a RAG system - Amazon (easy)
Amazon logo
Amazon
Oct 11, 2025, 12:00 AM
Data Scientist
Technical Screen
Machine Learning
4
0

You are interviewing for a Data Scientist role. Answer the following ML fundamentals questions clearly and concisely.

Concepts

  1. Explain the bias–variance tradeoff . How does it relate to overfitting vs. underfitting ?
  2. What are common forms of regularization (e.g., L1/L2, early stopping)? What problem does regularization solve?

Imbalanced classification

  1. For imbalanced datasets , which evaluation metrics do you prefer and why? Compare accuracy, precision, recall, F1, PR-AUC, ROC-AUC .
  2. Define precision and recall and provide their formulas. When would you optimize for one over the other?

Logistic regression / probabilities

  1. In logistic regression , what is the model’s raw output before converting to a probability? How is it mapped to a probability?

ROC-AUC interpretation

  1. What does it mean if ROC-AUC = 0.8 ? Provide an intuitive interpretation and at least one caveat.

Models

  1. What are ensemble models and why do they often outperform a single model?
  2. For tree-based models (decision trees / random forests / gradient boosting), name key hyperparameters and describe how they affect bias/variance.

Output activations

  1. Compare sigmoid vs. softmax : when do you use each, and how do their outputs differ?

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Machine Learning•More Amazon•More Data Scientist•Amazon Data Scientist•Amazon Machine Learning•Data Scientist Machine Learning
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.