Explain key ML concepts and techniques
Company: Amazon
Role: Machine Learning Engineer
Category: Machine Learning
Difficulty: hard
Interview Round: Onsite
Explain how XGBoost achieves parallel computation during training. Compare feature-parallel vs data-parallel (histogram-based) approaches, discuss distributed training across machines, and outline trade-offs in memory and accuracy.
Explain layer normalization in Transformers: where it is applied (pre-norm vs post-norm), the mathematical formulation, why it stabilizes training, and its effect on gradient flow.
Design a multimodal neural network that fuses text and images; describe early/late/cross-attention fusion strategies, how to align modalities, handle missing modalities, and choose loss functions and evaluation metrics.
Explain collaborative filtering approaches: user-based vs item-based, matrix factorization for implicit feedback, regularization, cold-start mitigation, and scaling to sparse, large datasets.
Discuss multi-armed bandits: define regret, compare epsilon-greedy, UCB, and Thompson sampling, address non-stationary and contextual settings, and describe offline policy evaluation and safe deployment.
For logistic regression, derive the log-likelihood and gradients, compare L1 vs L2 regularization, interpret coefficients as odds ratios, handle class imbalance and calibration, and choose decision thresholds.
Quick Answer: This multi-part question evaluates proficiency in core machine learning competencies including algorithmic understanding (XGBoost parallelism, bandit algorithms, collaborative filtering), model training and optimization (distributed training, layer normalization, logistic regression regularization and calibration), multimodal architecture design and modality alignment, and practical considerations for scalability and evaluation. It is commonly asked in Machine Learning interviews to probe both conceptual understanding and practical application — assessing reasoning about trade-offs, communication and aggregation patterns, regularization and metrics — and therefore targets a mixed level of abstraction combining conceptual depth with implementation-oriented system-level thinking.