PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Machine Learning/TikTok

Compare Random Forests and Boosted Trees: Bias, Variance, Speed

Last updated: Mar 29, 2026

Quick Overview

This question evaluates understanding of ensemble machine learning methods—specifically differences between Random Forests and gradient-boosted decision trees—and the impact of feature preprocessing on tree-based models.

  • medium
  • TikTok
  • Machine Learning
  • Data Scientist

Compare Random Forests and Boosted Trees: Bias, Variance, Speed

Company: TikTok

Role: Data Scientist

Category: Machine Learning

Difficulty: medium

Interview Round: Technical Screen

##### Scenario Product team wants to understand pros/cons of Random Forests versus Boosted Decision Trees and whether any feature scaling is required for these tree-based algorithms. ##### Question Compare Random Forests and Gradient-Boosted Decision Trees (e.g., XGBoost) in terms of bias/variance, interpretability, training speed, and robustness to overfitting. Is feature standardization or normalization necessary for tree-based models? Explain why or why not. ##### Hints Focus on ensemble construction, sequential vs. parallel learning, split criteria, and how trees handle monotonic transformations.

Quick Answer: This question evaluates understanding of ensemble machine learning methods—specifically differences between Random Forests and gradient-boosted decision trees—and the impact of feature preprocessing on tree-based models.

Related Interview Questions

  • Design multimodal deployment under compute limits - TikTok (easy)
  • Explain overfitting, dropout, normalization, RL post-training - TikTok (medium)
  • Write self-attention and cross-entropy pseudocode - TikTok (medium)
  • Implement AUC-ROC, softmax, and logistic regression - TikTok (medium)
  • Answer ML fundamentals and diagnostics questions - TikTok (hard)
TikTok logo
TikTok
Jul 12, 2025, 6:59 PM
Data Scientist
Technical Screen
Machine Learning
155
0

Scenario

A product/data science team is deciding between Random Forests and Gradient-Boosted Decision Trees (e.g., XGBoost) for a new predictive task. They also want to know whether they must standardize or normalize features for these tree-based models.

Task

Compare Random Forests and Gradient-Boosted Decision Trees in terms of:

  • Bias vs. variance
  • Interpretability
  • Training speed and inference speed
  • Robustness to overfitting

Then answer: Is feature standardization or normalization necessary for tree-based models? Explain why or why not.

Hints

  • Contrast how ensembles are constructed (bagging vs. boosting), including parallel vs. sequential learning.
  • Note how split criteria (e.g., Gini/entropy, squared error) work in trees.
  • Explain how trees are invariant to monotonic transformations of features (ordering-based splits).

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Machine Learning•More TikTok•More Data Scientist•TikTok Data Scientist•TikTok Machine Learning•Data Scientist Machine Learning
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.