PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Machine Learning/LinkedIn

Design a short-video recommendation system

Last updated: Mar 29, 2026

Quick Overview

This question evaluates a candidate's ability to design an end-to-end short-video recommendation system, testing competencies in label selection and trade-offs, feature and data engineering, multi-stage retrieval and ranking architectures, modeling challenges like cold start and feedback loops, and both offline and online evaluation.

  • medium
  • LinkedIn
  • Machine Learning
  • Data Scientist

Design a short-video recommendation system

Company: LinkedIn

Role: Data Scientist

Category: Machine Learning

Difficulty: medium

Interview Round: Technical Screen

Design a recommendation system for a short-video feed product. Your answer should cover the full pipeline: 1. **Objective and labels:** Define what the system should optimize. Discuss possible labels such as watch time, completion rate, likes, shares, follows, hides, reports, and long-term retention. Explain trade-offs between optimizing short-term engagement and long-term user satisfaction. 2. **Data and features:** Describe the training data you would collect, including impression logs, user actions, user history, creator features, video metadata, embeddings, freshness signals, and real-time context. Explain which features belong in candidate generation versus ranking. 3. **Three-stage recommendation architecture:** Propose a typical retrieval or candidate generation stage, a ranking stage, and a re-ranking or serving stage. Explain how each stage works and why the separation is useful for latency and scale. 4. **Modeling details:** Discuss negative sampling, delayed feedback, cold start, exploration versus exploitation, and how to avoid feedback loops or popularity bias. 5. **Offline evaluation:** Explain how you would evaluate the system offline, including ranking metrics and calibration. Discuss the limitations of offline evaluation when the training data comes from a previously deployed recommender and therefore has logging-policy bias. 6. **Online evaluation and guardrails:** Propose online experiment metrics and guardrails, including user experience, diversity, fairness, safety, and latency. Assume this is a large-scale consumer app with millions of users and videos, and that the interviewer wants both ML depth and product judgment.

Quick Answer: This question evaluates a candidate's ability to design an end-to-end short-video recommendation system, testing competencies in label selection and trade-offs, feature and data engineering, multi-stage retrieval and ranking architectures, modeling challenges like cold start and feedback loops, and both offline and online evaluation.

Related Interview Questions

  • Explain Logistic Regression, Backprop, and Adam - LinkedIn (medium)
  • Explain variance reduction in random forests - LinkedIn (medium)
  • Answer practical ML foundations questions - LinkedIn (medium)
  • Handle imbalance, sampling, and overfitting - LinkedIn (easy)
  • Handle imbalance, validate samples, and avoid overfitting - LinkedIn (easy)
LinkedIn logo
LinkedIn
Oct 12, 2025, 12:00 AM
Data Scientist
Technical Screen
Machine Learning
3
0

Design a recommendation system for a short-video feed product.

Your answer should cover the full pipeline:

  1. Objective and labels: Define what the system should optimize. Discuss possible labels such as watch time, completion rate, likes, shares, follows, hides, reports, and long-term retention. Explain trade-offs between optimizing short-term engagement and long-term user satisfaction.
  2. Data and features: Describe the training data you would collect, including impression logs, user actions, user history, creator features, video metadata, embeddings, freshness signals, and real-time context. Explain which features belong in candidate generation versus ranking.
  3. Three-stage recommendation architecture: Propose a typical retrieval or candidate generation stage, a ranking stage, and a re-ranking or serving stage. Explain how each stage works and why the separation is useful for latency and scale.
  4. Modeling details: Discuss negative sampling, delayed feedback, cold start, exploration versus exploitation, and how to avoid feedback loops or popularity bias.
  5. Offline evaluation: Explain how you would evaluate the system offline, including ranking metrics and calibration. Discuss the limitations of offline evaluation when the training data comes from a previously deployed recommender and therefore has logging-policy bias.
  6. Online evaluation and guardrails: Propose online experiment metrics and guardrails, including user experience, diversity, fairness, safety, and latency.

Assume this is a large-scale consumer app with millions of users and videos, and that the interviewer wants both ML depth and product judgment.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Machine Learning•More LinkedIn•More Data Scientist•LinkedIn Data Scientist•LinkedIn Machine Learning•Data Scientist Machine Learning
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.