How to deploy and tune multimodal models?
Company: Bytedance
Role: Data Scientist
Category: Machine Learning
Difficulty: hard
Interview Round: Onsite
You are interviewing for a new-grad machine learning role. Answer the following related machine-learning and LLM questions.
1. **Multimodal deployment under constraints**: Suppose you need to deploy a multimodal model (for example, text + image or text + video) under strict **GPU memory, compute, and latency constraints**. Describe how you would reduce memory usage and inference latency while preserving as much model quality as possible. Discuss model-level changes, compression, batching/serving, and modality-specific optimizations.
2. **Fast video retrieval with captions and embeddings**: You already have a caption and one or more embedding vectors for each video in a large corpus. How would you design a retrieval system that answers user queries quickly while maintaining high recall? Discuss offline preprocessing, index design, approximate nearest neighbor search, lexical vs dense retrieval, reranking, and freshness / update trade-offs.
3. **Overfitting**: What is overfitting, how would you detect it, and what are the most effective ways to mitigate it in deep learning systems?
4. **Dropout**: Explain the intuition behind dropout, why it can reduce overfitting, and why the common "inverted dropout" implementation keeps the expected activation scale consistent between training and inference. Also mention when dropout may be less effective or even harmful.
5. **Normalization layers**: Compare Batch Normalization, Layer Normalization, Group Normalization, Instance Normalization, and RMSNorm. What statistics do they use, and how are training-time and inference-time behaviors different? Why are some normalization layers preferred in transformers or small-batch settings?
6. **Reinforcement learning for LLM post-training**: Explain how reinforcement learning is used in LLM post-training, especially in RLHF. Describe the typical pipeline from supervised fine-tuning to preference modeling and policy optimization, the role of KL regularization, and common failure modes. You may also contrast PPO-style RLHF with newer preference-optimization approaches such as DPO.
Quick Answer: This question evaluates expertise in deploying and tuning multimodal deep-learning models and related competencies including model-level optimization for memory/latency, model compression and serving strategies, embedding-based and lexical retrieval design for large video corpora, generalization and overfitting detection/mitigation, regularization and normalization behaviors, and reinforcement-learning methods for LLM post-training. It is commonly asked in Machine Learning interviews to assess practical engineering trade-offs and theoretical foundations across ML systems, information retrieval, and reinforcement learning, testing both conceptual understanding and practical application at a systems-to-algorithm level.