PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Machine Learning/TikTok

How would you manage precision/recall for fraud detection?

Last updated: Mar 29, 2026

Quick Overview

This question evaluates a candidate's competency in applied machine learning for fraud detection, covering model performance measurement, thresholding, monitoring, and operational decisioning in a production setting; it is categorized under Machine Learning with a fraud-detection domain focus and tests both conceptual understanding and practical application for a data scientist role. It is commonly asked to probe the ability to select and balance a primary metric versus diagnostic metrics and operational guardrails, reason about cost asymmetry, label delay and distribution shift, and weigh trade-offs between product/user experience and fraud loss.

  • easy
  • TikTok
  • Machine Learning
  • Data Scientist

How would you manage precision/recall for fraud detection?

Company: TikTok

Role: Data Scientist

Category: Machine Learning

Difficulty: easy

Interview Round: Technical Screen

## Scenario You own (or significantly contribute to) a production **fraud detection** system that flags transactions/users as *fraud* vs *legit*. - The model outputs a fraud probability score \(p(\text{fraud})\). - A decision threshold determines whether to **block**, **step-up verify**, or **send to manual review**. - Labels may be delayed (chargebacks) and the data is **highly imbalanced**. ## Questions 1. **Precision/Recall management:** What concrete methods have you used (or would you use) to **measure, manage, and optimize precision and recall** in a real fraud system? 2. **False positives:** How would you diagnose and reduce **false positives** (legit users being flagged) without letting fraud through? 3. **Sudden fraud spike:** If you suddenly observe **many more fraud cases** than usual, what changes would you make (model, thresholding, monitoring, operations), and how would you validate them quickly? 4. **Specific fraud pattern:** If fraud shows a **very specific/pattern** (e.g., a new attack vector with clear signatures), what would you do—rules, model features, segmentation, retraining—and how would you prevent overfitting to a short-lived pattern? Please be explicit about: - The **primary metric** vs **diagnostic metrics** vs **guardrails** you would use. - How you handle **cost asymmetry** (FP vs FN), **label delay**, and **distribution shift/adversarial adaptation**. - The trade-off between **product/user experience** and **fraud loss**.

Quick Answer: This question evaluates a candidate's competency in applied machine learning for fraud detection, covering model performance measurement, thresholding, monitoring, and operational decisioning in a production setting; it is categorized under Machine Learning with a fraud-detection domain focus and tests both conceptual understanding and practical application for a data scientist role. It is commonly asked to probe the ability to select and balance a primary metric versus diagnostic metrics and operational guardrails, reason about cost asymmetry, label delay and distribution shift, and weigh trade-offs between product/user experience and fraud loss.

Related Interview Questions

  • Design multimodal deployment under compute limits - TikTok (easy)
  • Explain overfitting, dropout, normalization, RL post-training - TikTok (medium)
  • Write self-attention and cross-entropy pseudocode - TikTok (medium)
  • Implement AUC-ROC, softmax, and logistic regression - TikTok (medium)
  • Answer ML fundamentals and diagnostics questions - TikTok (hard)
TikTok logo
TikTok
Oct 26, 2025, 12:00 AM
Data Scientist
Technical Screen
Machine Learning
2
0
Loading...

Scenario

You own (or significantly contribute to) a production fraud detection system that flags transactions/users as fraud vs legit.

  • The model outputs a fraud probability score p(fraud)p(\text{fraud})p(fraud) .
  • A decision threshold determines whether to block , step-up verify , or send to manual review .
  • Labels may be delayed (chargebacks) and the data is highly imbalanced .

Questions

  1. Precision/Recall management: What concrete methods have you used (or would you use) to measure, manage, and optimize precision and recall in a real fraud system?
  2. False positives: How would you diagnose and reduce false positives (legit users being flagged) without letting fraud through?
  3. Sudden fraud spike: If you suddenly observe many more fraud cases than usual, what changes would you make (model, thresholding, monitoring, operations), and how would you validate them quickly?
  4. Specific fraud pattern: If fraud shows a very specific/pattern (e.g., a new attack vector with clear signatures), what would you do—rules, model features, segmentation, retraining—and how would you prevent overfitting to a short-lived pattern?

Please be explicit about:

  • The primary metric vs diagnostic metrics vs guardrails you would use.
  • How you handle cost asymmetry (FP vs FN), label delay , and distribution shift/adversarial adaptation .
  • The trade-off between product/user experience and fraud loss .

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Machine Learning•More TikTok•More Data Scientist•TikTok Data Scientist•TikTok Machine Learning•Data Scientist Machine Learning
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.