PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/ML System Design/Amazon

Design a fraud detection system

Last updated: Apr 16, 2026

Quick Overview

This question evaluates competency in ML system design for fraud detection, including real-time streaming inference, feature store architecture, delayed/noisy label handling, model selection and ensembling, latency budgeting, monitoring, and operational MLOps considerations.

  • hard
  • Amazon
  • ML System Design
  • Software Engineer

Design a fraud detection system

Company: Amazon

Role: Software Engineer

Category: ML System Design

Difficulty: hard

Interview Round: Technical Screen

Design a real-time payment fraud detection system. Discuss: events and labels (chargebacks, disputes), feature store (user, device, merchant, graph features), model selection (tree ensembles, deep models, anomaly detection), rule engine + model ensemble, data pipeline and streaming inference, latency budgets and fallbacks, thresholding to balance false positives vs. fraud loss, human-in-the-loop review, concept drift and adversarial adaptation, explainability requirements, online experiments, monitoring (precision at top-K, approval rate, fraud rate), and incident response/rollback.

Quick Answer: This question evaluates competency in ML system design for fraud detection, including real-time streaming inference, feature store architecture, delayed/noisy label handling, model selection and ensembling, latency budgeting, monitoring, and operational MLOps considerations.

Related Interview Questions

  • Design systems for global request detection and labeling - Amazon (hard)
  • Design a computer-use agent end-to-end - Amazon (medium)
  • Debug online worse than offline model performance - Amazon (medium)
  • Approach an ambiguous business problem - Amazon (medium)
  • Explain parallelism and collectives in training - Amazon (medium)
Amazon logo
Amazon
Aug 10, 2025, 12:00 AM
Software Engineer
Technical Screen
ML System Design
8
0

System Design: Real-Time Payment Fraud Detection

Context

Design a real-time fraud detection system for online payments (card-not-present). The system must score each transaction during authorization and decide whether to approve, decline, or route to manual review within a tight latency budget.

Assume:

  • End-to-end p95 decision latency budget: 100 ms (from feature retrieval to decision), with soft degradations permitted.
  • Labels (e.g., chargebacks) arrive with delays (weeks). You must train with delayed/noisy labels and operate with streaming features.

Requirements

Discuss and propose designs for:

  1. Events and Labels
  • What events to ingest (e.g., authorizations, captures, refunds, chargebacks, disputes, user actions).
  • How to define positive/negative labels (chargebacks, disputes) and handle label delay.
  1. Feature Store
  • Feature categories (user, device, merchant, payment instrument, velocity, graph/network features).
  • Offline vs. online stores, consistency, TTL, backfilling, and time-travel for training.
  1. Model Selection
  • Compare tree ensembles, deep models (e.g., sequence or representation models), and anomaly detection for cold start.
  • Calibration, class imbalance handling, and cost-sensitive learning.
  1. Rule Engine + Model Ensemble
  • Combining deterministic rules with ML scores, ensembling strategies, and reason codes.
  1. Data Pipeline and Streaming Inference
  • Ingestion, stream processing, feature computation, online retrieval, and a low-latency inference service.
  1. Latency Budgets and Fallbacks
  • Budget breakdown, caching, degradation paths (e.g., rules-only), and idempotency.
  1. Thresholding and Trade-offs
  • How to set thresholds to balance false positives vs. fraud loss; expected value formulation.
  1. Human-in-the-Loop Review
  • Review queue design, sampling strategies, SLAs, active learning, and feedback loops.
  1. Concept Drift and Adversarial Adaptation
  • Continuous training, drift detection, canaries, and defenses.
  1. Explainability Requirements
  • Feature attributions, rule traces, and audit logging.
  1. Online Experiments
  • A/B/shadow testing, guardrail metrics, ramp policy, and bias control.
  1. Monitoring and Alerting
  • Precision at top-K, approval rate, fraud rate, latency SLOs, data quality, and feature drift.
  1. Incident Response and Rollback
  • Kill switches, model/version rollback, runbooks, and postmortems.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More ML System Design•More Amazon•More Software Engineer•Amazon Software Engineer•Amazon ML System Design•Software Engineer ML System Design
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.