PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Analytics & Experimentation/TikTok

Investigate visit–report correlation causality

Last updated: Mar 29, 2026

Quick Overview

This question evaluates a candidate's competency in causal inference, experimental design, diagnostic analytics, and statistical modeling with ad-level observational data.

  • hard
  • TikTok
  • Analytics & Experimentation
  • Data Scientist

Investigate visit–report correlation causality

Company: TikTok

Role: Data Scientist

Category: Analytics & Experimentation

Difficulty: hard

Interview Round: Technical Screen

You observe a positive correlation between ad page visits and the probability an ad is reported as bad. Investigate whether more page visits cause more reports, or whether exposure/selection confounds the relationship. Provide: (1) at least three concrete hypotheses (e.g., pure exposure effect, targeting bias to sensitive cohorts, low-quality ads drive both attention and reporting, instrumentation effects), and the falsifiable prediction each makes; (2) a diagnostic plan including exact metrics and cuts: per-visit report rate vs. visit volume buckets; within-ad fixed-effects trends; cohorting by traffic source; funnel-based conditional probabilities; (3) a statistical model to estimate effect size controlling for exposure: e.g., ad-day logistic or Poisson regression with an offset log(visits), ad fixed effects, hour-of-day and geo controls; report the coefficient interpretation and how you would check overdispersion and multicollinearity; (4) at least one quasi-experimental approach (e.g., holdout geos, randomized delivery caps, or an IV using exogenous traffic shocks like site outages) and how you’d run a parallel A/A to validate; (5) criteria for action: define thresholds for an operational alert (e.g., ≥20% lift in per-visit report rate with p<0.01 after multiple-testing control) and what product/policy levers you’d recommend under each outcome.

Quick Answer: This question evaluates a candidate's competency in causal inference, experimental design, diagnostic analytics, and statistical modeling with ad-level observational data.

Related Interview Questions

  • Define Ultra success metrics and detect suspicious transactions - TikTok (easy)
  • Plan DS approach for biker delivery project - TikTok (easy)
  • Define and critique a user activity metric - TikTok (easy)
  • Design and decompose Trust & Safety risk metrics - TikTok (easy)
  • Analyze promo anomaly and design risk guardrails - TikTok (Medium)
TikTok logo
TikTok
Oct 13, 2025, 9:49 PM
Data Scientist
Technical Screen
Analytics & Experimentation
1
0

Causal Diagnosis: Do More Ad Page Visits Cause More Reports?

Context

You observe a positive correlation between the number of ad page visits and the probability that an ad is reported as "bad." Determine whether this is a causal exposure effect (more visits cause more reports) or is driven by selection/exposure confounding (e.g., certain traffic sources, targeting, or instrumentation issues).

Assume you have ad–day–geo–hour level data with counts of visits and reports, plus metadata (ad ID, traffic source, geo, device, time, user-level frequency where available).

Deliverables

  1. Hypotheses: Provide at least three concrete, falsifiable hypotheses (e.g., pure exposure effect, targeting bias to sensitive cohorts, low-quality ads drive both attention and reporting, instrumentation effects). For each, state a falsifiable prediction.
  2. Diagnostic plan: Specify exact metrics and cuts, such as per-visit report rate vs. visit volume buckets; within-ad fixed-effects trends; cohorting by traffic source; funnel-based conditional probabilities.
  3. Statistical model: Propose an ad–day model (logistic or Poisson/negative binomial) that controls for exposure via an offset log(visits), includes ad fixed effects and hour-of-day/geo controls. Explain coefficient interpretation and how you will check overdispersion and multicollinearity.
  4. Quasi-experimental approach: Propose at least one (e.g., holdout geos, randomized delivery caps, or an IV using exogenous traffic shocks like site outages) and describe how you’d run a parallel A/A to validate.
  5. Criteria for action: Define thresholds for an operational alert (e.g., ≥20% lift in per-visit report rate with p<0.01 after multiple-testing control) and recommend product/policy levers under each outcome.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Analytics & Experimentation•More TikTok•More Data Scientist•TikTok Data Scientist•TikTok Analytics & Experimentation•Data Scientist Analytics & Experimentation
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.