This question evaluates the ability to design observability dashboards and visualize streaming-quality metrics, testing competencies in monitoring, anomaly detection, cohort-based diagnosis, percentile analysis, and system-level troubleshooting within the Analytics & Experimentation category for Data Engineer roles, and it blends conceptual understanding with practical application. It is commonly asked because it reveals a candidate's skill in choosing real-time versus retrospective views, handling seasonality and smoothing, defining diagnostic breakdowns and guardrail metrics, and designing interaction patterns for drill-downs and comparisons.

You are building an observability dashboard for a global consumer video product. Choose one critical streaming quality metric and design how you would visualize it to monitor health and diagnose issues.
Assume you have standard playback events (play_attempt, start, first_frame, rebuffer, stop), client metadata (device, OS, app version, network), and infrastructure metadata (CDN/POP, ISP, region). The dashboard should support both real-time monitoring and retrospective diagnosis.
Pick one critical metric (e.g., start-failure rate, rebuffering ratio, time-to-first-frame, hours watched per user) and specify:
Login required